qdrant-loader

martin-papy/qdrant-loader

3.3

If you are the rightful owner of qdrant-loader and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.

The Model Context Protocol (MCP) server is a crucial component for integrating AI development tools with vector databases, enabling intelligent search and retrieval capabilities.

QDrant Loader

šŸ“‹ - Latest improvements and bug fixes (June 18, 2025)

A comprehensive toolkit for loading data into Qdrant vector database with advanced MCP server support for AI-powered development workflows.

šŸŽÆ What is QDrant Loader?

QDrant Loader is a powerful data ingestion and retrieval system that bridges the gap between your technical content and AI development tools. It collects, processes, and vectorizes content from multiple sources, then provides intelligent search capabilities through a Model Context Protocol (MCP) server.

Perfect for:

  • šŸ¤– AI-powered development with Cursor, Windsurf, and GitHub Copilot
  • šŸ“š Knowledge base creation from scattered documentation
  • šŸ” Intelligent code assistance with contextual documentation
  • šŸ¢ Enterprise content integration from Confluence, JIRA, and Git repositories

šŸ“¦ Packages

This monorepo contains two complementary packages:

šŸ”„

Data ingestion and processing engine

Collects and vectorizes content from multiple sources into QDrant vector database.

Key Features:

  • Multi-source connectors: Git, Confluence (Cloud & Data Center), JIRA (Cloud & Data Center), Public Docs, Local Files
  • Advanced file conversion: 20+ file types including PDF, Office docs, images with AI-powered processing
  • Intelligent chunking: Smart document processing with metadata extraction
  • Incremental updates: Change detection and efficient synchronization
  • Flexible embeddings: OpenAI, local models, and custom endpoints

šŸ”Œ

AI development integration layer

Model Context Protocol server providing RAG capabilities to AI development tools.

Key Features:

  • MCP protocol compliance: Full integration with Cursor, Windsurf, and Claude Desktop
  • Advanced search tools: Semantic, hierarchy-aware, and attachment-focused search
  • Confluence intelligence: Deep understanding of page hierarchies and relationships
  • File attachment support: Comprehensive attachment discovery with parent document context
  • Real-time processing: Streaming responses for large result sets

šŸš€ Quick Start

Installation

# Install both packages
pip install qdrant-loader qdrant-loader-mcp-server

# Or install individually
pip install qdrant-loader          # Data ingestion only
pip install qdrant-loader-mcp-server  # MCP server only

5-Minute Setup

  1. Create a workspace

    mkdir my-qdrant-workspace && cd my-qdrant-workspace
    
  2. Download configuration templates

    curl -o config.yaml https://raw.githubusercontent.com/martin-papy/qdrant-loader/main/packages/qdrant-loader/conf/config.template.yaml
    curl -o .env https://raw.githubusercontent.com/martin-papy/qdrant-loader/main/packages/qdrant-loader/conf/.env.template
    
  3. Configure your environment (edit .env)

    QDRANT_URL=http://localhost:6333
    QDRANT_COLLECTION_NAME=my_docs
    OPENAI_API_KEY=your_openai_key
    
  4. Configure data sources (edit config.yaml)

    sources:
      git:
        - url: "https://github.com/your-org/your-repo.git"
          branch: "main"
    
  5. Load your data

    qdrant-loader --workspace . init
    qdrant-loader --workspace . ingest
    
  6. Start the MCP server

    mcp-qdrant-loader
    

šŸŽ‰ You're ready! Your content is now searchable through AI development tools.

šŸ”§ Integration Examples

Cursor IDE Integration

Add to .cursor/mcp.json:

{
  "mcpServers": {
    "qdrant-loader": {
      "command": "/path/to/venv/bin/mcp-qdrant-loader",
      "env": {
        "QDRANT_URL": "http://localhost:6333",
        "QDRANT_COLLECTION_NAME": "my_docs",
        "OPENAI_API_KEY": "your_key",
        "MCP_DISABLE_CONSOLE_LOGGING": "true"
      }
    }
  }
}

Example Queries in Cursor

  • "Find documentation about authentication in our API"
  • "Show me examples of error handling patterns"
  • "What are the deployment requirements for this service?"
  • "Find all attachments related to database schema"

šŸ“ Project Structure

qdrant-loader/
ā”œā”€ā”€ packages/
│   ā”œā”€ā”€ qdrant-loader/           # Core data ingestion package
│   └── qdrant-loader-mcp-server/ # MCP server for AI integration
ā”œā”€ā”€ docs/                        # Comprehensive documentation
ā”œā”€ā”€ website/                     # Documentation website generator
└── README.md                   # This file

šŸ“š Documentation

šŸš€ Getting Started

  • - Project overview and use cases
  • - Complete installation instructions
  • - 5-minute getting started guide
  • - Vector databases and embeddings explained

šŸ‘„ For Users

  • - Comprehensive user guides
  • - Git, Confluence, JIRA, and more
  • - PDF, Office docs, images processing
  • - AI development integration
  • - Complete configuration reference

šŸ› ļø For Developers

  • - Architecture and contribution guides
  • - System design and components
  • - Testing guide and best practices
  • - Deployment guide and configurations
  • - Custom data sources and processors

šŸ“¦ Package Documentation

  • - Core loader documentation
  • - MCP server documentation
  • - Documentation website

šŸ¤ Contributing

We welcome contributions! Please see our for details on:

  • Setting up the development environment
  • Code style and standards
  • Pull request process
  • Issue reporting guidelines

Quick Development Setup

# Clone the repository
git clone https://github.com/martin-papy/qdrant-loader.git
cd qdrant-loader

# Create virtual environment
python -m venv venv
source venv/bin/activate  # On Windows: venv\Scripts\activate

# Install in development mode
pip install -e packages/qdrant-loader[dev]
pip install -e packages/qdrant-loader-mcp-server[dev]

# Run tests
pytest

šŸ†˜ Support

  • Issues - Bug reports and feature requests
  • Discussions - Community discussions and Q&A
  • - Comprehensive guides and references

šŸ“„ License

This project is licensed under the GNU GPLv3 - see the file for details.


Ready to supercharge your AI development workflow? Start with our or explore the .