Quick Start Guide
Get up and running with QDrant Loader in 5 minutes! This guide walks you through your first document ingestion and AI tool integration.
๐ฏ What You'll Accomplish
In one flow, you will:
- Install the packages
- Start QDrant
- Create a workspace
- Ingest your first content
- Connect AI tools through MCP
Estimated time: 10 to 15 minutes.
๐ง Prerequisites
- Python 3.12+
- Docker (or an existing QDrant instance)
- One LLM provider key (OpenAI, Azure OpenAI, Ollama, or OpenAI-compatible)
๐ Step 1. Install packages
pip install qdrant-loader qdrant-loader-mcp-server
Verify:
qdrant-loader --version
mcp-qdrant-loader --version
If you need OS-specific install help, see Installation Guide.
๐ Step 2. Start QDrant
Local Docker option:
docker run -p 6333:6333 -p 6334:6334 qdrant/qdrant
Or use QDrant Cloud and copy URL/API key.
๐ค Step 3. Create workspace
Recommended (wizard):
qdrant-loader setup --output-dir my-qdrant-workspace --mode default
cd my-qdrant-workspace
Alternative (manual):
mkdir my-qdrant-workspace
cd my-qdrant-workspace
qdrant-loader init --workspace .
Need more control over prompts and templates? See CLI setup command options.
๐ง Step 4. Configure environment
Create or edit .env:
QDRANT_URL=http://localhost:6333
QDRANT_COLLECTION_NAME=quickstart
LLM_PROVIDER=openai
LLM_BASE_URL=https://api.openai.com/v1
LLM_API_KEY=your-openai-key
LLM_EMBEDDING_MODEL=text-embedding-3-small
LLM_CHAT_MODEL=gpt-4o-mini
Canonical configuration references:
- LLM Provider Guide - Pick the right provider profile and copy a known-good
.envtemplate. - Environment Variables Reference - Validate required keys fast and avoid common startup/auth errors.
๐ Step 5. Add a minimal config and ingest
Create config.yaml:
global:
qdrant:
url: "${QDRANT_URL}"
collection_name: "${QDRANT_COLLECTION_NAME}"
llm:
provider: "${LLM_PROVIDER}"
base_url: "${LLM_BASE_URL}"
api_key: "${LLM_API_KEY}"
models:
embeddings: "${LLM_EMBEDDING_MODEL}"
chat: "${LLM_CHAT_MODEL}"
embeddings:
vector_size: 1536
projects:
quickstart:
project_id: "quickstart"
display_name: "Quick Start"
sources:
localfile:
docs:
base_url: "file://./docs"
include_paths: ["**/*.md"]
Create sample content and ingest:
mkdir docs
printf "# Hello QDrant Loader\n\nThis is my first document.\n" > docs/sample.md
qdrant-loader ingest --workspace .
# Expected output:
# ๐ Scanning directory: my-project/
# ๐ Processing: 2 files found
# โ
Ingested: 2 documents, multiple chunks
# ๐ Collection: quickstart
For Git/Confluence/Jira and advanced source filters, see Data Sources Guide.
๐ Step 6. Start MCP server
mcp-qdrant-loader
# Expected output:
# ๐ QDrant Loader MCP Server starting...
# ๐ก Server running on stdio
# ๐ Available tools: search, hierarchy_search, attachment_search
# โ
Ready for connections
Detailed integration guides:
- Setup and Integration Guide - Connect MCP in Cursor, Claude Desktop, and other clients step by step.
- Search Capabilities Guide - Learn each search tool, parameters, and practical query patterns.
๐ Step 7. Validate in your AI tool
In Cursor/Claude/Windsurf, run a query like: "Search my docs for QDrant Loader quick start notes"
If results are returned from ingested content, setup is complete.
๐ฏ Next steps
- Configuration Reference - Tune chunking, embeddings, and project-level behavior for production use.
- Data Sources Guide - Expand beyond local files with Git, Confluence, Jira, and public docs.
- Troubleshooting Guide - Diagnose ingestion/search issues quickly with practical fix paths.
๐งช Quick Success Checklist
qdrant-loader --versionandmcp-qdrant-loader --versionreturn successfullyqdrant-loader ingest --workspace .finishes without errors- MCP server starts with
mcp-qdrant-loader - Your AI tool returns results from ingested documents
๐ Quick Start Complete!
You're now ready to explore the full power of QDrant Loader. The next step is reviewing the Core Concepts summarized in Getting Started, or dive into the User Guides for specific features and workflows.