
qdrant_project
支持多种嵌入提供程序的 Qdrant MCP 服务器,适用于语义记忆系统。
Repository Info
About This Server
支持多种嵌入提供程序的 Qdrant MCP 服务器,适用于语义记忆系统。
Model Context Protocol (MCP) - This server can be integrated with AI applications to provide additional context and capabilities, enabling enhanced AI interactions and functionality.
Documentation
Enhanced Qdrant MCP Server for Claude Code
This repository contains an enhanced version of the Qdrant MCP server with support for multiple embedding providers, making it more flexible for use with Claude Code as a semantic memory system.
Features
-
Multiple Embedding Providers:
- FastEmbed - The original lightweight embedding provider with ONNX Runtime
- GGUF - Direct integration with GGUF-format embedding models via llama.cpp
- LM Studio - Connectivity to LM Studio's API for optimized local embedding generation
-
Docker Compose Setup:
- Complete stack with Qdrant database, MCP servers, and UI
- Multiple MCP server instances for different providers
- Easy configuration through environment variables
-
Comprehensive Makefile:
- Simple commands for building, starting, and managing the stack
- Tools for adding different MCP configurations to Claude Code
- Utilities for downloading models and dependencies
Quick Start
Prerequisites
- Docker and Docker Compose
- Python 3.8+ with pip
- For GGUF: llama.cpp embedding binary
- For LM Studio: LM Studio application
Installation
-
Clone and set up:
cd /Users/oliviercontant/Sandbox/AI/mcp/qdrant make install-deps -
Start with FastEmbed (simplest option):
make start-fastembed -
Add to Claude Code:
make add-to-claude # For GGUF provider # OR make add-lmstudio-to-claude # For LM Studio provider
Using Different Embedding Providers
FastEmbed (Default)
The original provider that uses lightweight ONNX Runtime for generating embeddings:
# Run with Docker
make start-fastembed
# Run locally
make run-fastembed
# Add to Claude Code
claude mcp add qdrant-fastembed \
-e QDRANT_URL="http://localhost:6333" \
-e COLLECTION_NAME="code-memory" \
-e EMBEDDING_PROVIDER="fastembed" \
-e EMBEDDING_MODEL="sentence-transformers/all-MiniLM-L6-v2" \
-- uvx mcp-server-qdrant
GGUF Models
Use quantized GGUF embedding models for efficient memory usage:
# Run with Docker
make start-gguf
# Run locally
make run-gguf
# Add to Claude Code with custom model
make add-to-claude MODEL_NAME=all-MiniLM-L6-v2-Q8_0.gguf HUGGINGFACE_REPO=TheBloke/all-MiniLM-L6-v2-GGUF
For more details, see GGUF Support Documentation.
LM Studio
Connect to LM Studio for optimized local embedding generation:
# Start LM Studio and load your embedding model first
# Run with Docker
make start-lmstudio
# Run locally
make run-lmstudio
# Add to Claude Code with custom settings
make add-lmstudio-to-claude LMSTUDIO_API_BASE=http://localhost:1234/v1 LMSTUDIO_MODEL=Snowflake/snowflake-arctic-embed-l-v2.0 LMSTUDIO_VECTOR_SIZE=768
For more details, see LM Studio Support Documentation.
Environment Variables
Each embedding provider has its own set of environment variables:
Common Variables
QDRANT_URL- URL of the Qdrant serverCOLLECTION_NAME- Name of the Qdrant collectionEMBEDDING_PROVIDER- Provider type (fastembed,gguf, orlmstudio)EMBEDDING_MODEL- Model name or pathEMBEDDING_VECTOR_SIZE- Size of embedding vectors (optional)
GGUF-specific Variables
LLAMA_CPP_PATH- Path to llama.cpp embedding binaryEMBEDDING_MAX_CONTEXT_LENGTH- Maximum context length for embeddings
LM Studio-specific Variables
LMSTUDIO_API_BASE- Base URL of the LM Studio APILMSTUDIO_API_KEY- API key for authentication (optional)
Using with Claude Code
-
Start the Qdrant server:
make start -
Add to Claude Code:
# Choose one provider: make add-to-claude # GGUF # OR make add-lmstudio-to-claude # LM Studio -
Verify in Claude Code:
claude mcp list
Troubleshooting
- Connection issues: Ensure Qdrant is running (
make start) - GGUF model errors: Check that llama.cpp is properly installed (
make download-llama-cpp) - LM Studio errors: Verify LM Studio is running with your model loaded (
make check-lmstudio)
Advanced Usage
See the Makefile for more commands:
make help
Additional Documentation
- GGUF Support Documentation
- LM Studio Support Documentation
- Original Qdrant MCP Server Documentation
Quick Start
Clone the repository
git clone https://github.com/ocontant/qdrant_projectInstall dependencies
cd qdrant_project
npm installFollow the documentation
Check the repository's README.md file for specific installation and usage instructions.
Repository Details
Recommended MCP Servers
Discord MCP
Enable AI assistants to seamlessly interact with Discord servers, channels, and messages.
Knit MCP
Connect AI agents to 200+ SaaS applications and automate workflows.
Apify MCP Server
Deploy and interact with Apify actors for web scraping and data extraction.
BrowserStack MCP
BrowserStack MCP Server for automated testing across multiple browsers.
Zapier MCP
A Zapier server that provides automation capabilities for various apps.