files db mcp
A local vector database system that provides LLM coding agents with fast, efficient semantic search capabilities for software projects via the Message Control Protocol.
A local vector database system that provides LLM coding agents with fast, efficient semantic search capabilities for software projects via the Message Control Protocol.
A local vector database system that provides LLM coding agents with fast, efficient search capabilities for software projects via the Message Control Protocol (MCP).
# Using SSH (recommended if you have SSH keys set up with GitHub)
git clone [email protected]:randomm/files-db-mcp.git ~/.files-db-mcp && bash ~/.files-db-mcp/install/setup.sh
# Using HTTPS (if you Do not have SSH keys set up)
git clone https://github.com/randomm/files-db-mcp.git ~/.files-db-mcp && bash ~/.files-db-mcp/install/setup.sh
curl -fsSL https://raw.githubusercontent.com/randomm/files-db-mcp/main/install/install.sh | bash
After installation, run in any project directory:
files-db-mcp
The service will: 1. Detect your project files 2. Start indexing in the background 3. Begin responding to MCP search queries immediately
Files-DB-MCP works without configuration, but you can customize it with environment variables:
EMBEDDING_MODEL
- Change the embedding model (default: 'jinaai/jina-embeddings-v2-base-code' or project-specific model)FAST_STARTUP
- Set to 'true' to use a smaller model for faster startup (default: 'false')QUANTIZATION
- Enable/disable quantization (default: 'true')BINARY_EMBEDDINGS
- Enable/disable binary embeddings (default: 'false')IGNORE_PATTERNS
- Comma-separated list of files/dirs to ignoreOn first run, Files-DB-MCP will download embedding models which may take several minutes depending on: - The size of the selected model (300-500MB for high-quality models) - Your internet connection speed
Subsequent startups will be much faster as models are cached in a persistent Docker volume. For faster initial startup, you can:
# Use a smaller, faster model (90MB)
EMBEDDING_MODEL=sentence-transformers/all-MiniLM-L6-v2 files-db-mcp
# Or enable fast startup mode
FAST_STARTUP=true files-db-mcp
Files-DB-MCP automatically persists downloaded embedding models, so you only need to download them once:
model_cache
Add to your Claude Code configuration:
{
"mcpServers": {
"files-db-mcp": {
"command": "python",
"args": ["/path/to/src/claude_mcp_server.py", "--host", "localhost", "--port", "6333"]
}
}
}
For details, see Claude MCP Integration.
/src
- Source code/tests
- Unit and integration tests/docs
- Documentation/scripts
- Utility scripts/install
- Installation scripts/.docker
- Docker configuration/config
- Configuration files/ai-assist
- AI assistance filesContributions welcome! Please feel free to submit a pull request.