MCP server and CLI for semantic search over your Obsidian vault. Generates embeddings with OpenAI, Ollama, or LM Studio. Stores vectors locally in sqlite-vec (~200KB, no telemetry, no network calls).
Search your notes by meaning, not just keywords:
obsidian-rag search "project architecture decisions" -n 5
obsidian-rag similar "Projects/Platform Hub.md"
obsidian-rag context "Daily Notes/2026-02-14.md"As an MCP server, it gives any compatible AI assistant the same capabilities — searching your notes, finding related content, and pulling context during conversations.
- Python 3.11+
- uv (for running and installing)
- One of:
OPENAI_API_KEY, Ollama, or LM Studio for embeddings
uvx obsidian-notes-rag setupThis creates a config at ~/.config/obsidian-notes-rag/config.toml with your vault path, embedding provider, and API key.
uvx obsidian-notes-rag indexParses your markdown files, chunks them by heading structure (using Chonkie RecursiveChunker), generates embeddings, and stores everything in a local SQLite database.
Works with any MCP-compatible client. Examples:
Claude Code:
claude mcp add -s user obsidian-notes-rag -- uvx obsidian-notes-rag serveClaude Desktop, Cursor, Windsurf, etc. (JSON config):
Add to your client's MCP config file (e.g. ~/Library/Application Support/Claude/claude_desktop_config.json for Claude Desktop on macOS):
{
"mcpServers": {
"obsidian-notes-rag": {
"command": "uvx",
"args": ["obsidian-notes-rag", "serve"]
}
}
}If you want obsidian-rag available as a standalone command:
uv tool install obsidian-notes-ragThis installs both obsidian-rag and obsidian-notes-rag to ~/.local/bin/.
Instead of running the MCP server, you can have your AI assistant call the CLI directly via shell commands. This avoids loading MCP tool definitions into the context window, freeing up tokens for your actual work.
To do this, create a rule or skill that tells your assistant when and how to use the CLI:
- Claude Code: Create a skill with CLI usage instructions
- Cursor: Add a rule to
.cursor/rules/ - Windsurf: Add a rule to
.windsurfrules
The rule should describe when to use each command (search, similar, context) and any project-specific conventions. This gives the assistant enough context to run the right CLI commands without the overhead of an MCP connection.
# Search
obsidian-rag search "query" # semantic search
obsidian-rag search "standup" --type daily # filter by note type
obsidian-rag search "design" -n 10 # more results
# Explore
obsidian-rag similar "Path/To/Note.md" # find related notes
obsidian-rag context "Path/To/Note.md" # show note + related context
# Index
obsidian-rag index # re-index vault
obsidian-rag index --clear # rebuild from scratch
obsidian-rag index --path-filter "Daily Notes/" # index subset
# Info
obsidian-rag stats # show index size
# Services
obsidian-rag serve # start MCP server
obsidian-rag watch # watch for changes, auto-reindex
obsidian-rag install-service # macOS launchd auto-start
obsidian-rag uninstall-service # remove service
obsidian-rag service-status # check service statusOnce connected, your AI assistant has access to:
| Tool | What it does |
|---|---|
search_notes |
Find notes matching a query |
get_similar |
Find notes similar to a given note |
get_note_context |
Get a note with related context |
get_stats |
Show index statistics |
reindex |
Rebuild the index |
Manual: obsidian-rag index
Auto-reindex on file changes: obsidian-rag watch (run in a terminal or background)
macOS background service: obsidian-rag install-service (starts on login, appears in System Settings > Login Items)
ollama pull nomic-embed-text
obsidian-rag --provider ollama indexLoad an embedding model in LM Studio, then:
obsidian-rag --provider lmstudio indexThe setup wizard writes to ~/.config/obsidian-notes-rag/config.toml. You can also override with environment variables:
| Variable | Description |
|---|---|
OPENAI_API_KEY |
OpenAI API key |
OBSIDIAN_RAG_PROVIDER |
openai (default), ollama, or lmstudio |
OBSIDIAN_RAG_VAULT |
Path to Obsidian vault |
OBSIDIAN_RAG_DATA |
Index storage path (default: platform-specific) |
OBSIDIAN_RAG_OLLAMA_URL |
Ollama URL (default: http://localhost:11434) |
OBSIDIAN_RAG_LMSTUDIO_URL |
LM Studio URL (default: http://localhost:1234) |
OBSIDIAN_RAG_MODEL |
Override embedding model |
- Parses markdown files, strips YAML frontmatter
- Chunks content using Chonkie's RecursiveChunker (splits by headings > paragraphs > lines > sentences, max 1500 tokens per chunk)
- Generates embeddings via your chosen provider
- Stores metadata in SQLite, vectors in sqlite-vec (KNN search via vec0 virtual tables)
- MCP server and CLI both query the same local database
If you installed the CLI with uv tool install, upgrade with:
uv tool upgrade obsidian-notes-ragIf you use uvx to run commands or the MCP server, it automatically uses the latest version.
v1.0.0 replaces ChromaDB with sqlite-vec. After upgrading, rebuild your index:
obsidian-rag index --clearThe old ChromaDB data at ~/.local/share/obsidian-notes-rag/ (or your configured path) can be deleted.
See CONTRIBUTING.md for development setup.
MIT