AI‑assisted CLI to keep growing collections of files tidy. Organize folders with safe renames/moves and undo, watch directories for changes, and search collections with substring or semantic queries - all powered by portable per‑collection state.
Read the documentation: https://bryaneburr.github.io/dorgy/
- DSPy - Structured LLM queries and responses
- Docling - Document processing
- ChromaDB - Document search
- Durango - CLI Configuration management
Before (a messy folder):
my_docs/
IMG_0234.jpg
Scan_001.pdf
taxes.txt
contract_final_FINAL.docx
notes (1).txt
2023-05-07 14.23.10.png
invoice.pdf
After (organized by category/date with safe renames, hyphenated lower‑case folders):
my_docs/
.dorgy/ # state, history, search index, logs
documents/
contracts/
Employment Agreement (2023-06-15).pdf
taxes/
2023/
Tax Notes.txt
photos/
2023/05/
2023-05-07 14-23-10.png
invoices/
2023/
ACME - April.pdf
Exact destinations depend on your config and prompts; all moves are reversible via dorgy undo using the state in .dorgy.
Requires Python 3.11 or newer on macOS, Linux, or Windows.
pip install dorgygit clone https://github.com/bryaneburr/dorgy.git
cd dorgy
# Optional: install dev dependencies
uv sync --extra dev
# Optional: editable install
uv pip install -e .# Inspect available commands
dorgy --help
dorgy version
dorgy version --json
# Organize a directory in place (dry run first)
dorgy org ./documents --dry-run
dorgy org ./documents
# Monitor a directory and emit JSON batches
dorgy watch ./inbox --json --once
dorgy watch ./inbox # keep watching after the initial sweep
# Undo the latest plan
dorgy undo ./documents --dry-run
dorgy status ./documents --json- File events are debounced into batches so bursts of changes process together, and the CLI reports each completed batch.
- Before ingesting anything, every batch re-checks that files still exist, so deletes or moves that happen mid-queue become safe no-ops.
- Moves and deletes show up in CLI/JSON output as
removalsorsuppressed_deletions, with destructive changes gated byprocessing.watch.allow_deletions/--allow-deletions.
See the docs for guides on Organize, Watch, Search, Move/Undo, and configuration details.
Set language model credentials and defaults via dorgy config commands or the YAML file at ~/.dorgy/config.yaml. Important fields include:
llm.model— full LiteLLM/DSPy model identifier (e.g.,openai/gpt-4o-mini,openrouter/gpt-4.1).llm.api_key— API token for the selected provider (keep this in environment variables for security, e.g.,export DORGY__LLM__API_KEY=...).llm.api_base_url— optional custom gateway URL (useful for openrouter, proxies, or self-hosted backends).llm.temperature/llm.max_tokens— sampling parameters that shape response creativity and length.
To override values temporarily, export environment variables following the DORGY__SECTION__KEY scheme—for example:
export DORGY__LLM__MODEL="openai/gpt-4o-mini"
export DORGY__LLM__API_KEY="sk-example"
export DORGY__LLM__API_BASE_URL="https://api.openai.com/v1"Then run CLI commands as usual (dorgy org, dorgy watch, etc.).
We've tested dorgy with a number of LLMs and providers, and we've found the following to perform well:
- Gemini 2.5 (Best)
- Claude Sonnet 4.5
- GPT-5
- If you use OpenRouter, the
openrouter/automodel can give interesting results.
For many use cases, dorgy already performs well. YMMV depending on how much text content is in your files, the amount of context sent to the LLM, how good the LLM you're using is at this task, etc. That said, we are always looking to improve dorgy's performance and accuracy across a wide range of scenarios.
dorgy's development is ongoing. Here are some areas I'd like to explore next:
- Keep on adding file types and specialized handlers for: audio files, OCR'ing PDFs and other images containing text, and tabular data (CSV, Excel, etc.). Let me know if there's any other interesting/special file types you'd like
dorgyto handle. - Improve search beyond simple vector/semantic similarity and exact match, including for specialized file types.
- Improve the confidence scoring and "needs-review" system.
- Use
DSPy's evalution and optimization framework to makedorgyperform even better. If you would like to contribute the JSON output of your "good" and "bad" runs, reach out on our discussions page! Let's experiment!
We welcome issues and pull requests. See docs/development/contributing.md for environment setup, pre‑commit hooks, and CI guidance.
This repository includes Invoke tasks that wrap our uv commands. After installing dependencies, run:
uv run invoke --listCommon tasks include:
uv run invoke sync— update the virtual environment (installsdevanddocsextras by default).uv run invoke ci— replicate the CI pipeline locally (lint, mypy, tests, docs).uv run invoke docs-serve— launch the MkDocs server for live documentation previews.
Released under the MIT License. See LICENSE for details.
