Skip to content

An LLM-powered command line tool to automatically organize your files and directories

License

Notifications You must be signed in to change notification settings

bryaneburr/dorgy

Repository files navigation

CI PyPI - Version PyPI - Python Version GitHub License

Dorgy

dorgy logo

AI‑assisted CLI to keep growing collections of files tidy. Organize folders with safe renames/moves and undo, watch directories for changes, and search collections with substring or semantic queries - all powered by portable per‑collection state.

Read the documentation: https://bryaneburr.github.io/dorgy/

Powered by:

  • DSPy - Structured LLM queries and responses
  • Docling - Document processing
  • ChromaDB - Document search
  • Durango - CLI Configuration management

What It Does

Before (a messy folder):

my_docs/
  IMG_0234.jpg
  Scan_001.pdf
  taxes.txt
  contract_final_FINAL.docx
  notes (1).txt
  2023-05-07 14.23.10.png
  invoice.pdf

After (organized by category/date with safe renames, hyphenated lower‑case folders):

my_docs/
  .dorgy/                     # state, history, search index, logs
  documents/
    contracts/
      Employment Agreement (2023-06-15).pdf
    taxes/
      2023/
        Tax Notes.txt
  photos/
    2023/05/
      2023-05-07 14-23-10.png
  invoices/
    2023/
      ACME - April.pdf

Exact destinations depend on your config and prompts; all moves are reversible via dorgy undo using the state in .dorgy.

Installation

Requires Python 3.11 or newer on macOS, Linux, or Windows.

PyPI (recommended)

pip install dorgy

From source (contributors)

git clone https://github.com/bryaneburr/dorgy.git
cd dorgy

# Optional: install dev dependencies
uv sync --extra dev

# Optional: editable install
uv pip install -e .

Getting Started

# Inspect available commands
dorgy --help
dorgy version
dorgy version --json

# Organize a directory in place (dry run first)
dorgy org ./documents --dry-run
dorgy org ./documents

# Monitor a directory and emit JSON batches
dorgy watch ./inbox --json --once
dorgy watch ./inbox            # keep watching after the initial sweep

# Undo the latest plan
dorgy undo ./documents --dry-run
dorgy status ./documents --json

Watch batching & safety

  • File events are debounced into batches so bursts of changes process together, and the CLI reports each completed batch.
  • Before ingesting anything, every batch re-checks that files still exist, so deletes or moves that happen mid-queue become safe no-ops.
  • Moves and deletes show up in CLI/JSON output as removals or suppressed_deletions, with destructive changes gated by processing.watch.allow_deletions / --allow-deletions.

See the docs for guides on Organize, Watch, Search, Move/Undo, and configuration details.

Configuring LLM access

Set language model credentials and defaults via dorgy config commands or the YAML file at ~/.dorgy/config.yaml. Important fields include:

  • llm.model — full LiteLLM/DSPy model identifier (e.g., openai/gpt-4o-mini, openrouter/gpt-4.1).
  • llm.api_key — API token for the selected provider (keep this in environment variables for security, e.g., export DORGY__LLM__API_KEY=...).
  • llm.api_base_url — optional custom gateway URL (useful for openrouter, proxies, or self-hosted backends).
  • llm.temperature / llm.max_tokens — sampling parameters that shape response creativity and length.

To override values temporarily, export environment variables following the DORGY__SECTION__KEY scheme—for example:

export DORGY__LLM__MODEL="openai/gpt-4o-mini"
export DORGY__LLM__API_KEY="sk-example"
export DORGY__LLM__API_BASE_URL="https://api.openai.com/v1"

Then run CLI commands as usual (dorgy org, dorgy watch, etc.).

Notes on LLM Performance

We've tested dorgy with a number of LLMs and providers, and we've found the following to perform well:

  • Gemini 2.5 (Best)
  • Claude Sonnet 4.5
  • GPT-5
  • If you use OpenRouter, the openrouter/auto model can give interesting results.

For many use cases, dorgy already performs well. YMMV depending on how much text content is in your files, the amount of context sent to the LLM, how good the LLM you're using is at this task, etc. That said, we are always looking to improve dorgy's performance and accuracy across a wide range of scenarios.

What's Next?

dorgy's development is ongoing. Here are some areas I'd like to explore next:

  • Keep on adding file types and specialized handlers for: audio files, OCR'ing PDFs and other images containing text, and tabular data (CSV, Excel, etc.). Let me know if there's any other interesting/special file types you'd like dorgy to handle.
  • Improve search beyond simple vector/semantic similarity and exact match, including for specialized file types.
  • Improve the confidence scoring and "needs-review" system.
  • Use DSPy's evalution and optimization framework to make dorgy perform even better. If you would like to contribute the JSON output of your "good" and "bad" runs, reach out on our discussions page! Let's experiment!

Contributing

We welcome issues and pull requests. See docs/development/contributing.md for environment setup, pre‑commit hooks, and CI guidance.

Local Workflow Helpers

This repository includes Invoke tasks that wrap our uv commands. After installing dependencies, run:

uv run invoke --list

Common tasks include:

  • uv run invoke sync — update the virtual environment (installs dev and docs extras by default).
  • uv run invoke ci — replicate the CI pipeline locally (lint, mypy, tests, docs).
  • uv run invoke docs-serve — launch the MkDocs server for live documentation previews.

License

Released under the MIT License. See LICENSE for details.

About

An LLM-powered command line tool to automatically organize your files and directories

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages