Skip to content

AntigmaLabs/ante-preview

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

153 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Ante

Ante

⚠️ Alpha Preview Ante is currently in alpha and provided as a research preview. Expect breaking changes and incomplete functionality. macOS and Linux only.

Ante is an AI-native, cloud-native, local-first agent runtime built by Antigma Labs. A single ~15MB Rust binary with zero runtime dependencies — designed from the ground up for security, performance, and resistance to AI-generated slop.

Key Features

  • Lightweight agent core — ~15MB binary, zero dependencies. Built for minimal overhead and maximum throughput.
  • Native local models — Built-in local inference integration. No API keys, no internet, no data leaving your device.
  • Zero vendor lock-in — Bring your own API key or local model. Switch between 12+ providers freely. No account required.
  • Client-daemon architecture — Run as an interactive TUI, headless CLI, or long-lived server (ante serve).
  • Channel integrations — Run Ante as a Slack or Discord bot with ante gateway.
  • Multi-agent orchestration — Spawn sub-agents, coordinate complex tasks across independent, decentralized, or centralized architectures.
  • Extensible — Custom skills, sub-agents, and persistent memory across sessions.
  • Benchmark proven — Topped the Terminal Bench 1.0 and 2.0 leaderboards. Public, reproducible evals.

Performance

We care about the harness not the model nor the prompts.

Ante is designed for the cellular-native thesis: agents lightweight enough to run hundreds of replicas in parallel on a single machine. Its ~15MB Rust core uses a fraction of the memory, CPU, and disk I/O of comparable agents — making mass parallelism practical without specialized infrastructure.

Docker resource usage across 20 parallel tasks (Ante vs Claude Code vs Opencode):

Resource Usage Comparison

Across 20 parallel tasks, Ante uses ~7× less peak memory, ~9× less average CPU, and generates ~5× less total disk I/O than Claude Code — while completing the same workload. See the comparison table and the benchmark details for the evaluation methodology and results.

Quick Start

Installation

Ante is distributed as a single, self-contained binary with no external dependencies — just download and run.

curl -fsSL https://ante.run/install.sh | bash

Interactive TUI

ante

Headless Mode

# Fix a bug
ante -p "find and fix the failing test in src/auth"

# Review a diff
git diff | ante -p "review this for security issues"

# Use a different provider
ante --provider openai --model gpt-5.4 -p "refactor the database module"

# Resume a saved session
ante --resume ses_01ARZ3NDEKTSV4RRFFQ69G5FAV -p "now add tests"

# Run fully offline with a local GGUF model
ante --offline-model ~/.ante/models/Qwen3.5-9B-Q4_K_M.gguf \
  -p "add error handling to src/main.rs"

Server Mode

ante serve

Gateway Mode

ante gateway

Update Ante

ante update

Example Usages with TUI

Providing Context: Files & Folders

Adding file context with @ mentions

Interrupting & Steering

Interrupting the agent with Escape

Models, Providers & Thinking

Selecting a model and provider

Subscription Login

Connecting to a provider via /connect

See all cookbook guides

Architecture

┌─────────────────────────────────────────────────────────────┐
│                         Clients                             │
│                                                             │
│   ┌───────────┐    ┌───────────┐    ┌────────────────────┐  │
│   │    TUI    │    │ Headless  │    │    ante serve      │  │
│   │  (ante)   │    │ (ante -p) │    │  (stdio / ws)      │  │
│   └─────┬─────┘    └─────┬─────┘    └─────────┬──────────┘  │
└─────────┼────────────────┼─────────────────────┼────────────┘
          │     Op         │                     │
          ▼                ▼                     ▼
┌─────────────────────────────────────────────────────────────┐
│                         Daemon                              │
│                                                             │
│   Session ──▶ Turn ──▶ Step                                 │
│                                                             │
│   ┌──────────┐  ┌──────────────┐  ┌───────────────────┐     │
│   │  Tools   │  │  Permission  │  │  Skills / Agents  │     │
│   └──────────┘  └──────────────┘  └───────────────────┘     │
└────────────────────────┬────────────────────────────────────┘
                         │
                         ▼
┌─────────────────────────────────────────────────────────────┐
│                     LLM Providers                           │
│                                                             │
│   Anthropic · OpenAI · Gemini · Grok · Open Router · Local  │
└─────────────────────────────────────────────────────────────┘

Supported Providers

Ante works with 12+ providers out of the box:

Provider Example Models
Anthropic Claude Sonnet 4.5, Opus 4.6
OpenAI GPT-5 family
Google Gemini Gemini 3 family
Grok (xAI) Grok 4
Open Router Multiple providers
Local (GGUF) Any GGUF model via built-in llama.cpp
...and more Vertex AI, Zai, Antix, OpenAI-compatible

Configure providers via environment variables (ANTHROPIC_API_KEY, OPENAI_API_KEY, etc.) or OAuth. Add custom providers in ~/.ante/catalog.json.

FAQ

Why Another Terminal agent

Ante is fast, lightweight, and the only terminal agent with native local inference support built in. We believe this self-contained agent core that self organize is the centre of the future of agent economy.

It is just built different.

How is Ante different than other agents On the high level, it has most of your favorite features (Multi-agents, skills, etc.) of your favorite agents (like Claude Code, Codex, etc.)
  • Ante is built from scratch in native Rust, we are obsessed with being self contained, so only essential libraries without framework or runtime dependencies.

  • You only need a llm provider configured to run it. Actually if you have the hardware, you don't even need a llm provider because Ante natively support private inference engine.

  • This resulted in ~15MB self-contained binary and multi-agent orchestration designed to run hundreds of replicas in parallel at scale. See the benchmark details across 20 parallel tasks for concrete numbers.

  • No vendor lock-ins, not even ourself. You don't need an account and can reuse your favorite api credentials.

Why care about runtime optimization like memory and I/O if model inference is usually the biggest bottleneck?

For one-on-one agent interactions, runtime overhead like memory usage and I/O is often less important than model inference.

But our vision is much bigger: millions of agents self-organizing and communicating at massive scale. At that point, even small inefficiencies get multiplied millions or billions of times, so runtime optimization becomes economically significant.

Can I run Ante completely offline?

Yes. Ante has a built-in llama.cpp engine that runs GGUF models locally. It handles engine installation, model discovery, and memory management automatically. No API keys or internet connection required.

Can I use my own custom models or providers?

Yes. Create a ~/.ante/catalog.json file to add or override providers and models with custom endpoints, API keys, and configurations. Any OpenAI-compatible API works.

What is the ante serve mode for?

Server mode runs Ante as a long-lived daemon that communicates over a structured JSONL protocol. It's ideal for building editor plugins, web UIs, and custom integrations on top of Ante.

How do I configure Ante?

Settings live in ~/.ante/settings.json. You can set your default model, provider, theme, and permission policy. CLI flags override settings for individual sessions. See the configuration docs for details.

Can I extend Ante with custom skills or sub-agents?

Yes. Drop skill files in ~/.ante/skills/ (user-level) or .ante/skills/ (project-level) using the Open Agent Skills format. Custom sub-agents go in ~/.ante/agents/ with their own prompts, tool sets, and model overrides.

Documentation

Full documentation is available at docs.antigma.ai. The source code is in docs-site/docs