Introduce Ante: Self-Contained Agent That Self-Organize
Today Anthropic "open-sourced" Claude Code — and it's the perfect day to introduce Ante, the precursor of our mission, a self-contained agent built from first principles.
And why yet ANother TErminal agent? Well, "it's just built different."
Built Different
Here is a simple benchmark running same 20 parallel containerized tasks. Same models. Same workload. The only thing different is runtime.
Compared with Claude Code, Ante consumes:
- ~7x less peak memory
- ~9x less average CPU
- ~5x less disk I/O
while achieving same accuracy.
We believe agents should be like cells in a living organism — tiny, expendable, massively replicated. Lightweight enough to run by the thousands. Reliable enough that the system self-heals when any one of them fails.
For one agent doing one task, none of this matters. The model is the bottleneck. You could wrap the API in a shell script and it would mostly work.
But we're not building for one agent. We're building for hundreds — thousands — of agents that coordinate, delegate, fail, recover, and keep going. At that scale, every byte per instance, every wasted cycle, every unnecessary dependency gets multiplied a million times.
We call this the cellular-native thesis. Three pillars:
- Lightweight. Each replica can't cost gigabytes of memory — especially now, when RAM is expensive. Every line of code is a liability. Keep the core tight.
- Reliable. The return on reliability is non-linear. Below a threshold, the system is unusable. Above it, the game changes. There's a phase transition — you need to be on the right side of it.
- Closed-loop. Orchestration should be self-repairing — converging to a desired state automatically. Homeostasis in biology. Kubernetes in infrastructure. Individual agents are expendable; the organism persists.
Measure the Harness, Not the Model
We started with eval first. Benchmarks are the primary way to get real data before getting users. The principle: we would rather burn tokens than burn customer trust.
We use Terminal Bench and Harbor as our primary external benchmark. The team behind Harbor and Terminal Bench are genuine researchers — rigorous, principled, and motivated by getting it right rather than hype.
We're proud of the results — topped Terminal Bench 1.0 and Terminal Bench 2.0.
Most of the magic comes from the model — but the agent harness is the critical conduit between human and AI.
We evaluate the agent and how well it channels the model's power — not the model itself.
We aim to improve the harness, not the prompt.
Origin
Ante started as a side project to run open models on multiple heterogeneous commodity hardware with basic agentic capabilities backed by MCP, focusing on inference in embedding space.
Ante started as a side project to run open models on multiple heterogeneous commodity hardware
Then Claude Code showed us — and the industry — how much agent development is still uncharted territory. Full of excitement and unsolved problems, and critical to our mission of self-organizing intelligence.
Since then, we went all in, deep into the token space.
Self-Contained First
However, we also didn't forget our commitment to privacy and individual empowerment
Sovereign Compute and Network State, Second Amendment of the AI Era
Inspired by @karpathy's nanochat, we stripped down our prototype inference engine in pure Rust into a tiny hackable repo and open-sourced it (github.com/AntigmaLabs/nanochat-rs), as our tribute to the open-source community. And we are the first production-grade coding agent with first-class support for local models.
Alpha Preview
Try it, currently on macOS and Linux. Expect breaking changes.
curl -fsSL https://ante.run/install.sh | bash
- Docs: docs.antigma.ai
- GitHub: ante-preview
- Discord: https://discord.gg/CbAsUR434B
- HuggingFace: https://huggingface.co/Antigma
We'll share more in detail in the coming weeks. In the meantime — come try it and tell us what breaks.