Cost per task — Benchmark Sources & Consensus
Total dollar cost to complete a representative agent workflow.
Platforms tracked: Openclaw · Nemoclaw · Ironclaw · Hermes · Claude Cowork · Chatgpt
Consensus across 2 sources
Across 2 sources, coding agent costs range from near-zero (local tools) to $100+/month (Claude Code); open-source entrants like DeepSeek V4 claim 3-50x lower per-token cost than Claude tiers.
All Sources
We aggregate published benchmarks; we never run our own tests and never pick winners. Each row links back to the original publication.
| Source | Date | Finding | Methodology | Quality |
|---|---|---|---|---|
| GitHub | 2026-01-01 | 80+ coding agents surveyed: free local tools to $100+/month for Claude Code; Claude Code ranks highest by adoption; pricing varies widely by workflow and model tier | Survey of 80+ agents with self-reported or public pricing data; SWE-bench scores where available | medium |
| Hacker News | 2026-04-24 | DeepSeek V4-Pro claims open-source SOTA on agentic coding; V4-Flash at $0.14/$0.28/M tokens is 3-50x cheaper than Claude tiers | Self-reported SOTA ranking; official pricing data from announcement | medium |
How we work
OpenClawDatabase aggregates and links to published benchmarks. We don't run our own tests, and we don't pick winners. Our weekly benchmark-aggregator routine scans 7+ live leaderboards (OpenRouter, Aider, SWE-bench, GAIA, LMSYS, BigCodeBench, MMLU-Pro) plus relevant Reddit and Hacker News threads, then writes structured entries into /assets/benchmarks.json. Every row here links back to the original publication.
← Back to all benchmark tasks · See also: Decision guide · Cost calculator