Compare AI model pricing for any agent workflow
estimate_agent_cost compares the cost of running any AI agent task across Claude, GPT-4, Gemini, Llama, Mistral, and 20+ other models, returning per-call and per-day cost breakdowns. Free to use, no API key required.
Specify your input tokens, output tokens, and number of API calls per task. The tool returns a ranked comparison table so you can pick the cheapest model that meets your quality requirements.
Pricing data covers the latest model versions including Claude Opus 4, Claude Sonnet 4, GPT-4o, Gemini 2.5 Pro, Llama 4, and Mistral Large. The table updates as providers change their rates.
claude mcp add agentic-platform --transport http https://api.asalvocreative.com/mcp
{
"mcpServers": {
"agentic-platform": {
"url": "https://api.asalvocreative.com/mcp"
}
}
}
npx -y @smithery/cli@latest install agentic-platform--andysalvo.run.tools --client claude
model -- Model name to highlight in the comparison (optional)input_tokens -- Estimated input tokens per API call (default: 1000)output_tokens -- Estimated output tokens per API call (default: 500)num_calls -- Number of API calls per task run (default: 1)task_description -- What the agent does, for optimization tips (optional)estimate_agent_cost({
model: "claude-sonnet-4",
input_tokens: 2000,
output_tokens: 1000,
num_calls: 5,
task_description: "Code review agent that reads PRs and suggests fixes"
})
Agent Cost Estimate
===================
Task: Code review agent (5 calls/run, 2000 in + 1000 out per call)
Model Per Call Per Run Per Day (10 runs)
---------------------------------------------------------------
claude-sonnet-4 > $0.0090 $0.0450 $0.4500
claude-haiku-3.5 $0.0013 $0.0063 $0.0625
gpt-4o $0.0075 $0.0375 $0.3750
gpt-4o-mini $0.0005 $0.0023 $0.0225
gemini-2.5-flash $0.0004 $0.0019 $0.0188
llama-4-scout $0.0003 $0.0016 $0.0163
...
Optimization tips:
- Consider Haiku for routine reviews, Sonnet for complex ones
- Cache system prompt to reduce input token costs
- Batch small PRs to reduce per-call overhead
It depends on the model, token usage, and call volume. A typical 5-call agent task costs $0.002 to $0.045 per run depending on the model. Gemini 2.5 Flash and GPT-4o mini are the cheapest at scale. Claude Sonnet and GPT-4o cost more but produce higher-quality output. Use estimate_agent_cost with your exact parameters to get a full comparison.
For most workloads, Gemini 2.5 Flash and GPT-4o mini offer the lowest per-call cost. Llama 4 Scout via API providers is also competitive. The best choice depends on your input/output token ratio and whether you need frontier-level reasoning. The tool ranks every model by cost for your specific use case.
GPT-4o runs approximately $2.50/$10 per million tokens (input/output). Claude Sonnet 4 runs $3/$15. Claude Haiku 3.5 is significantly cheaper at $0.80/$4. For high-volume agent tasks, the difference compounds quickly. Run estimate_agent_cost to see exact numbers for your token counts.
Set your expected input_tokens, output_tokens, and num_calls per task. The tool returns a side-by-side table with per-call, per-run, and per-day costs across all major models. No signup needed. Connect to the MCP endpoint and call estimate_agent_cost.
The Agentic Platform is an MCP server providing free agent diagnostics and expert skill files. In addition to the cost estimator, it offers an MCP linter, agent health check, and service trust evaluator -- all free, no API key required.