← back to notes

Personal reference note. Numbers, prices, and other specifics may be outdated or wrong. Verify anything important before you rely on it.

AI Models — Use Cases & Pricing Reference

What GPT, Claude, Gemini and Cursor are good at, vision capabilities, and API pricing per token. Numbers may be outdated — verify before use.

📅 Apr 2026
AI LLM tools comparison

I keep forgetting which model to reach for and what things cost, so I wrote it down. Numbers are from April 2026 — prices change, so double-check before making decisions.


Models overview

Three main providers plus Cursor, which routes between them automatically for coding.

ProviderModelsKnown for
AnthropicHaiku 4.5 · Sonnet 4.6 · Opus 4.6Natural writing, long context, coding
OpenAIGPT-4o-mini · GPT-4o · GPT-5.2 · o3Broad ecosystem, all-round reliability
GoogleGemini 2.5 Flash · 2.5 Pro · 3.1 ProReasoning, long documents, cost efficiency
CursorRoutes between the aboveCoding — picks the model per task

Use cases

Coding

Claude Opus 4.6 leads on complex, multi-file work — 80.8% on SWE-bench Verified. Gemini 3.1 Pro is close behind (78.8%) at roughly 60% lower cost. Cursor’s auto mode routes mostly to Sonnet 4.6 and Gemini day-to-day, Opus for harder problems.

Reasoning and research

Gemini 3.1 Pro: 94.3% on GPQA Diamond, 77.1% on ARC-AGI-2. OpenAI’s o3 is designed for step-by-step reasoning — good for math and logic.

Writing and long documents

Claude. Most natural prose, up to 128K tokens output per pass. Sonnet 4.6 is the cost-efficient pick here.

High-volume / budget

Gemini 2.5 Flash at $0.075/M input. GPT-4o-mini at $0.15. Both fine for classification, summarization, scale tasks.

General everyday use

GPT-5.2 — widest ecosystem, solid at most things. Gemini 2.5 Pro is close at lower cost.


Vision and image understanding

Vision taskBest pick
Text/OCR from photosGPT-4o
Charts, diagrams, academic figuresGemini 3.1 Pro
Non-English document OCRGemini
UI screenshots → codeClaude Sonnet/Opus
Technical diagram interpretationClaude or GPT-4o
General photo descriptionGPT-4o or Gemini

Notes on each:

  • GPT-4o — best for text extraction from photos, layout-aware document parsing. Struggles with non-Latin scripts and medical imagery.
  • Gemini 3.1 Pro / Flash — 79.0 on MMMU Pro (charts, diagrams, academic figures). Best non-English OCR. Integrates natively with Google Docs/Sheets.
  • Claude Sonnet / Opus — best for UI screenshots and generating code from them. Solid on technical diagrams.

API pricing (April 2026)

These are API prices — developer pay-per-token, not subscriptions. ChatGPT Plus / Claude Pro / Gemini Advanced are flat monthly fees regardless of usage.

ModelInput / 1M tokensOutput / 1M tokens
Gemini 2.5 Flash$0.075$0.30
GPT-4o-mini$0.15$0.60
Claude Haiku 4.5$1.00$5.00
GPT-5.2$1.75$14.00
Gemini 3.1 Pro$2.00$12.00
GPT-4o$2.50$10.00
Claude Sonnet 4.6$3.00$15.00
Gemini 2.5 Pro$4.00$20.00
Claude Opus 4.6$5.00$25.00
o3 (reasoning)$10.00

Output tokens cost roughly 4–5× more than input. Prompt caching can cut input costs by up to 90%.


Quick pick table

TaskReach for
Complex multi-file codingClaude Opus 4.6 or Gemini 3.1 Pro
Everyday codingClaude Sonnet 4.6 (or Cursor auto)
Writing and long documentsClaude Sonnet or Opus
Step-by-step reasoningGemini 3.1 Pro or o3
General-purpose assistantGPT-5.2
High-volume / budgetGemini 2.5 Flash or GPT-4o-mini
OCR / text from photosGPT-4o
Charts and diagram readingGemini 3.1 Pro
UI screenshotsClaude Sonnet/Opus
Non-English documentsGemini

Sources — verify numbers here

← back to notes