Gemini 2.5 Pro — March–June 2025¶
What Changed¶
Google DeepMind's Gemini 2.5 Pro became the top-ranked model on LMArena (Elo 1470) and WebDevArena (1443) after its full release in June 2025. Its defining features: 1 million token context, Deep Think mode (configurable reasoning budget up to 32K thinking tokens), and grounded search natively integrated.
Gemini 2.5 Pro moved the context scaling frontier from 128K to 1M tokens while maintaining competitive quality on multi-hop reasoning and coding tasks.
Key Technical Details¶
Configurable thinking budget: unlike o1/o3 which have opaque reasoning, Gemini 2.5 Pro exposes thinking_budget as a parameter controlling how many tokens the model may spend on internal reasoning before answering. This makes cost/quality trade-offs explicit and debuggable.
Thought summaries expose a compressed version of the reasoning trace — valuable for enterprise auditability without exposing full internal chains.
Technical Details
- 1,048,576 input tokens: enables loading entire codebases, legal corpora, or book-length documents in a single prompt.
- Deep Think: routes complex requests to a reasoning mode analogous to o1 — math, code, planning tasks. Light requests bypass it.
- Multimodal: text, code, images, audio, video in the same context.
- Knowledge cutoff: January 2025 (at GA).
{
"model": "gemini-2.5-pro",
"thinking_config": {
"thinking_budget": 8000,
"include_thoughts": false
},
"contents": [...]
}
Practical Implications¶
Use thinking_budget when you want explicit control over internal reasoning depth versus models with opaque reasoning. Bill and forecast cost using prefill + thinking + output tokens. Use thought summaries for enterprise auditability without exposing full internal chains. The 1,048,576-token context window enables whole codebases, legal corpora, or book-length documents in a single prompt.
Interview Questions
- How does exposing
thinking_budgetas an API parameter change the cost/quality trade-off compared to a model with a fixed reasoning depth? - Why does extending context from 128K to 1M tokens create new serving challenges even if the model quality holds — what tensor shapes and memory layouts are affected?
- What is the "lost-in-the-middle" problem, and how does it interact with 1M-token contexts?