Gemini 2.5 Pro Experimental is our most advanced coding model yet and is state-of-the-art across a range of benchmarks requiring enhanced reasoning.
Pro performance
-
Enhanced reasoning
State-of-the-art in key math and science benchmarks.
-
Advanced coding
Easily generate code for web development tasks.
-
Natively multimodal
Understands input across text, audio, images and video.
-
Long context
Explore vast datasets with a 1-million token context window.
Hands-on with 2.5 Pro
See how Gemini 2.5 Pro Experimental uses its reasoning capabilities to create interactive simulations and do advanced coding.
Benchmarks
Gemini 2.5 Pro leads common benchmarks by meaningful margins.
Benchmark |
Gemini 2.5 Pro
Experimental (03-25)
|
OpenAI o3-mini
High
|
OpenAI GPT-4.5
|
Claude 3.7 Sonnet
64k Extended thinking
|
Grok 3 Beta
Extended thinking
|
DeepSeek R1
|
|
---|---|---|---|---|---|---|---|
Reasoning & knowledge
Humanity's Last Exam (no tools)
|
18.8% | 14.0%* | 6.4% | 8.9% | — | 8.6%* | |
Science
GPQA diamond
|
single attempt (pass@1) | 84.0% | 79.7% | 71.4% | 78.2% | 80.2% | 71.5% |
|
multiple attempts | — | — | — | 84.8% | 84.6% | — |
Mathematics
AIME 2025
|
single attempt (pass@1) | 86.7% | 86.5% | — | 49.5% | 77.3% | 70.0% |
|
multiple attempts | — | — | — | — | 93.3% | — |
Mathematics
AIME 2024
|
single attempt (pass@1) | 92.0% | 87.3% | 36.7% | 61.3% | 83.9% | 79.8% |
|
multiple attempts | — | — | — | 80.0% | 93.3% | — |
Code generation
LiveCodeBench v5
|
single attempt (pass@1) | 70.4% | 74.1% | — | — | 70.6% | 64.3% |
|
multiple attempts | — | — | — | — | 79.4% | — |
Code editing
Aider Polyglot
|
74.0% / 68.6%
whole / diff
|
60.4%
diff
|
44.9%
diff
|
64.9%
diff
|
— |
56.9%
diff
|
|
Agentic coding
SWE-bench Verified
|
63.8% | 49.3% | 38.0% | 70.3% | — | 49.2% | |
Factuality
SimpleQA
|
52.9% | 13.8% | 62.5% | — | 43.6% | 30.1% | |
Visual reasoning
MMMU
|
single attempt (pass@1) | 81.7% | no MM support | 74.4% | 75.0% | 76.0% | no MM support |
|
multiple attempts | — | no MM support | — | — | 78.0% | no MM support |
Image understanding
Vibe-Eval (Reka)
|
69.4% | no MM support | — | — | — | no MM support | |
Long context
MRCR
|
128k (average) | 94.5% | 61.4% | 64.0% | — | — | — |
|
1M (pointwise) | 83.1% | — | — | — | — | — |
Multilingual performance
Global MMLU (Lite)
|
89.8% | — | — | — | — | — |
Model information
Model deployment status | Experimental |
Supported data types for input | Text, Image, Video, Audio |
Supported data types for output | Text |
Supported # tokens for input | 1M |
Supported # tokens for output | 64k |
Knowledge cutoff | January 2025 |
Tool use |
Function calling Structured output Search as a tool Code execution |
Best for |
Reasoning Coding Complex prompts |
Availability |
Google AI Studio Gemini API Gemini App |