Handle complex tasks
Gemma 3's 128K-token context window lets your applications process and understand vast amounts of information, enabling more sophisticated AI features.
Explore Gemma 3
Gemma 3 is the most capable model that can run on a single GPU or TPU. Efficient on workstations, laptops, and even smartphones, allowing developers to build responsible AI applications at scale.
Gemma 3's 128K-token context window lets your applications process and understand vast amounts of information, enabling more sophisticated AI features.
Unparalleled multilingual capabilities let you communicate effortlessly across countries and cultures. Develop applications that reach a global audience, with support for over 140 languages.
Easily build applications that analyze images, text, and video opening up new possibilities for interactive and intelligent applications.
The MMLU benchmark is a test that measures the breadth of knowledge and problem-solving ability acquired by large language models during pretraining.
Assesses code generation capabilities on real-world coding problems from platforms like LeetCode and Codeforces.
Tests a model's ability to translate natural language questions into complex SQL queries across various domains.
Challenges models with difficult questions written by Ph.D. holders in biology, physics, and chemistry.
Evaluates a model's ability to answer simple, factual questions with short phrases.
Evaluates if LLM responses are factually accurate and detailed enough, based on given input documents.
MATH evaluates a language model's ability to solve complex mathematical word problems, requiring reasoning, multi-step problem-solving, and the understanding of mathematical concepts.
An internal holdout set of competition math problems.
Evaluates multimodal understanding and reasoning across various disciplines requiring college-level knowledge.
Gemma QAT dramatically reduces memory requirements while maintaining high quality. This lets you run powerful models like Gemma 3 27B locally on consumer-grade GPUs like an NVIDIA RTX 3090.
Discover the latest advancements in Gemma, Google's family of lightweight, state-of-the-art open models.
Hear how the Gemma research team unveil the architecture, design principles, and innovations behind Google's family of lightweight, state-of-the-art open models.
Explore the development of intelligent agents using Gemma models, with core components that facilitate agent creation, including capabilities for function calling, planning, and reasoning.
Discover how with Gemma 3, we have tried to push many of the limits of what makes a model usable and practical.
Building multilingual AI applications is crucial for reaching global audiences, and varied language proficiency remains a top developer priority.