Google Unleashes Gemma 4: Open Models with Gemini 3 Tech
AI News

Google Unleashes Gemma 4: Open Models with Gemini 3 Tech

3 min
4/3/2026
artificial intelligencemachine learningopen sourcegoogle

Google DeepMind Launches Next-Gen Open AI with Gemma 4

Google DeepMind has officially unveiled Gemma 4, the latest generation of its open-weight, state-of-the-art AI model family. Announced on April 2, 2026, via its official models page, Gemma 4 represents a significant leap, being the first open models built directly from the research and technology underpinning Google's flagship Gemini 3 models. This marks a strategic move to push advanced AI capabilities into the open-source ecosystem and onto local hardware.

A Four-Model Family for Every Scale

The Gemma 4 release is not a single model but a quartet designed for different compute environments. The family consists of two larger, reasoning-focused models and two ultra-efficient variants built for the edge.

  • Gemma 4 31B IT Thinking: The flagship model, optimized for advanced reasoning tasks.
  • Gemma 4 26B A4B IT Thinking: A high-performance alternative balancing size and capability.
  • Gemma 4 E4B IT Thinking: An efficient model for mobile and IoT devices.
  • Gemma 4 E2B IT Thinking: The smallest model, designed for maximum compute and memory efficiency on constrained hardware.

Google emphasizes "intelligence-per-parameter" as a core design goal, aiming to deliver frontier-level capabilities on consumer-grade hardware, from personal computers to Raspberry Pi and Jetson Nano boards.

continue reading below...

Technical Capabilities and Performance Benchmarks

Gemma 4 introduces several advanced features previously reserved for closed, proprietary models. Key capabilities include native support for agentic workflows, enabling autonomous agents that can plan and execute tasks using function calling. The models also boast multimodal reasoning with strong audio and visual understanding, support for 140 languages with cultural context, and extensive fine-tuning support.

Performance metrics released by Google show Gemma 4 models leading in several key benchmarks. The 31B variant scores 1452 on Arena AI (text) as of April 2, 2026, and achieves 85.2% on MMMLU. It also shows dramatic improvements in specialized tasks, scoring 89.2% on AIME 2026 Mathematics and 86.4% on the τ2-bench for agentic tool use, a massive leap from Gemma 3 27B's 6.6%.

Open Ecosystem and Deployment

True to its open-model ethos, Gemma 4 weights are immediately available for download on popular platforms including Hugging Face, Ollama, Kaggle, LM Studio, and Docker. For development and deployment, Google provides integration guides for Jax, Keras, Google AI Edge, Google Kubernetes Engine (GKE), and Vertex AI.

The smaller E2B and E4B models are highlighted for their ability to run completely offline with near-zero latency, opening new possibilities for real-time AI processing on edge devices without cloud dependency. The larger 26B and 31B models are positioned to turn workstations into "local-first AI servers" for developers and researchers.

Safety and Strategic Implications

Google states that Gemma 4 models undergo the same rigorous infrastructure security protocols as its proprietary models, aiming to provide a "trusted, transparent foundation" for enterprises and sovereign organizations. This release signals a continued commitment to the open-source AI community while directly competing with other open model families like Llama and Mistral.

By decoupling advanced AI from cloud-only access and leveraging Gemini 3's architectural advances, Google DeepMind is potentially accelerating the democratization of cutting-edge AI. The focus on edge efficiency and local deployment could reshape how AI is integrated into applications, from personal coding assistants to on-device multimodal agents.