The Launch That Should Have Been Front-Page News
On April 2, while most of the tech press was still writing Liberation Day anniversary retrospectives, Google DeepMind quietly released what might be the most consequential AI model of 2026 [1]. Gemma 4 is a family of open-source AI models released under the Apache 2.0 license — the most permissive open-source license that exists. Anyone can download them, run them, modify them, and build commercial products with them. No API keys. No per-token charges. No terms of service that change quarterly. The family includes four variants: a 2B-parameter model small enough for IoT devices, a 4B model for phones, a 26B model that uses a mixture-of-experts architecture for efficiency, and a 31B dense model that represents the full capability ceiling [1]. All of them support 256K token context windows, process text, images, and audio natively, and work in over 140 languages. That spec sheet alone would be noteworthy. What makes Gemma 4 genuinely important is the benchmark performance — numbers that put it in direct competition with models costing tens of millions of dollars to access at scale.
The Numbers That Matter
The 31B model scores 89.2% on AIME 2026, 80% on LiveCodeBench v6, and 84.3% on GPQA Diamond [1]. For context, those are numbers that would have been a closed-model flex not very long ago. The 26B mixture-of-experts variant scores 88.3% on AIME and 77.1% on LiveCodeBench while using significantly less compute per inference than the dense 31B model [1]. That matters because it suggests you can get most of the capability at a much lower hardware cost. The most striking metric, though, may be the agentic one: 86.4% on τ2-bench, a benchmark measuring an AI model's ability to use tools autonomously and complete multi-step workflows [1]. That is not a small incremental upgrade over the previous generation. It's a leap. And because it's available under an open license, that leap belongs to anybody willing to download the weights and start building.

