Why the Open License Matters
Previous versions of Google's open models used a custom license that many developers found overly restrictive, including unilateral update clauses and requirements to enforce Google's policies on Gemma-derived projects Ars Technica. This ambiguity and control mechanism created apprehension, hindering broader adoption and development within the open-source community. The switch to Apache 2.0 eliminates these concerns.The Apache 2.0 license is an industry standard for open-source software, providing clear, permissive terms that allow users to freely use, modify, and distribute the software for any purpose, including commercial applications, without significant legal overhead or fear of future restrictions ZDNET. This change signals Google's commitment to fostering a truly open ecosystem for its models, directly challenging the perception that its "open" models were still tightly controlled. Developers can now build with Gemma 4, confident in the stability and freedom of its licensing.
What's New in Gemma 4
Gemma 4 introduces four distinct variants designed to scale from powerful data centers down to mobile devices. The larger models, a 26B Mixture-of-Experts (MoE) and a 31B Dense model, are optimized for offline use on developer hardware. The 26B MoE model is engineered for speed, activating only 3.8 billion of its 26 billion parameters during inference, leading to higher tokens-per-second than similarly sized models. The 31B Dense model prioritizes quality and is designed for fine-tuning to specific use cases Ars Technica.These larger models can run unquantized in bfloat16 format on a single Nvidia H100 GPU, though they can also be quantized to run on consumer-grade GPUs. For on-device applications, Google introduced the Effective 2B (E2B) and Effective 4B (E4B) models. These "edge" models are specifically tailored for minimal memory usage and "near-zero latency" on hardware like smartphones, Raspberry Pi, and Jetson Nano.
Gemma's Technical Specs and Capabilities
Gemma 4 models are built upon the same underlying research and technology as Google's proprietary Gemini 3 closed models, translating into improved reasoning, mathematical abilities, and instruction-following for the open-source variants Engadget. The 31B Dense model has already secured the third spot on the Arena AI text leaderboard, competing with models significantly larger in parameter count The Next Web. Google claims this represents an "intelligence-per-parameter" breakthrough, making Gemma 4 extremely efficient.The new models are also designed for modern AI applications, featuring native function calling, structured JSON output, and integrated instructions for common tools and APIs. Code generation, a critical emerging application, is another area where Gemma 4 excels, offering high-quality results in an offline environment. Furthermore, Gemma 4 supports visual input processing for tasks like OCR (Optical Character Recognition) and chart understanding, and the E2B and E4B models include native speech recognition capabilities. These capabilities are supported across more than 140 languages, with context windows up to 256k tokens for the larger models and 128k for the edge variants.







