
The Apache 2.0 license is an industry standard for open-source software, providing clear, permissive terms that allow users to freely use, modify, and distribute the software for any purpose, including commercial applications, without significant legal overhead or fear of future restrictions ZDNET. This change signals Google's commitment to fostering a truly open ecosystem for its models, directly challenging the perception that its "open" models were still tightly controlled. Developers can now build with Gemma 4, confident in the stability and freedom of its licensing.
These larger models can run unquantized in bfloat16 format on a single Nvidia H100 GPU, though they can also be quantized to run on consumer-grade GPUs. For on-device applications, Google introduced the Effective 2B (E2B) and Effective 4B (E4B) models. These "edge" models are specifically tailored for minimal memory usage and "near-zero latency" on hardware like smartphones, Raspberry Pi, and Jetson Nano.
The new models are also designed for modern AI applications, featuring native function calling, structured JSON output, and integrated instructions for common tools and APIs. Code generation, a critical emerging application, is another area where Gemma 4 excels, offering high-quality results in an offline environment. Furthermore, Gemma 4 supports visual input processing for tasks like OCR (Optical Character Recognition) and chart understanding, and the E2B and E4B models include native speech recognition capabilities. These capabilities are supported across more than 140 languages, with context windows up to 256k tokens for the larger models and 128k for the edge variants.
Developers
Leverage the Apache 2.0 license to integrate advanced local AI capabilities into commercial applications without licensing concerns. Use the E2B and E4B models for prototyping efficient on-device AI for mobile apps and embedded systems, knowing they are forward-compatible with upcoming Gemini Nano 4 releases. Founders & Startups: Build innovative, privacy-focused products that perform complex AI tasks offline. The improved reasoning, code generation, and agentic workflow support in Gemma 4 can power new business models in areas like secure enterprise tools or local content creation. Researchers: Access state-of-the-art models for experimentation and fine-tuning. The public availability of model weights on Hugging Face, Kaggle, and Ollama provides a powerful foundation for advancing local AI research. Research Sources
Google Gemma 4 represents the company's next generation of open-weight AI models, now released under the highly permissive Apache 2.0 license. This strategic shift unlocks powerful local AI capabilities for developers, enabling use across a wide range of devices from mobile to high-end workstations.
The Apache 2.0 license is crucial because it provides clear, permissive terms, allowing developers to freely use, modify, and distribute Gemma 4 for any purpose, including commercial applications. This change removes previous restrictive licensing concerns, fostering broader adoption and innovation within the open-source community.
Google Gemma 4 offers four distinct variants: Effective 2B (E2B) and 4B (E4B) are optimized for mobile and edge devices, providing near-zero latency. The larger 26B Mixture-of-Experts (MoE) and 31B Dense models are designed for developer hardware, excelling in enhanced reasoning, math, and code generation.
Google Gemma 4 models provide enhanced reasoning, mathematics, and code generation, built upon the same research as Gemini 3. They also support advanced functionalities like agentic workflows and native function calling, with the 31B Dense model demonstrating high performance on AI text leaderboards.
More insights on trending topics and technology







