IBL News | New York
Google introduced an advanced open-source model, Gemma 4, under a commercially permissive Apache 2.0 license. This model has been built for advanced reasoning and agentic workflows, according to the company.
Gemma 4 was released in four sizes: Effective 2B (E2B), Effective 4B (E4B), 26B Mixture of Experts (MoE), and 31B Dense.
The 31B model currently ranks #3 among open models on the industry-standard Arena AI text leaderboard, and the 26B model is securing the #6 spot.
Gemma 4 models specifically can run on Android devices and laptop GPUs.
Agentic workflows feature native support for function calling, structured JSON output, and system instructions. It enables the building of autonomous agents that can interact with various tools and APIs and execute workflows reliably.
Gemma 4 supports high-quality offline code generation, turning the user’s workstation into a local-first AI code assistant.
The models natively process video and images, support variable resolutions, and excel at visual tasks such as OCR and chart understanding. Additionally, the E2B and E4B models feature native audio input for speech recognition and understanding.
Gemma 4 presents a longer context. The edge models feature a 128K context window, while the larger models offer up to 256K, allowing users to pass repositories or long documents in a single prompt.
According to Google, these multimodal models run completely offline with near-zero latency across edge devices like phones, Raspberry Pi, NVIDIA, and Jetson Orin Nano.
Android developers can now prototype agentic flows in the AICore Developer Preview for forward compatibility with Gemini Nano 4.
Google is offering Gemma 4 in Google AI Studio (31B and 26B MoE) or in Google AI Edge Gallery (E4B and E2B).
First-generation Gemma models have registered over 400 million downloads, generating a Gemmaverse of more than 100,000 variants.
