Google DeepMind's Gemma 4 family launches on Hugging Face with Apache 2 licenses, multimodal support (image, text, audio), and sizes ranging from 2B to 27B parameters with up to 128K context windows. Models are available across transformers, llama.cpp, MLX, WebGPU, and Rust, with on-device deployment as a first-class target. Hugging Face reports the models score at the Pareto frontier on arena benchmarks and are strong enough out of the box to make fine-tuning examples hard to find.
Models
Welcome Gemma 4: Frontier multimodal intelligence on device
Google releases Gemma 4, an open-source multimodal family (2B–27B parameters) scoring at the performance frontier while optimized for on-device deployment without fine-tuning needed.
Friday, April 3, 2026 12:00 PM UTC2 MIN READSOURCE: Hugging FaceBY sys://pipeline
Tags
models
/// RELATED
War4d ago
Elon Musk had a bad week in court
Musk's testimony in his lawsuit against OpenAI unraveled in court as he contradicted himself and argued with counsel, potentially crippling his case to reclaim control of the nonprofit.
Products4d ago
Introducing Dynamic Workflows: durable execution that follows the tenant
Cloudflare Workers now supports durable, dynamic code execution on multi-tenant platforms — enabling AI agents and CI/CD systems to safely run versioned, isolated workloads without rebuilds.