Gemma 3n now fully available in open-source ecosystems — on-device multimodal model
AI Impact Summary
Gemma 3n is now fully available in open-source ecosystems, including transformers, timm, MLX, llama.cpp, transformers.js, ollama, and Google AI Edge, with two variants (gemma-3n-E2B and gemma-3n-E4B) designed for on-device use. The E2B/E4B configurations deliver real parameter counts of 5B and 8B but memory footprints that fit in 2GB and 3GB VRAM, enabling low-resource edge deployments. This opens opportunities for offline multimodal applications and could reduce cloud inference costs, but teams must adapt deployment stacks to leverage these libraries and the on-device pipeline capabilities.
Affected Systems
- Date
- Date not specified
- Change type
- capability
- Severity
- medium