Overview
- Gemma 3n is optimized for on-device AI, enabling real-time performance on mobile devices while reducing reliance on cloud processing.
- The model introduces innovations like Per-Layer Embeddings, allowing 5B and 8B parameter models to run with minimal memory usage comparable to smaller models.
- Gemma 3n supports multimodal inputs, including audio, text, images, and video, with enhanced multilingual capabilities for languages such as Japanese, German, and Spanish.
- Developers can access Gemma 3n through Google AI Studio for cloud-based use or Google AI Edge for on-device integration starting today.
- Google collaborated with Qualcomm, MediaTek, and Samsung to optimize Gemma 3n's mobile-first architecture, while also announcing MedGemma for healthcare AI and SignGemma for sign language translation.