Today's AI/ML headlines are brought to you by ThreatPerspective

Digital Event Horizon

Gemma 3n: A Revolutionary Breakthrough in Multimodal AI Models



Gemma 3n, a revolutionary multimodal AI model, is now available on various open-source libraries. This game-changing breakthrough has significant implications for industries worldwide, enabling diverse applications across natural language processing and computer vision. Learn more about Gemma 3n's features, availability, and potential applications in this exciting new development.

  • Gemma 3n is a cutting-edge multimodal AI model announced at Google I/O.
  • The model supports image, text, audio, and video inputs and has been made available on open-source libraries.
  • Gemma 3n comes in two model sizes: gemma-3n-E2B and gemma-3n-E4B with comparable performance to larger models.
  • The model leverages advanced architectures like MatFormer and Per-Layer Embeddings for improved memory efficiency.
  • Gemma 3n has been integrated into various open-source libraries, including transformers, timm, and Google AI Edge.
  • A special notebook allows users to fine-tune the model on free Google Colab, while a repository offers additional resources and scripts for running models and fine-tuning them.



  • Gemma 3n, a cutting-edge multimodal AI model announced at Google I/O, has finally been made available on the most-used open-source libraries. This monumental achievement marks a significant milestone in the field of natural language processing and computer vision, paving the way for a new era of innovation and collaboration.

    Developed from scratch to run locally on hardware, Gemma 3n is an natively multimodal model that supports image, text, audio, and video inputs. Its groundbreaking design enables it to process various modalities seamlessly, making it an ideal solution for diverse applications across industries.

    The release of Gemma 3n comes with two model sizes: gemma-3n-E2B and gemma-3n-E4B. Despite their smaller parameter counts, these models boast impressive performance comparable to larger models. The E2B variant requires only 2GB of GPU RAM, while the E4B variant can run with just 3GB.

    Gemma 3n leverages advanced architectures such as MatFormer and Per-Layer Embeddings (PLE), which have been added to transformers and timm. These innovations have significantly improved memory efficiency, enabling the model to operate within restricted hardware constraints.

    The vision encoder in Gemma 3n is powered by a new version of MobileNet-v5-300, which boasts an impressive 60 FPS on Google Pixel devices while using only 3x fewer parameters than its predecessor, ViT Giant. The audio encoder is based on the Universal Speech Model (USM) and processes audio inputs in 160ms chunks.

    Gemma 3n has been integrated into various open-source libraries, including transformers, timm, MLX, llama.cpp (text inputs), transformers.js, ollama, Google AI Edge, and others. This widespread availability ensures that developers from diverse backgrounds can harness the power of Gemma 3n for their projects.

    The model's architecture has been added to the new version of transformers released today. Users can explore this implementation through demos, fine-tune the model for specific downstream tasks across modalities, or experiment with other libraries like MLX and llama.cpp.

    To celebrate the release of Gemma 3n, Hugging Face has introduced a special notebook that allows users to fine-tune the model on free Google Colab. This innovative feature empowers developers to adapt the model to their speech datasets and benchmarks in an effortless manner.

    Moreover, Hugging Face has launched the Hugging Face Gemma Recipes repository, where users can discover notebooks and scripts for running models and fine-tuning them. The community is encouraged to contribute more recipes, making the platform even more comprehensive and accessible.

    The release of Gemma 3n marks a significant breakthrough in multimodal AI research. Its availability on open-source libraries has opened up new possibilities for developers worldwide. As researchers and engineers continue to push the boundaries of this technology, we can expect even more groundbreaking innovations in the years to come.




    Related Information:
  • https://www.digitaleventhorizon.com/articles/Gemma-3n-A-Revolutionary-Breakthrough-in-Multimodal-AI-Models-deh.shtml

  • https://huggingface.co/blog/gemma3n


  • Published: Thu Jun 26 15:30:31 2025 by llama3.2 3B Q4_K_M











    © Digital Event Horizon . All rights reserved.

    Privacy | Terms of Use | Contact Us