Google Releases Open Source Multimodal AI Model Gemma 3

Top post
Google's Gemma 3: A New Multimodal Open-Source AI Player
Google has introduced Gemma 3, a new generation of its open-source AI models. Gemma 3 extends the existing Gemma family with multimodal capabilities, meaning the models can now process both text and images. The models are available in various sizes, from 1 to 27 billion parameters, making them suitable for different use cases and hardware resources.
A particular focus is on the expansion of the context window. Gemma 3 can work with at least 128,000 tokens, significantly more than many comparable models. This allows for the processing of longer texts and more complex tasks. To manage the increased memory requirements when processing long contexts, the model's architecture has been adjusted. The ratio of local to global attention has been increased, and the span of local attention has been kept short. These modifications reduce the need for KV-cache memory, which would otherwise increase dramatically with long contexts.
The Gemma 3 models were trained using distillation, a technique that transfers knowledge from larger models to smaller ones. This leads to improved performance compared to Gemma 2, both in the pre-trained version and after fine-tuning for specific tasks. A new post-training procedure improves capabilities in mathematics, chat, instruction following, and multilingualism. Google states that Gemma3-4B-IT can compete with Gemma2-27B-IT in benchmarks and Gemma3-27B-IT is comparable to Gemini-1.5-Pro.
Improved Performance and Expanded Functionalities
The multimodal orientation of Gemma 3 opens up new possibilities in AI development. The combination of text and image processing enables applications such as image captioning, visual question-answering systems, and the generation of content based on visual input. The improved performance compared to previous models and the extended context length make Gemma 3 a promising tool for developers and researchers.
For Mindverse, as a provider of AI solutions, Gemma 3 opens up interesting perspectives. Integration into the Mindverse platform could provide users with access to powerful multimodal AI functions. Applications in the areas of content creation, chatbots, knowledge databases, and custom AI solutions are conceivable.
Open Source and the AI Community
The release of Gemma 3 as an open-source model underscores Google's commitment to promoting AI research and development. The open availability of the model allows the community to test, improve, and use Gemma 3 for their own projects. This contributes to the democratization of AI technologies and accelerates innovation in this field.
The release of Gemma 3 is an important step in the development of open-source AI models. The combination of multimodality, improved performance, and long context length makes Gemma 3 a promising tool for a variety of applications. It remains to be seen how the community will embrace the model and what innovative applications will emerge.
Bibliography: - https://arxiv.org/abs/2503.19786 - https://blog.google/technology/developers/gemma-3/ - https://storage.googleapis.com/deepmind-media/gemma/Gemma3Report.pdf - https://huggingface.co/google/gemma-3-27b-it - https://www.linkedin.com/posts/data-science-dojo_gemma-3-technical-report-activity-7305950645147070464-Pgou - https://blog.roboflow.com/gemma-3/ - https://ai.google.dev/gemma/docs/core - https://simonw.substack.com/p/notes-on-googles-gemma-3 - https://medium.com/@ritvik19/papers-explained-329-gemma-3-153803a2c591 - https://build.nvidia.com/google/gemma-3-1b-it/modelcard