Advertisement Gemma 3n Preview: powerful, efficient, mobile AI

After exciting premieres Gemma 3 AND Gemma 3 KatOur family of the most modern open models, which can work on one acceleratories of clouds or desktop computers, we fuel our vision of available artificial intelligence even more. Gemma 3 has provided powerful opportunities for programmers, and now we are expanding this vision to highly talented artificial intelligence in real time operating directly on devices that you use every day-telephones, tablets and laptops.

To power the next generation of artificial intelligence and support various applications, including the development of Gemini Nano's capabilities, we designed a new, most modern architecture. This new generation foundation was created in close cooperation with mobile equipment leaders, such as Qualcomm Technologies, MediaTek and Samsung Lsi Business system, and is optimized at the angle of lightning, multimodal artificial intelligence, enabling truly personal and private experience directly on your device.

Gemma 3N is our first open model built on this groundbreaking, common architecture, enabling programmers to start experiments with this technology today in the early preview. The same advanced architecture also supplies the next generation of Gemini Nano, which leads these possibilities to a wide range of functions on Google Apps and our ecosystem on the device and will become available later this year. Gemma 3N allows you to start building this foundation that will appear on the main platforms such as Android and Chrome.

This chart occupies AI models according to the results of Elo Chatbot Arena; Higher results (the highest numbers) indicate greater user preferences. Gemma 3N is high among popular and open models.

GEMMA 3N uses Google Deepmind innovation called Embeddings per laystander (Ple), which ensures a significant reduction in RAM consumption. While the number of RAW parameters is 5b and 8b, this innovation allows you to launch larger models on mobile devices or live from the cloud, with digging memory comparable to the 2B and 4b model, which means that models can work with a dynamic memory trace of only 2 GB and 3GB. Learn more in ours documentation.

By studying Gemma 3N, programmers can get an early view of the basic possibilities of the open model and mobile architectural innovations that will be available on Android and Chrome from Gemini Nano.

In this post, we will examine the new capabilities of Gemma 3N, our approach to responsible development and how to access today.


Key capabilities of Gemmy 3N

Designed for fast, low artificial intelligence AI, Gemma 3N provides:

  • Optimized performance and performance on the device: Gemma 3N begins to react about 1.5x faster on mobile devices of much better quality (compared to Gemma 3 4B) and a reduced memory trace achieved thanks to innovations, such as layer deposition, KVC sharing and advanced acantization of activation.
  • Much in 1 flexibility: Model with an active 4B memory trace, which natively contains the most modern submodel of the 2B memory trace (thanks Form of food training). This ensures the flexibility of dynamic performance and quality exchange in flight without hosting separate models. In addition, we introduce the possibilities of Mix'n'Match in Gemma 3N to dynamically create submodels from the 4B model, which can optimally adapt to a specific use case – and a related quality/delay compromise. Stay up to date to get more research in our upcoming technical report.
  • Ready and offline privacy Local performance enables reliable functions that estimate the privacy and function of the user, even without an internet connection.
  • Extended multimodal understanding with sound: Gemma 3N can understand and process audio, text and images and offers a much better understanding of video. Its audio capabilities allow the model to perform high quality automatic speech recognition (transcription) and translation (speech for translated text). In addition, the model accepts the interspers of input data between the methods, enabling understanding of complex multimodal interactions. (Soon public implementation)
  • Improved multilingual possibilities: Improved multilingual results, especially in Japanese, German, Korean, Spanish and French. Expensive performance reflected on multilingual comparative tests, such as 50.1% in WMT24 ++ (ChRF).

MMLU performance

MMLU performance compared to the size of the MMLU model compared to the possibilities of MMLU-N-Match (Pretraced).

Unlocking new traffic experiences

Gemma 3N will strengthen the new wave of intelligent applications in motion, enabling programmers:

  1. Build vivid, interactive experiences These understand and respond to visual and auditory tips in real time from the user's environment.


2. Power deeper understanding and generating contextual text using combined audio input data, image, video and text-all processed privately to the device.


3. Develop advanced applications focused on audioincluding real -time speech transcription, translation and rich interactions based on the voice.

Here is a review and types of experiences that you can build:

Building responsibly, together

Our commitment to the responsible development of AI is the most important. Gemma 3N, like all Gemma models, has undergone strict security assessments, data management and tuning adaptation to our safety rules. We approach open models with a thorough risk assessment, constantly improving our practices as the AI ​​landscape evolutions.


Start: Gemma 3N preview today

We are excited that Gemma 3N in our hands through a preview starting today:


Initial access (available now):

  • Exploration in a cloud with Google AI Studio: Try Gemma 3N directly in the browser Google to learn – No configuration. Immediately explore the possibilities of entering the text.
  • Development on a device with Google AI Edge: For programmers who want to integrate Gemma 3N locally, Google Ai Edge Provides tools and libraries. You can start with the possibility of understanding the text and image/generation.

Gemma 3N means another step in the democratization of access to the latest, efficient artificial intelligence. We are incredibly excited, seeing what you will build when we provide this technology, starting with today's preview.

Examine this advertisement and all updates We/W Google IO.GOogle From May 22.

LEAVE A REPLY

Please enter your comment!
Please enter your name here