Technology

Google unveils open source Gemma 3 model with 128k context window


Join daily and weekly newsletters to obtain the latest updates and exclusive content to cover the leading artificial intelligence in the industry. Learn more


Even as large language models and logic remain common, institutions are increasingly heading to smaller models to operate artificial intelligence with less concerns of energy and cost.

Although some institutions distort larger models for smaller versions, models such as Google continue to issue small language models (SLMS) as an alternative to large LLMS models, which may be assigned to operate without sacrificing performance or accuracy.

With this in mind, Google released its latest small -style version, GEMMA, which features expanded windows, larger parameters, and multi -media thinking capabilities.

GEMMA 3, which has the same processing strength as large GEINII 2.0 models, remains better used by smaller devices such as phones and laptops. The new model has four sizes: 1b, 4B, 12B and 27B.

With a larger context window of 128 kilometers – in contrast to that, GEMMA 2 had a context window of 80k – GMMA 3 you can understand more information and complex requests. Google has updated GEMMA 3 to work in 140 languages, analyze images, texts, short videos and support support for the automation of tasks and the progress of the parents.

Gemma gives a strong performance

To reduce computing costs further, Google has provided quantitative versions of GEMMA. Think of quantitative models as compressed models. This occurs through the process of “reducing the accuracy of numerical values ​​in the weight of the model” without sacrificing accuracy.

Google said that GEMMA 3 “offers the latest performance for its size” and outperformed LLMS such as Llama-405b, Deepseek-V3 and O3-MINI. Gemma 3 2 2b, specifically, ranked second to Deepseek-R1 in Chatbot Arena Elo tests. I have topped the smaller Deepseek model, Deepseek V3, Openai’s O3-MINI, Meta’s Llama-405B and great Mistral.

By selecting the amount of GEMMA 3, users can improve performance, run the form and build applications “can suit the host of GPU processing unit and tensioner treatment unit (TPU).”

GEMMA 3 is integrated with developers tools such as Luging Face Transformers, OLLAMA, Jax, KERAS, PyTorch and others. Users can also access GEMMA 3 through Google Ai Studio, Hugging Face or Kagge. Corporates and developers can request access to GEMMA 3 API through AI Studio.

Gima Shield for Security

Google said she built safety protocols in GEMMA 3, including a safety auditor called ShieldGEMA 2.

“The development of GEMMA 3 guarantees extensive data governance, compatibility with our safety policies through standard assessments and strong measurement,” Google is written in a blog post. “While the comprehensive test of models more capable of reporting our evaluation of the lowest ability to do so, the STEM’s augmented GMMA 3 performance led to specific assessments that focus on the possibility of misuse in creating harmful substances; its results indicate a low level of risk.”

ShieldGEMA 2 is a 4B photo auditor, a parameter based on the Gemma 3 Foundation. Users can customize ShieldGEMA 2 to suit their own needs.

Small models and overflow

Since Google was first released in February 2024, SLMS has witnessed an increase in interest. Other small models such as Microsoft Phi-4 and Mistral Small 3 indicate that institutions want to build applications with strong models such as LLMS, but do not necessarily use the range of what LLM can.

Institutions have also started to turn into smaller versions of LLMS that you prefer through distillation. In order to be clear, GEMMA is not the distillation of Gemini 2.0; Instead, it is trained with the same set of data and architecture. The distilled model learns from a larger model, which is GEMMA.

Organizations are often preferred to suit the usage cases designated for a model. Instead of publishing LLM like O3-MINI or Claude 3.7 Sonnet on a simple symbol editor, a smaller model, whether SLM or a distilled version, can do these tasks easily without outperforming a huge model.



2025-03-12 07:30:00

Related Articles

Back to top button