Google’s Gemma 3 LLMs, the new iteration in its already existing portfolio of lightweight Gemma LLMs, come in four variants: 1 billion parameters, 4 billion parameters, 12 billion parameters and 27 billion parameters. These LLMs, Google says, allow AI developers to choose the best model for their available hardware and performance when developing AI applications.
The Gemma 3 LLMs are designed to run quickly and directly on devices ranging from workstations to laptops and smartphones. For example, to run a Gemma 3 LLM on a device with a GPU or a TensorFlow processor, a larger model can be used than for a smartphone processor. The latter then uses a smaller model.
According to the tech giant, the new Gemma 3 LLMs’ performance is very high, surpassing that of OpenAI’s Llama-40B, DeepSeek-V3, and o3-mini, according to benchmarks. However, they still have to beat DeepSeek R1.
I didn’t know about Chatbot arena, turns out its a 100% subjective rating system where visitors are asked compare 2 LLMs and vote for the “best” after giving them one prompt and displaying their output side by side (sending more than 2 prompts crashed the page for me). That’s a crazy metric to include in your promo material imo.