Google’s Gemma 3 rivals DeepSeek R1 with 98% accuracy—on just one GPU

Written by

Published 14 Mar 2025

Fact checked by

NSFW AI Why trust Greenbot

We maintain a strict editorial policy dedicated to factual accuracy, relevance, and impartiality. Our content is written and edited by top industry professionals with first-hand experience. The content undergoes thorough review by experienced editors to guarantee and adherence to the highest standards of reporting and publishing.

Disclosure

 

googles gemma3 rivals deepseek r1

Source: Google for Developers

Google launched Gemma 3, a new artificial intelligence (AI) model achieving 98% of DeepSeek R1‘s accuracy while using just one NVIDIA H100 GPU—compared to R1’s estimated 32 GPUs.

This leap in efficiency could make high-performance AI far more accessible.

    In Chatbot Arena Elo ratings, Gemma 3 scored 1338 compared to DeepSeek R1’s 1363. While R1 maintains a slight lead, Google’s achievement focuses on delivering similar results with far less computing power.

    google-2025-gemma-3-elo-comparison

    Source: Google

    “Gemma 3 delivers state-of-the-art performance for its size, outperforming Llama-405B, DeepSeek-V3, and o3-mini in preliminary human preference evaluations,” Google stated in its developer blog. “This helps you to create engaging user experiences that can fit on a single GPU or TPU host.”

    The model features four sizes (1B, 4B, 12B, and 27B parameters) built to run on devices from phones to workstations. Even its largest version is dramatically smaller than DeepSeek R1’s 671 billion parameters.

    Gemma 3 provides “out-of-the-box” support for 35 languages with pre-trained capabilities for 105 more. This represents a dramatic expansion from Gemma 2, which primarily supported English.

    google-2025-gemma-3-and-gemma-2-comparison

    Source: Google

    Beyond language support, Gemma 3 brings multimodal capabilities. This means the AI can understand and work with words, pictures, and videos all at once, rather than just processing text. Its expanded 128K token context window processes approximately 300 pages of text or 30 high-resolution images at once.

    These efficiency improvements come from model distillation and training techniques like Reinforcement Learning from Human Feedback (RLHF) and Reinforcement Learning from Machine Feedback (RLMF). Full technical details are available in Google’s 26-page technical report on the Gemma website.

    Google paired the release with ShieldGemma 2, a specialized 4B parameter model for content safety. The company’s technical report notes, “We find that Gemma 3 models memorize long-form text at a much lower rate than prior models,” suggesting better data privacy than earlier versions.

    The push for AI efficiency has grown stronger with Microsoft’s Phi-4 and Mistral Small 3 joining the competition. These smaller models target businesses that need AI capabilities without the massive computing resources of data centers. Running advanced AI on a single GPU dramatically reduces the entry barriers for smaller organizations.

    Developers can access Gemma 3 through Google AI Studio, Hugging Face, and Kaggle. Academic researchers can apply for Google’s program, which offers $10,000 in cloud credits for Gemma 3 research.

    While Google calls Gemma “open,” some debate exists around its licensing terms restricting certain uses. Still, the model represents progress in making advanced AI work with reasonable computing resources.