The two largest Gemma 4 models – 26B Mixture of Experts and 31B Dense – require an 80GB Nvidia H100 GPU to run unquantized in bfloat16 format. Google claims these models deliver “frontier intelligence on personal computers” for students, researchers, and developers, providing advanced reasoning capabilities for IDEs, coding assistants, and agentic workflows.
