In partnership with

I've been testing this setup by myself to see how really measures up in the real-world and if it can actually beat Claude.

I also discovered an even faster way to set things up -- without even needing to permanently switch away from Claude models.

I talk more about it in our community -- where I'm also currently sharing my experiences with Gemma 4 in Claude Code.

The past few days have had so many devs going crazy over Google's new open-source Gemma 4.

And for very good reason -- suddenly so many AI-powered tools like Claude Code have now become FREE and accessible to everyone -- without any compromises in intelligence.

And the best part is it's so ridiculously easy to set up locally -- thanks to ingenious connector tools like Ollama.

Gemma 4 + Ollama + Claude Code.

Ollama exposes an Anthropic-compatible API -- which allows Claude Code to talk to a local model instead of a hosted endpoint.

With Gemma 4 running locally, you get a Claude-style coding workflow without relying on remote inference.

This gives you:

  • local coding model

  • Claude Code terminal workflow

  • no hosted inference calls

  • fast iteration

  • full repo privacy

  • easy model swapping

What more could you even ask for?

1. Get started: Install and Run Gemma 4 with Ollama

Installing or updating Ollama is just too easy:

curl -fsSL https://ollama.com/install.sh | sh

Then pull a Gemma 4 model based on your hardware:

Model sizes to pick from

E2B

  • 2.3B effective (~5.1B w/ embeddings)

  • ~1.7GB download

  • ~1.5–2GB RAM

ollama pull gemma4:e2b

E4B

  • 4.5B effective (~8B w/ embeddings)

  • ~3.2GB download

  • ~3–4GB RAM

ollama pull gemma4:e4b

26B A4B

  • 26B total (4B active)

  • ~17GB download

  • ~18–20GB RAM

ollama pull gemma4:26b

31B Dense

  • 31B

  • ~19GB download

  • ~20-24GB RAM

ollama pull gemma4:31b

Verify the model works:

LLM traffic converts 3× better than Google search

58% of buyers now start their research in ChatGPT or Gemini, not Google. Most startups aren't showing up there yet.

The ones that are get cited by the AI tools their buyers, investors, and future hires already use. And they convert at 3×.

Download the free AEO Playbook for Startups from HubSpot and get the exact steps to start showing up. Five minutes to read.

Learn how to code faster with AI in 5 mins a day

You're spending 40 hours a week writing code that AI could do in 10.

While you're grinding through pull requests, 200k+ engineers at OpenAI, Google & Meta are using AI to ship faster.

How?

The Code newsletter teaches them exactly which AI tools to use and how to use them.

Here's what you get:

  • AI coding techniques used by top engineers at top companies in just 5 mins a day

  • Tools and workflows that cut your coding time in half

  • Tech insights that keep you 6 months ahead

Sign up and get access to the Ultimate Claude code guide to ship 5X faster.

Keep Reading