Forem

Gemma

Gemma is a collection of lightweight, state-of-the-art open models built from the same technology that powers our Gemini models.

Posts

👋 Sign in for the ability to sort posts by relevant, latest, or top.
We Gave an AI Agent a Long Context Caching Idea. Here's what happened next!
Cover image for We Gave an AI Agent a Long Context Caching Idea. Here's what happened next!

We Gave an AI Agent a Long Context Caching Idea. Here's what happened next!

1
Comments
7 min read
Local AI Node on Xiaomi 12 Pro Needs Cooling, Battery Discipline

Local AI Node on Xiaomi 12 Pro Needs Cooling, Battery Discipline

Comments
7 min read
Gemma 4 and the Architecture of On-Device AI

Gemma 4 and the Architecture of On-Device AI

Comments
3 min read
Gemma 3 on a Raspberry Pi 5: I Benchmarked Google's Open Model on a $80 Computer [2026]

Gemma 3 on a Raspberry Pi 5: I Benchmarked Google's Open Model on a $80 Computer [2026]

Comments
7 min read
Gemma 4 E4B on Mac Mini M2: Real Benchmarks for a Companion Robot

Gemma 4 E4B on Mac Mini M2: Real Benchmarks for a Companion Robot

Comments
4 min read
Mastering Gemma 4: A Comprehensive Deep Dive into Google's Next-Generation Open Model Architecture and Deployment
Cover image for Mastering Gemma 4: A Comprehensive Deep Dive into Google's Next-Generation Open Model Architecture and Deployment

Mastering Gemma 4: A Comprehensive Deep Dive into Google's Next-Generation Open Model Architecture and Deployment

1
Comments
7 min read
Same model. Different results. — AgentKit Benchmark + OpenCode Integration
Cover image for Same model. Different results. — AgentKit Benchmark + OpenCode Integration

Same model. Different results. — AgentKit Benchmark + OpenCode Integration

Comments
2 min read
Gemma 4: Byte for byte, the most capable open models

Gemma 4: Byte for byte, the most capable open models

8
Comments
15 min read
I Ran Google's New Gemma 4 Models Locally (26B and 31B) — Here's What I Found

I Ran Google's New Gemma 4 Models Locally (26B and 31B) — Here's What I Found

Comments
4 min read
Cut AI Costs: Flutter On-Device LLM Integration Works

Cut AI Costs: Flutter On-Device LLM Integration Works

Comments
10 min read
I Ran Google's latest Gemma 4 Models on 48GB GPU. Here's What Actually Happened.
Cover image for I Ran Google's latest Gemma 4 Models on 48GB GPU. Here's What Actually Happened.

I Ran Google's latest Gemma 4 Models on 48GB GPU. Here's What Actually Happened.

1
Comments
6 min read
Gemma on the Edge: Building Private, Low-Cost AI Features for Real Apps
Cover image for Gemma on the Edge: Building Private, Low-Cost AI Features for Real Apps

Gemma on the Edge: Building Private, Low-Cost AI Features for Real Apps

Comments
9 min read
Fine-Tuning Gemma 3 with Cloud Run Jobs: Serverless GPUs (NVIDIA RTX 6000 Pro) for pet breed classification 🐈🐕

Fine-Tuning Gemma 3 with Cloud Run Jobs: Serverless GPUs (NVIDIA RTX 6000 Pro) for pet breed classification 🐈🐕

31
Comments 1
14 min read
Building a Multimodal Local AI Stack: Gemma 4 E2B, vLLM, and Hermes Agent
Cover image for Building a Multimodal Local AI Stack: Gemma 4 E2B, vLLM, and Hermes Agent

Building a Multimodal Local AI Stack: Gemma 4 E2B, vLLM, and Hermes Agent

1
Comments
2 min read
On-Device AI with the Google AI Edge Gallery and Gemma 4
Cover image for On-Device AI with the Google AI Edge Gallery and Gemma 4

On-Device AI with the Google AI Edge Gallery and Gemma 4

28
Comments
6 min read
👋 Sign in for the ability to sort posts by relevant, latest, or top.