Forem

# llm

Posts

đź‘‹ Sign in for the ability to sort posts by relevant, latest, or top.
I built TrustLayer — an open-source trust layer for every AI tool you use

I built TrustLayer — an open-source trust layer for every AI tool you use

Comments
2 min read
I got tired of my agents repeating the same mistakes, so I built a feedback loop for them — here's How it is worked

I got tired of my agents repeating the same mistakes, so I built a feedback loop for them — here's How it is worked

Comments
2 min read
Five Hard Problems in the MCP Ecosystem

Five Hard Problems in the MCP Ecosystem

3
Comments 2
9 min read
Prompt Injection Doesn't Come from Your Users
Cover image for Prompt Injection Doesn't Come from Your Users

Prompt Injection Doesn't Come from Your Users

Comments
10 min read
Ollama, LM Studio, and GPT4All Are All Just llama.cpp — Here's Why Performance Still Differs

Ollama, LM Studio, and GPT4All Are All Just llama.cpp — Here's Why Performance Still Differs

Comments
6 min read
Per-customer cost attribution without a proxy

Per-customer cost attribution without a proxy

Comments
3 min read
Anthropic Just Did Something Unprecedented: They Kept a Model Because It Was Too Good at Hacking

Anthropic Just Did Something Unprecedented: They Kept a Model Because It Was Too Good at Hacking

Comments
3 min read
GLM-5.1: The 754B Open Model That Writes Animated SVG

GLM-5.1: The 754B Open Model That Writes Animated SVG

Comments
1 min read
The Chinese Open-Source Model That Draws Pelicans Better Than GPT-4o

The Chinese Open-Source Model That Draws Pelicans Better Than GPT-4o

Comments
2 min read
LLM-as-Judge: using Claude to review a Gemini agent

LLM-as-Judge: using Claude to review a Gemini agent

Comments
7 min read
TurboQuant: How a Simple Spin Saves Gigabytes of GPU Memory
Cover image for TurboQuant: How a Simple Spin Saves Gigabytes of GPU Memory

TurboQuant: How a Simple Spin Saves Gigabytes of GPU Memory

Comments
6 min read
99.8% of LLM Inference Power Isn't Spent on Computation

99.8% of LLM Inference Power Isn't Spent on Computation

Comments
7 min read
Stop Paying Frontier Prices for Tasks a Local Model Handles Fine

Stop Paying Frontier Prices for Tasks a Local Model Handles Fine

Comments
3 min read
When Your AI Wiki Outgrows the Context Window — A Practical Guide to RAG

When Your AI Wiki Outgrows the Context Window — A Practical Guide to RAG

Comments
6 min read
Building a Voice-Controlled Local AI Agent on a 4GB GPU

Building a Voice-Controlled Local AI Agent on a 4GB GPU

Comments
3 min read
đź‘‹ Sign in for the ability to sort posts by relevant, latest, or top.