Forem

# llm

Posts

👋 Sign in for the ability to sort posts by relevant, latest, or top.
Built a Predictive Incident Response Agent with LLMs and Vector Memory

Built a Predictive Incident Response Agent with LLMs and Vector Memory

Comments
6 min read
Aria: Building an AI Customer Support Agent with Persistent Memory

Aria: Building an AI Customer Support Agent with Persistent Memory

Comments
8 min read
Fixing blind spots in code reviews with Hindsight memory

Fixing blind spots in code reviews with Hindsight memory

Comments
2 min read
MCP Tool Poisoning: The AI Supply Chain Attack Nobody Is Talking About
Cover image for MCP Tool Poisoning: The AI Supply Chain Attack Nobody Is Talking About

MCP Tool Poisoning: The AI Supply Chain Attack Nobody Is Talking About

Comments 2
4 min read
TOON File Format Anatomy: Schema-Once, Data-Many for LLM Pipelines 🎯📄
Cover image for TOON File Format Anatomy: Schema-Once, Data-Many for LLM Pipelines 🎯📄

TOON File Format Anatomy: Schema-Once, Data-Many for LLM Pipelines 🎯📄

Comments
6 min read
Semantic caching thresholds and why they matter

Semantic caching thresholds and why they matter

Comments
12 min read
Five habits that separate the operator from the vibe-coder

Five habits that separate the operator from the vibe-coder

Comments
6 min read
Gate Zero: stop unfalsifiable prompts before they canonicalize as specs

Gate Zero: stop unfalsifiable prompts before they canonicalize as specs

Comments
5 min read
Eval-driven development for a local-LLM agent: how I shipped Lore 0.2.0 with confidence

Eval-driven development for a local-LLM agent: how I shipped Lore 0.2.0 with confidence

1
Comments
6 min read
Qwen 3.6 Ollama Release, Consumer GPU Benchmarks, GGUF Quantization Fixes

Qwen 3.6 Ollama Release, Consumer GPU Benchmarks, GGUF Quantization Fixes

Comments
4 min read
Your Claude Code rules are a liability you'll never audit

Your Claude Code rules are a liability you'll never audit

Comments
6 min read
Traditional Quantization vs 1.58-Bit Ternary Models: A Practical Comparison
Cover image for Traditional Quantization vs 1.58-Bit Ternary Models: A Practical Comparison

Traditional Quantization vs 1.58-Bit Ternary Models: A Practical Comparison

Comments 1
5 min read
How to structure JSON for LLMs (and stop wasting tokens)
Cover image for How to structure JSON for LLMs (and stop wasting tokens)

How to structure JSON for LLMs (and stop wasting tokens)

Comments
4 min read
dealmind
Cover image for dealmind

dealmind

Comments
1 min read
DEALMIND
Cover image for DEALMIND

DEALMIND

Comments
1 min read
👋 Sign in for the ability to sort posts by relevant, latest, or top.