Forem

# llm

Posts

đź‘‹ Sign in for the ability to sort posts by relevant, latest, or top.
Compilation for LLMs: Why a Language for Models Needs Native Code
Cover image for Compilation for LLMs: Why a Language for Models Needs Native Code

Compilation for LLMs: Why a Language for Models Needs Native Code

Comments
4 min read
Why RAG Falls Short for Documentation Search (and What to Try Instead)
Cover image for Why RAG Falls Short for Documentation Search (and What to Try Instead)

Why RAG Falls Short for Documentation Search (and What to Try Instead)

1
Comments
5 min read
Show HN: Isartor – Pure-Rust prompt firewall, deflects 60-95% of LLM traffic

Show HN: Isartor – Pure-Rust prompt firewall, deflects 60-95% of LLM traffic

Comments
1 min read
Your Knowledge, Your Model — Part 2: Agents, Iatrogenics
Cover image for Your Knowledge, Your Model — Part 2: Agents, Iatrogenics

Your Knowledge, Your Model — Part 2: Agents, Iatrogenics

Comments
4 min read
Remote Ollama access via Tailscale or WireGuard, no public ports

Remote Ollama access via Tailscale or WireGuard, no public ports

Comments
9 min read
The Tool Calling Problem: Why Most Agents Are Just Chatbots with Buttons

The Tool Calling Problem: Why Most Agents Are Just Chatbots with Buttons

Comments
2 min read
The Real Cost: Token Savings Calculator for Engineering Teams
Cover image for The Real Cost: Token Savings Calculator for Engineering Teams

The Real Cost: Token Savings Calculator for Engineering Teams

Comments
2 min read
JIT vs Interpreters: Benchmarking LLM-Generated Code Execution
Cover image for JIT vs Interpreters: Benchmarking LLM-Generated Code Execution

JIT vs Interpreters: Benchmarking LLM-Generated Code Execution

Comments
3 min read
Building Production-Ready Agentic AI Systems for Enterprise Software Delivery
Cover image for Building Production-Ready Agentic AI Systems for Enterprise Software Delivery

Building Production-Ready Agentic AI Systems for Enterprise Software Delivery

Comments 1
3 min read
Hindley-Milner for LLMs: Type Inference Without Annotations
Cover image for Hindley-Milner for LLMs: Type Inference Without Annotations

Hindley-Milner for LLMs: Type Inference Without Annotations

Comments
2 min read
Prompts Are Infrastructure. Here's What That Actually Means.

Prompts Are Infrastructure. Here's What That Actually Means.

Comments
7 min read
We built an LLM proxy that adds 47ms of latency. Here's every millisecond accounted for.
Cover image for We built an LLM proxy that adds 47ms of latency. Here's every millisecond accounted for.

We built an LLM proxy that adds 47ms of latency. Here's every millisecond accounted for.

1
Comments
8 min read
Claude Feels Slow. But Is Moving a Team to Open-Weight Models Actually the Fix?
Cover image for Claude Feels Slow. But Is Moving a Team to Open-Weight Models Actually the Fix?

Claude Feels Slow. But Is Moving a Team to Open-Weight Models Actually the Fix?

Comments
5 min read
The Billion Dollar While Loop: Emergent Architecture in the Agentic AI Era

The Billion Dollar While Loop: Emergent Architecture in the Agentic AI Era

Comments
7 min read
Closing the knowledge gap with agent skills

Closing the knowledge gap with agent skills

Comments 1
4 min read
đź‘‹ Sign in for the ability to sort posts by relevant, latest, or top.