Latest Posts

How LLM applications learned to remember

We went from 4K token context windows to virtual memory filesystems in four years. Here's the engineering story of how LLM memory evolved - and what you should actually use today.

The hard part of AI engineering isn't the AI

I run a 19-node LangGraph pipeline serving 20000+ users. I've never written a PyTorch training loop for it. Here's what actually matters - and a 24-week roadmap built around it.

What Happens When You Let an AI Rewrite Its Own Instructions?

Most of us are stuck on the prompt treadmill - manually tweaking instructions that break every time the task shifts. This post lays out an architecture where the AI agent grades its own work, rewrites its own prompts, builds its own tools, and rolls back when things get worse. Every idea is backed by published research. No jargon, just the blueprint.

Today I Learned