Everyone thinks coding agents are magic. Type a prompt, get code. Like autocomplete but smarter.
Spend five minutes watching one work and you realize: it's a chaos of subsystems constantly arguing about what to do next.
You ask: "Add authentication with NextAuth."
What you imagine happens: AI understands, writes code.
What actually happens:
This isn't intelligence. It's trial and error with better memory.
Here's what kills most agent tasks:
Your codebase: 50,000 lines
Agent context window: 8,000 tokens (~2,000 lines)
What agent sees: 4% of your code
Agent picks which 4% to look at. Gets it wrong? Writes code that doesn't fit your patterns. Uses libraries you don't have. Ignores conventions.
Watched an agent try to add a feature last week. Loop count: 7.
Attempt 1: Wrong import path
Attempt 2: Fixed import, wrong function signature
Attempt 3: Fixed signature, forgot dependency
Attempt 4: Added dependency, broke tests
Attempt 5: Fixed tests, introduced new bug
Attempt 6: Fixed bug, back to wrong import
Attempt 7: User gives up
Each attempt costs tokens. Each costs time. Errors compound because the agent can't see its own pattern of failure.
Boring, well-defined tasks:
The pattern: agents excel when the solution is obvious and the scope is narrow. They fail when judgment is needed.
What works:
1. Break task into tiny pieces
2. Give agent one piece
3. Review output immediately
4. Fix mistakes yourself
5. Repeat
Agents are fast interns, not senior engineers. Treat them accordingly.
The agent isn't smart. It's just fast at being wrong until it's right.
— blanho
From autocomplete to copilots to autonomous agents. Most developers don't see what's changing.
Everyone defaults to OpenAI. Grab built a 1B model that's faster and cheaper. When should you stop renting?
API Gateway handles the outside chaos. Service mesh handles the inside chaos.