Article cover image

Anthropic shipped March update that introduced a noticeable regression - a drop in Claude output quality, widespread complaints on X, Reddit and a recent 800-upvote thread on Hacker News surrounding a blog post critiquing anthropic.

Anthropic by then investigated and fixed the issues. Their post-mortem is a great read, but this one line stood out to me:

*...*one addition to the system prompt caused an outsized effect on intelligence in Claude Code

I'm not piling on - Anthropic probably has better eval infrastructure than 99% of companies. The point is: even with that, a single prompt line caused a production regression nobody caught before it shipped.

Testing Agents is Hard

Modern agent stack is complex - harness code, system prompts, tool schemas, context engineering, short and long-term memory, caching - all layered on top of the underlying model. The failure surface is large, and a change in one layer can silently break another. A prompt tweak that improves one task can regress next.

Most teams I talk to don't have any systematic way to catch that. Some run ad hoc scripts. Some lean on generic LLM eval tools that weren't built with agents in mind. Plenty are still doing it manually with their fingers crossed.

And honestly, even teams trying to do it right; struggle because aggregate scores their eval tools emit - hide the failure. You can have score going up yet part of your agent behavior regressing silently. In dev, these failures keep agents in demo rut. In production, Agent failures have business implications.

Robustness through Evaluations

This is the problem we're focused on at HaliosAI - not just telling you that something failed, but which layer of the agent failed and why. That distinction matters a lot when you're trying to actually fix it. In fact we discussed similar regression scenario for AI sales assistant and how we navigated through it on our website:

... We deployed updated candidate prompt and ran it against the same scenario bank. The results initially looked fantastic - the qualitative feel of the conversations improved drastically. 

But when we looked at the structured evaluation metrics, Halios caught a catastrophic regression: among evaluation tasks - search relevance and trajectory validation improved significantly but tool schema validation regressed by 30%. 

Had this been shipped, agent would have missed capturing leads from 30% of conversations - certainly not a good show for a sales assistant. 

- from Lynon agent optimization story on Halios blog (2)

If Agent Reliability is something your team is wrestling with, happy to talk through it. HaliosAI is an operating layer for agent reliability across full agent lifecycle.

My DM's open or email me at first name at halios ai.