Skip to main content
Claudeprompt-engineeringcoding-agents

CLAUDE.MD: Making Claude Try Harder

A practical trick has emerged for prompting Claude: explicitly stating that its output will be audited by Codex encourages the model to be more thorough and avoid shortcuts. This is a significant insight for AI automation, as this small change can noticeably improve the quality of coding agents without needing a model upgrade.

Technical Context

I wasn't hooked by the joke about "Claude will try harder," but by the mechanism itself. A simple frame is added to CLAUDE.MD: your result will be checked by Codex. And suddenly, Claude starts cutting corners less often, skips fewer testing steps, and is more careful about completing the task.

From an engineering perspective, this is very recognizable. I constantly see in AI implementation that a model's behavior changes not just from an instruction like "do better," but from its role, the associated risks, and the expectation of a review. When an agent understands that its output isn't final and will go to an external audit, it more often chooses a more conservative path.

In the original discussion, an important detail was that this all worked in a "clean context," but the author also had a custom adapter. And this is where I'd put a big asterisk. This isn't news on the level of "Anthropic officially rolled out a new feature," but rather a field insight born from a real-world setup around the model.

So, our primary fact doesn't come from Anthropic's documentation or a release note. It's a user's observation of a coding agent's behavior within their own experimental matrix. It's April 2026, and I would honestly present this not as a proven law, but as a strong hypothesis worth testing in your own environment.

Why might this trick work at all? Because LLMs have a good sense of a task's social structure. If I write in the system prompt not just "do a quality job," but "your code will be reviewed, errors will surface, and your solution will be compared to an alternative," I create pressure for completeness and self-checking.

And yes, the mention of Codex here isn't magical. It's highly likely that the brand isn't what's working, but the very fact of an external auditor. Today it's Codex, tomorrow another agent, the day after an internal review bot. The essence is that the model gains a context of accountability.

What This Changes for Business and Automation

The most interesting part begins not in the prompts, but in the architecture. If I'm building AI automation for development, support, or QA, it's no longer enough to just pick the "best model." I think in layers: who generates, who reviews, who debates, who finalizes with tests.

This is where small tricks like this are actually worth money. One paragraph in a system prompt can eliminate some lazy responses without a plan upgrade or complex fine-tuning. For a business, this is often a better deal than immediately throwing more tokens and more agents at the problem.

But there's a downside. If you overdo it with "you'll be checked, you'll be compared, don't make mistakes," the agent might become slower, more cautious, and start over-explaining the obvious. I've broken this many times: quality seems to have improved, but throughput drops so much that the entire automation deflates.

Therefore, I would only test this pattern in measurable scenarios. For example: the percentage of tasks where Claude actually runs or suggests tests; the proportion of missed edge cases; the number of iterations to a working PR. Without metrics, this quickly turns into a nice legend about prompt engineering.

Who wins? Teams that already have coding agents integrated into their pipeline and have their own custom wrapper. Who loses? Those who expect one phrase in CLAUDE.MD to suddenly replace proper AI integration, validation, and task routing between agents.

At Nahornyi AI Lab, this is exactly how we break things down into layers: where an external critic is needed, where a self-check is sufficient, and where it's better not to let the model run autonomously at all. This is no longer about a magic prompt, but about AI solutions architecture, where an agent's behavior is based on roles, checks, and the cost of an error.

If your coding agent writes code but sometimes "cuts corners" in the most critical places, I would start with this external audit frame and A/B test it on your tasks. And if you want to stop guessing and build a working AI automation system for your process, at Nahornyi AI Lab I can help turn these experiments into a system that actually eliminates routine work, not creates another layer of chaos.

Share this article