Anthropic’s Claude Code pushes into regulated enterprises as devs demand more agent transparency
Anthropic is expanding Claude Code from internal-heavy code generation to regulated enterprise use while shipping updates and fielding developer concerns about opaque agent behavior. Anthropic says its AI systems now generate nearly all of the company’s internal code, reframing engineers’ roles toward system design and review as described in this report from Moneycontrol ([source](https://www.moneycontrol.com/news/business/information-technology/why-anthropic-says-engineers-matter-more-than-ever-even-as-ai-writes-the-code-13830811.html)). Building on that, Anthropic announced a collaboration with Infosys to deliver agentic AI for telecom, financial services, and manufacturing via Infosys Topaz and the Claude Agent SDK, targeting persistent, multi-step workflows with governance needs ([announcement](https://www.anthropic.com/news/anthropic-infosys)). AWS also outlined how to run Claude Code in compliance-sensitive environments on Amazon Bedrock, aimed at aligning AI-assisted dev work with strict controls ([AWS blog](https://aws.amazon.com/blogs/machine-learning/supercharge-regulated-workloads-with-claude-code-and-amazon-bedrock/)). On the ground, developers called out visibility gaps around what agents do to their codebases in a widely discussed Hacker News thread ([discussion](https://news.ycombinator.com/item?id=47033622)), even as Anthropic continues frequent incremental fixes such as auth refresh repairs and improved error messaging in recent Claude Code releases ([release notes](https://github.com/anthropics/claude-code/releases)). Community demos show evolving workflows—like Plan Mode and multi-agent patterns in Opus 4.6—that promise more autonomous execution but heighten the need for auditability ([Plan Mode walkthrough](https://www.youtube.com/watch?v=fxj82iBWypA&pp=ygUSQ2xhdWRlIENvZGUgdXBkYXRl), [Agent Teams demo](https://www.youtube.com/watch?v=6UKUQNcRk2k&pp=ygUYQUkgY29kaW5nIGFnZW50IHdvcmtmbG93)).