AI5 min readThe Verge AI

Anthropic’s Claude Code gets ‘safer’ auto mode

P
Redakcja Pixelift0 views
Share
Anthropic’s Claude Code gets ‘safer’ auto mode

Foto: The Verge AI

Claude Code has just gained a new "auto mode" feature, designed to be the middle ground between the tedious manual approval of every operation and the risky, full autonomy of artificial intelligence. Anthropic is introducing this solution to enable the AI agent to make decisions at the system permission level on behalf of the user while simultaneously minimizing threats. The tool can independently flag and block potentially dangerous actions—such as deleting key files, sending sensitive data, or executing malicious code—before they are launched. Currently, the feature is available in research preview for Team plan subscribers and will reach Enterprise and API users in the coming days. This is a significant step for so-called "vibe coders" and professional developers who want to accelerate the software development process without fear of accidental destruction of their workspace. However, Anthropic loyally warns that auto mode remains an experiment and does not eliminate risk 100%, therefore recommending testing it in isolated environments. For the global creator community, this signifies a transition from simple code-writing assistants to advanced agents capable of independently correcting their own errors and requesting human intervention only in critical situations. Programming automation is thus entering a phase of controlled independence, where security becomes an integral part of the creative process.

In the world of AI-assisted software engineering, the line between utility and loss of control is becoming increasingly thin. Anthropic, one of the leading players in the language model market, has just taken a significant step toward civilizing autonomous AI agents. The company announced the introduction of the auto mode feature for its Claude Code tool, promising developers a middle ground between the tedious approval of every line of code and the risky surrender of the controls to the machine.

This premiere comes at a time when the tech industry is experimenting more boldly with so-called vibe coding — a work style where the programmer operates at a high level of abstraction, delegating the dirty implementation work to the AI. However, the previous freedom of Claude Code carried real risks, from the accidental deletion of critical files to the unconscious execution of malicious instructions hidden in the source code. The new mode is intended to be a safety fuse that allows the AI to make decisions regarding permissions without constant human supervision, but within strictly defined security frameworks.

Security architecture instead of blind trust

The mechanism behind auto mode is based on intelligent filtering of high-risk actions. Instead of asking the user about every minor change, Claude Code is now able to independently assess whether a given operation falls within safety limits. If the tool encounters a potentially dangerous action — such as an attempt to send sensitive data externally or modify system files — the system automatically flags and blocks it. In such a scenario, the AI agent is given a chance to approach the problem again in a safer way or, as a last resort, requests human intervention.

Claude Code interface during project work
Claude Code is set to become a safer partner in a programmer's daily work thanks to new control mechanisms.

For developers, this means a significant reduction in so-called decision noise. Previous "agentic AI" tools often fell into two extremes: they either required an "Enter" confirmation for every operation, which killed productivity, or they operated in full autonomy mode, which was sometimes compared to letting a digital elephant into a china shop. Anthropic positions its new mode as an intermediate layer designed to protect against common errors, such as the execution of malicious code or the leakage of sensitive information.

Availability and research phase limitations

Despite the optimistic announcements, Anthropic maintains a high degree of caution. Currently, auto mode is available exclusively as a research preview for Team plan users. However, the company announced that in the coming days, access will be extended to Enterprise plan subscribers and users utilizing the API. This gradual rollout suggests that the technology is still under the scrutiny of engineers and requires gathering more data from real production environments.

It is worth emphasizing that the manufacturer does not hide the experimental nature of the tool. The company's official position clearly indicates that auto mode "does not eliminate" risk entirely. It only serves as a protective barrier intended to minimize the effects of potential model hallucinations or incorrect interpretations of commands. Consequently, Anthropic recommends that developers use Claude Code in isolated environments, which is standard practice when testing autonomous AI agents.

AI Logo
New security features in Claude Code are part of a broader trend of raising awareness of the risks associated with autonomous agents.

A pragmatic approach to autonomy

The introduction of auto mode is a clear signal that Anthropic understands real market needs. Programmers are not looking for a magic wand that will write an entire system for them, but a solid assistant that won't destroy their repository during a brief absence from the keyboard. The ability to delegate decision-making at the permissions-level decisions is the key to scaling the work of programming teams, provided these mechanisms are trustworthy.

From an industry perspective, this move places Claude Code in an interesting position relative to competitors such as GitHub Copilot or Cursor. While others focus on the speed of syntax suggestions, Anthropic strikes a chord with security and responsible autonomy. In a world where cyberattacks based on prompt injection into AI models are becoming a real threat, the ability to block risky actions before they are executed may turn out to be the strongest selling point.

  • Safer autonomy: The system automatically flags and blocks risky operations on files and data.
  • Reduction of micromanagement: The programmer does not have to approve every minor decision of the agent.
  • Availability: Currently for the Team plan, soon for Enterprise and API.
  • Recommendation: Use in isolated environments is recommended due to experimental status.

The direction taken by the creators of Claude Code suggests that the era of "wild" AI autonomy in programming is coming to an end. Instead of giving models unlimited power over operating systems, market leaders will strive to create control systems that run in the background, allowing for fluid work without compromising project integrity. The success of auto mode will depend on how precisely Anthropic can calibrate the sensitivity of its security filters — too restrictive and they will paralyze work, too loose and they will remain merely a facade of protection.

The introduction of this feature is evidence of Anthropic's technological maturity. The company is not just racing on model parameters but building an ecosystem of tools that have a chance to survive in rigorous corporate conditions. For vibe coders, this is good news: their "vibes" will now have a more solid digital foundation, protecting them from the consequences of their own inattention or model errors.

Source: The Verge AI
Share

Comments

Loading...