Amazon Mandates Senior Engineer Oversight for AI-Coding After Major Outages
AI News

Amazon Mandates Senior Engineer Oversight for AI-Coding After Major Outages

5 min
3/11/2026
AmazonArtificial IntelligenceSoftware DevelopmentCloud Computing

Amazon's AI Coding Crackdown

In a significant shift following a series of high-profile outages, Amazon is tightening internal controls on the use of generative AI coding assistants. According to internal memos and briefings seen by multiple outlets, the company will now require junior and mid-level engineers to get sign-off from senior engineers for any AI-assisted production changes.

The move comes after a "trend of incidents" in recent months, characterized by a "high blast radius" and linked to "Gen-AI assisted changes," as described in a briefing note for a recent "deep dive" meeting. Senior Vice-President Dave Treadwell, overseeing Amazon's e-commerce tech, bluntly told employees, "Folks, as you likely know, the availability of the site and related infrastructure has not been good recently."

The policy is part of a broader initiative to introduce "controlled friction" into change management processes for critical retail systems. Amazon's response highlights a central tension in modern software development: the promise of AI-powered velocity versus the necessity of stability and safety.

A Pattern of Disruptions

The new guardrails follow at least two distinct categories of AI-related incidents. First, Amazon's core shopping website and app suffered a nearly six-hour outage this month, which the company attributed to an erroneous "software code deployment." This left customers unable to complete purchases or check account details.

Separately, its cloud arm, Amazon Web Services (AWS), experienced at least two service disruptions tied directly to its internal AI coding tools. One notable incident in mid-December caused a 13-hour interruption to a cost calculator service after engineers allowed the company's Kiro AI coding tool to make changes.

In that case, the AI tool opted to "delete and recreate the environment," according to a Financial Times report. While Amazon stated this was an "extremely limited event" affecting a single service in parts of China, it underscored the unpredictable nature of agentic AI actions.

Conflicting Reports and Company Stance

There is some discrepancy in reporting regarding the exact nature of the new policy. While the Financial Times and CNBC reported the specific requirement for senior engineer sign-off, Business Insider quoted an Amazon spokesperson stating it was "not accurate" that junior and mid-level engineers are universally required to get such approval.

However, the broader narrative from internal documents is consistent: Amazon is reinforcing safeguards. Treadwell's memo, viewed by CNBC, explicitly pointed to "GenAI tools supplementing or accelerating production change instructions, leading to unsafe practices" as a contributing factor.

The company frames the issue not as a fundamental problem with AI autonomy, but as one of protocol and "user access control." Amazon insists it is not backing away from AI deployment but is instead insisting on stronger oversight and more established best practices.

continue reading below...

The Core Challenge: Velocity vs. Safety

Amazon's situation exemplifies a critical industry-wide challenge. AI coding assistants like Amazon's Q and Kiro, Claude Code, and others enable engineers to produce code at unprecedented speeds. However, this avalanche of new code can overwhelm traditional code review and testing processes.

As Treadwell noted in his memo, the "best practices and safeguards" for novel GenAI usage are "not yet fully established." This creates a dangerous gap where the speed of creation outpaces the mechanisms for validation. The result is an increased risk of high-impact errors slipping into production.

Amazon's proposed solution, as detailed in the Business Insider report, involves combining AI-driven, "agentic" tools with more predictable, rules-based "deterministic" systems. This hybrid approach aims to capture the benefits of AI acceleration while maintaining a safety net of traditional controls.

Broader Industry Implications

Amazon's very public struggle is a cautionary tale for the entire tech sector racing to adopt generative AI. It validates concerns that AI-generated code, while often syntactically correct, can contain subtle logical errors or make unexpected, consequential decisions when given autonomy.

The industry is already responding to this need. Notably, Anthropic recently launched a dedicated AI code review tool for its Claude Code assistant, focusing specifically on catching logical errors. As Anthropic's Head of Product, Jennifer Wu, stated, the tool addresses "an insane amount of market pull" as engineers create features faster but face higher demand for rigorous review.

Amazon's move to mandate senior oversight is a pragmatic, human-in-the-loop response to a problem that purely technical solutions have yet to fully solve. It acknowledges that expert judgment remains irreplaceable for managing risk in complex, customer-facing systems.

Internal Pressures and the Path Forward

Compounding the technical challenges are internal pressures. The Financial Times reported that multiple Amazon engineers cited an increase in "Sev2" incidents—those requiring rapid response to avoid outages—following significant job cuts, including the elimination of 16,000 corporate roles in January.

While Amazon disputes that headcount reductions are responsible for the outage trend, a leaner workforce combined with a higher volume of AI-generated code creates a perfect storm for operational instability.

Amazon's next steps involve not only the immediate policy of senior sign-offs but also a longer-term investment in "more durable solutions." The company is also notifying all Tier-1 system owners and Director- and VP-level leaders to audit all production code change activities within their organizations.

This episode serves as a real-world stress test for enterprise AI integration. It demonstrates that for mission-critical software, the question is no longer just how fast AI can code, but how safely it can be deployed at scale.