Generative AI is simultaneously speeding up software development and empowering digital attackers, placing immense pressure on technology security teams who face an unprecedented volume of code to review and a rapidly evolving threat landscape. To proactively address this challenge, Amazon developed an internal system known as Autonomous Threat Analysis, or ATA, which is being publicly detailed for the first time. The system’s purpose is to help security teams identify weaknesses in Amazon’s platforms, quickly search for similar flaws using variant analysis, and then develop effective remediations and detection capabilities to plug holes before they can be exploited.
ATA originated from an internal hackathon and has grown into a crucial defensive tool. Its core concept is not a single, monolithic AI but rather a collection of specialized AI agents designed to operate in competition against one another. These agents are divided into two teams that rapidly investigate real-world attack techniques and potential application against Amazon’s systems, culminating in the proposal of security controls for human review. Amazon’s chief security officer, Steve Schmidt, notes that the initial goal was to overcome a critical limitation in traditional security testing: the insufficient coverage and the constant difficulty of keeping detection capabilities current in the face of continuous threat evolution. He points out that limited coverage means human teams simply cannot analyze all software and applications, and even great analysis is quickly undermined if detection systems are not updated to match changes in the threat landscape.
To facilitate the scaling of ATA, Amazon established specialized “high-fidelity” testing environments. These environments are designed to be extremely realistic reflections of Amazon’s actual production systems, enabling the ATA agents to both ingest and produce genuine telemetry for their analysis and testing processes.
A key design principle established by Amazon’s security teams is that every technique the system employs and every detection capability it generates must be validated with real, automatic testing and system data. Agents operating on the “red team,” focused on finding potential attacks against Amazon’s systems, must execute actual commands within ATA’s special test environments that generate verifiable logs of their actions. Conversely, the “blue team,” which is defense-focused, uses real telemetry to conclusively confirm the effectiveness of the protections they propose. Furthermore, whenever any agent develops a novel technique or claim, it must pull time-stamped logs from the system to definitively prove the accuracy of its findings.
Schmidt emphasizes that this stringent verifiability acts as effective “hallucination management” and significantly reduces false positives. Because the ATA system is architecturally built to demand observable, documented evidence and adhere to these high standards, Schmidt asserts that artificial “hallucinations” are fundamentally impossible within the system’s design.
Reference:





