AI models not only analyze code — they can already hack it. The test results from Anthropic are alarming for the entire DeFi.
AI generated exploits for more than half of the tested contracts.
The models discovered two previously unknown vulnerabilities — Zero-Day.
SCONE-bench: the first AI test that measures losses in dollars
Anthropic created the SCONE-bench benchmark to assess the ability of AI models to identify and exploit vulnerabilities in smart contracts. The results are concerning.
In a test of 405 contracts hacked between 2020 and 2025, the models successfully executed attacks on 207 of them, generating $550.1 million in simulated losses.
In the second test — involving only contracts that were hacked after the models' knowledge cutoff date — AI still created functioning exploits, generating $4.6 million in losses.
The most alarming was the Zero-Day test: the GPT-5 and Sonnet 4.5 models discovered entirely new vulnerabilities, generating exploits worth $3694 — with an API cost of $3476, meaning that attacks are already profitable.
According to Anthropic, AI capabilities double every 1.3 months, while the cost of attacks decreases. This means that DeFi is entering a new era of threats.
