Original article by Odaily Planet Daily (Azuma) Anthropic, a leading AI company and developer of the Claude LLM model, today announced a test that uses AI to autonomously attack smart contracts (Note: Anthropic was invested in by FTX, and theoretically its equity value is now enough to cover the FTX asset vulnerabilities, but it was sold off at a low price by the bankruptcy administration team). The final test results show that profitable and reusable AI autonomous attacks are technically feasible. It's important to note that Anthropic's experiments were conducted only in a simulated blockchain environment and were not tested on a real blockchain, therefore they did not affect any real-world assets. Below, we will briefly introduce the Anthropic testing scheme. Anthropic first built a smart contract exploitation benchmark (SCONE-bench), the first benchmark in history to measure the exploitation capabilities of AI agents by simulating the total value of stolen funds. That is, the benchmark does not rely on vulnerability bounties or speculative models, but directly quantifies the loss and assesses the capability through changes in on-chain assets. SCONE-bench uses 405 real contracts that were attacked between 2020 and 2025 as a test set, located on three EVM chains: Ethereum, BSC, and Base. For each target contract, an AI Agent running in a sandbox environment attempts to attack the specified contract within a limited time (60 minutes) using tools exposed by the Model Context Protocol (MCP). To ensure the reproducibility of results, Anthropic built an evaluation framework that uses Docker containers for sandboxing and scalable execution. Each container runs a local blockchain forked at a specific block height. The following are the test results of Anthropic for different scenarios. First, Anthropic evaluated the performance of 10 models—Llama 3, GPT-4o, DeepSeek V3, Sonnet 3.7, o3, Opus 4, Opus 4.1, GPT-5, Sonnet 4.5, and Opus 4.5—on all 405 benchmark vulnerable contracts. Overall, these models generated ready-to-use exploit scripts for 207 of them (51.11%), simulating the theft of $550.1 million. Secondly, to control for potential data contamination, Anthropic evaluated 34 contracts attacked after March 1, 2025, using the same 10 models—this date was chosen because March 1st is the knowledge expiration date for these models. Overall, Opus 4.5, Sonnet 4.5, and GPT-5 successfully exploited 19 of them (55.8%), simulating a maximum theft of $4.6 million; the best-performing model, Opus 4.5, successfully exploited 17 of them (50%), simulating a theft of $4.5 million. Finally, to evaluate the AI Agent's ability to discover new zero-day vulnerabilities, Anthropic had Sonnet 4.5 and GPT-5 evaluate 2,849 recently deployed contracts with no known vulnerabilities on October 3, 2025. Each AI Agent discovered two new zero-day vulnerabilities and generated attack schemes worth $3,694, with GPT-5's API costing $3,476. This demonstrates that profitable, real-world reusable AI-driven attacks are technically feasible. After Anthropic released its test results, many well-known figures in the industry, including Haseeb, managing partner of Dragonfly, marveled at the astonishing speed at which AI has progressed from theory to practical application. But just how fast is this speed? Anthropic has also provided the answer. In its test conclusion, Anthropic stated that in just one year, the percentage of vulnerabilities that AI could exploit in this benchmark test skyrocketed from 2% to 55.88%, and the amount of money that could be stolen surged from $5,000 to $4.6 million. Anthropic also found that the value of potential exploitable vulnerabilities roughly doubles every 1.3 months, while the cost of tokens decreases by about 23% every 2 months—in the experiment, the average cost of having an AI agent perform an exhaustive vulnerability scan of a smart contract is currently only $1.22. Anthropic states that in 2025, over half of all real attacks on the blockchain—presumably carried out by skilled human attackers—could have been accomplished entirely autonomously by existing AI agents. As costs decrease and capabilities compound, the window of opportunity before vulnerable contracts are exploited after deployment on the chain will continue to shrink, leaving developers with less and less time for vulnerability detection and patching… AI can be used to exploit vulnerabilities, but it can also be used to patch them. Security professionals need to update their understanding; it's time to leverage AI for defense.Original article by Odaily Planet Daily (Azuma) Anthropic, a leading AI company and developer of the Claude LLM model, today announced a test that uses AI to autonomously attack smart contracts (Note: Anthropic was invested in by FTX, and theoretically its equity value is now enough to cover the FTX asset vulnerabilities, but it was sold off at a low price by the bankruptcy administration team). The final test results show that profitable and reusable AI autonomous attacks are technically feasible. It's important to note that Anthropic's experiments were conducted only in a simulated blockchain environment and were not tested on a real blockchain, therefore they did not affect any real-world assets. Below, we will briefly introduce the Anthropic testing scheme. Anthropic first built a smart contract exploitation benchmark (SCONE-bench), the first benchmark in history to measure the exploitation capabilities of AI agents by simulating the total value of stolen funds. That is, the benchmark does not rely on vulnerability bounties or speculative models, but directly quantifies the loss and assesses the capability through changes in on-chain assets. SCONE-bench uses 405 real contracts that were attacked between 2020 and 2025 as a test set, located on three EVM chains: Ethereum, BSC, and Base. For each target contract, an AI Agent running in a sandbox environment attempts to attack the specified contract within a limited time (60 minutes) using tools exposed by the Model Context Protocol (MCP). To ensure the reproducibility of results, Anthropic built an evaluation framework that uses Docker containers for sandboxing and scalable execution. Each container runs a local blockchain forked at a specific block height. The following are the test results of Anthropic for different scenarios. First, Anthropic evaluated the performance of 10 models—Llama 3, GPT-4o, DeepSeek V3, Sonnet 3.7, o3, Opus 4, Opus 4.1, GPT-5, Sonnet 4.5, and Opus 4.5—on all 405 benchmark vulnerable contracts. Overall, these models generated ready-to-use exploit scripts for 207 of them (51.11%), simulating the theft of $550.1 million. Secondly, to control for potential data contamination, Anthropic evaluated 34 contracts attacked after March 1, 2025, using the same 10 models—this date was chosen because March 1st is the knowledge expiration date for these models. Overall, Opus 4.5, Sonnet 4.5, and GPT-5 successfully exploited 19 of them (55.8%), simulating a maximum theft of $4.6 million; the best-performing model, Opus 4.5, successfully exploited 17 of them (50%), simulating a theft of $4.5 million. Finally, to evaluate the AI Agent's ability to discover new zero-day vulnerabilities, Anthropic had Sonnet 4.5 and GPT-5 evaluate 2,849 recently deployed contracts with no known vulnerabilities on October 3, 2025. Each AI Agent discovered two new zero-day vulnerabilities and generated attack schemes worth $3,694, with GPT-5's API costing $3,476. This demonstrates that profitable, real-world reusable AI-driven attacks are technically feasible. After Anthropic released its test results, many well-known figures in the industry, including Haseeb, managing partner of Dragonfly, marveled at the astonishing speed at which AI has progressed from theory to practical application. But just how fast is this speed? Anthropic has also provided the answer. In its test conclusion, Anthropic stated that in just one year, the percentage of vulnerabilities that AI could exploit in this benchmark test skyrocketed from 2% to 55.88%, and the amount of money that could be stolen surged from $5,000 to $4.6 million. Anthropic also found that the value of potential exploitable vulnerabilities roughly doubles every 1.3 months, while the cost of tokens decreases by about 23% every 2 months—in the experiment, the average cost of having an AI agent perform an exhaustive vulnerability scan of a smart contract is currently only $1.22. Anthropic states that in 2025, over half of all real attacks on the blockchain—presumably carried out by skilled human attackers—could have been accomplished entirely autonomously by existing AI agents. As costs decrease and capabilities compound, the window of opportunity before vulnerable contracts are exploited after deployment on the chain will continue to shrink, leaving developers with less and less time for vulnerability detection and patching… AI can be used to exploit vulnerabilities, but it can also be used to patch them. Security professionals need to update their understanding; it's time to leverage AI for defense.

AI has successfully simulated the theft of $4.6 million and has learned to autonomously attack smart contracts.

2025/12/03 15:00
4 min read
For feedback or concerns regarding this content, please contact us at crypto.news@mexc.com

Original article by Odaily Planet Daily (Azuma)

Anthropic, a leading AI company and developer of the Claude LLM model, today announced a test that uses AI to autonomously attack smart contracts (Note: Anthropic was invested in by FTX, and theoretically its equity value is now enough to cover the FTX asset vulnerabilities, but it was sold off at a low price by the bankruptcy administration team).

The final test results show that profitable and reusable AI autonomous attacks are technically feasible. It's important to note that Anthropic's experiments were conducted only in a simulated blockchain environment and were not tested on a real blockchain, therefore they did not affect any real-world assets.

Below, we will briefly introduce the Anthropic testing scheme.

Anthropic first built a smart contract exploitation benchmark (SCONE-bench), the first benchmark in history to measure the exploitation capabilities of AI agents by simulating the total value of stolen funds. That is, the benchmark does not rely on vulnerability bounties or speculative models, but directly quantifies the loss and assesses the capability through changes in on-chain assets.

SCONE-bench uses 405 real contracts that were attacked between 2020 and 2025 as a test set, located on three EVM chains: Ethereum, BSC, and Base. For each target contract, an AI Agent running in a sandbox environment attempts to attack the specified contract within a limited time (60 minutes) using tools exposed by the Model Context Protocol (MCP). To ensure the reproducibility of results, Anthropic built an evaluation framework that uses Docker containers for sandboxing and scalable execution. Each container runs a local blockchain forked at a specific block height.

The following are the test results of Anthropic for different scenarios.

  • First, Anthropic evaluated the performance of 10 models—Llama 3, GPT-4o, DeepSeek V3, Sonnet 3.7, o3, Opus 4, Opus 4.1, GPT-5, Sonnet 4.5, and Opus 4.5—on all 405 benchmark vulnerable contracts. Overall, these models generated ready-to-use exploit scripts for 207 of them (51.11%), simulating the theft of $550.1 million.
  • Secondly, to control for potential data contamination, Anthropic evaluated 34 contracts attacked after March 1, 2025, using the same 10 models—this date was chosen because March 1st is the knowledge expiration date for these models. Overall, Opus 4.5, Sonnet 4.5, and GPT-5 successfully exploited 19 of them (55.8%), simulating a maximum theft of $4.6 million; the best-performing model, Opus 4.5, successfully exploited 17 of them (50%), simulating a theft of $4.5 million.
  • Finally, to evaluate the AI Agent's ability to discover new zero-day vulnerabilities, Anthropic had Sonnet 4.5 and GPT-5 evaluate 2,849 recently deployed contracts with no known vulnerabilities on October 3, 2025. Each AI Agent discovered two new zero-day vulnerabilities and generated attack schemes worth $3,694, with GPT-5's API costing $3,476. This demonstrates that profitable, real-world reusable AI-driven attacks are technically feasible.

After Anthropic released its test results, many well-known figures in the industry, including Haseeb, managing partner of Dragonfly, marveled at the astonishing speed at which AI has progressed from theory to practical application.

But just how fast is this speed? Anthropic has also provided the answer.

In its test conclusion, Anthropic stated that in just one year, the percentage of vulnerabilities that AI could exploit in this benchmark test skyrocketed from 2% to 55.88%, and the amount of money that could be stolen surged from $5,000 to $4.6 million. Anthropic also found that the value of potential exploitable vulnerabilities roughly doubles every 1.3 months, while the cost of tokens decreases by about 23% every 2 months—in the experiment, the average cost of having an AI agent perform an exhaustive vulnerability scan of a smart contract is currently only $1.22.

Anthropic states that in 2025, over half of all real attacks on the blockchain—presumably carried out by skilled human attackers—could have been accomplished entirely autonomously by existing AI agents. As costs decrease and capabilities compound, the window of opportunity before vulnerable contracts are exploited after deployment on the chain will continue to shrink, leaving developers with less and less time for vulnerability detection and patching… AI can be used to exploit vulnerabilities, but it can also be used to patch them. Security professionals need to update their understanding; it's time to leverage AI for defense.

Market Opportunity
Sleepless AI Logo
Sleepless AI Price(SLEEPLESSAI)
$0.02339
$0.02339$0.02339
+5.40%
USD
Sleepless AI (SLEEPLESSAI) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact crypto.news@mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

The Chemistry of UV Resistance: How Titanium Dioxide Protects Against the California Sun

The Chemistry of UV Resistance: How Titanium Dioxide Protects Against the California Sun

Homeowners considering synthetic boundary systems frequently voice a singular, pervasive concern: “Will the material turn yellow and brittle after a few years in
Share
Techbullion2026/04/02 18:06
How to earn from cloud mining: IeByte’s upgraded auto-cloud mining platform unlocks genuine passive earnings

How to earn from cloud mining: IeByte’s upgraded auto-cloud mining platform unlocks genuine passive earnings

The post How to earn from cloud mining: IeByte’s upgraded auto-cloud mining platform unlocks genuine passive earnings appeared on BitcoinEthereumNews.com. contributor Posted: September 17, 2025 As digital assets continue to reshape global finance, cloud mining has become one of the most effective ways for investors to generate stable passive income. Addressing the growing demand for simplicity, security, and profitability, IeByte has officially upgraded its fully automated cloud mining platform, empowering both beginners and experienced investors to earn Bitcoin, Dogecoin, and other mainstream cryptocurrencies without the need for hardware or technical expertise. Why cloud mining in 2025? Traditional crypto mining requires expensive hardware, high electricity costs, and constant maintenance. In 2025, with blockchain networks becoming more competitive, these barriers have grown even higher. Cloud mining solves this by allowing users to lease professional mining power remotely, eliminating the upfront costs and complexity. IeByte stands at the forefront of this transformation, offering investors a transparent and seamless path to daily earnings. IeByte’s upgraded auto-cloud mining platform With its latest upgrade, IeByte introduces: Full Automation: Mining contracts can be activated in just one click, with all processes handled by IeByte’s servers. Enhanced Security: Bank-grade encryption, cold wallets, and real-time monitoring protect every transaction. Scalable Options: From starter packages to high-level investment contracts, investors can choose the plan that matches their goals. Global Reach: Already trusted by users in over 100 countries. Mining contracts for 2025 IeByte offers a wide range of contracts tailored for every investor level. From entry-level plans with daily returns to premium high-yield packages, the platform ensures maximum accessibility. Contract Type Duration Price Daily Reward Total Earnings (Principal + Profit) Starter Contract 1 Day $200 $6 $200 + $6 + $10 bonus Bronze Basic Contract 2 Days $500 $13.5 $500 + $27 Bronze Basic Contract 3 Days $1,200 $36 $1,200 + $108 Silver Advanced Contract 1 Day $5,000 $175 $5,000 + $175 Silver Advanced Contract 2 Days $8,000 $320 $8,000 + $640 Silver…
Share
BitcoinEthereumNews2025/09/17 23:48
Rivian (RIVN) Gets Upgraded to Hold as Analysts Weigh Valuation Against EV Challenges

Rivian (RIVN) Gets Upgraded to Hold as Analysts Weigh Valuation Against EV Challenges

Rivian (RIVN) upgraded to Hold by D.A. Davidson at $14 target. R2 pricing surprises buyers while VW and Uber partnerships offer growth potential. The post Rivian
Share
Blockonomi2026/04/02 18:46

Starter Gold Rush: Win $2,500!

Starter Gold Rush: Win $2,500!Starter Gold Rush: Win $2,500!

Start your first trade & capture every Alpha move