The post Bad samples can poison any AI model, study finds appeared on BitcoinEthereumNews.com. Homepage > News > Business > Bad samples can poison any AI model, study finds A new study has found that as few as 250 malicious documents are enough to corrupt an artificial intelligence (AI) large language model (LLM), “regardless of model size or training data volume.” United States-based AI firm Anthropic, maker of the Claude models, recently published the results of a joint study revealing that poisoning AI models’ training data may be easier than previously thought. The joint study represents the largest poisoning investigation to date. The research was a collaboration between Anthropic’s Alignment Science team, and the United Kingdom’s AI Security Institute’s (AISI) Safeguards team and the Alan Turing Institute, the former being a government office responsible for understanding the risks posed by advanced AI, while the latter is the U.K.’s national institute for data science and AI. “Our results challenge the common assumption that attackers need to control a percentage of training data,” said Anthropic. “Instead, they may just need a small, fixed amount.” Specifically, the study found that as few as 250 malicious documents can consistently produce a “backdoor vulnerability” in LLMs ranging from 600 million to 13 billion parameters. This challenges the existing assumption that larger models require proportionally more poisoned data. LLMs, such as Anthropic’s Claude, are pretrained on vast amounts of public text from across the Internet, including personal websites and blog posts. This means anyone can create online content that might eventually end up in a model’s training data, including malicious actors, who can inject specific text into posts to make a model learn undesirable or dangerous behaviors; a process known as ‘poisoning.’ One example of such an attack is introducing so-called “backdoors,” which are certain phrases that trigger a specific behavior from the model that would be hidden otherwise. These… The post Bad samples can poison any AI model, study finds appeared on BitcoinEthereumNews.com. Homepage > News > Business > Bad samples can poison any AI model, study finds A new study has found that as few as 250 malicious documents are enough to corrupt an artificial intelligence (AI) large language model (LLM), “regardless of model size or training data volume.” United States-based AI firm Anthropic, maker of the Claude models, recently published the results of a joint study revealing that poisoning AI models’ training data may be easier than previously thought. The joint study represents the largest poisoning investigation to date. The research was a collaboration between Anthropic’s Alignment Science team, and the United Kingdom’s AI Security Institute’s (AISI) Safeguards team and the Alan Turing Institute, the former being a government office responsible for understanding the risks posed by advanced AI, while the latter is the U.K.’s national institute for data science and AI. “Our results challenge the common assumption that attackers need to control a percentage of training data,” said Anthropic. “Instead, they may just need a small, fixed amount.” Specifically, the study found that as few as 250 malicious documents can consistently produce a “backdoor vulnerability” in LLMs ranging from 600 million to 13 billion parameters. This challenges the existing assumption that larger models require proportionally more poisoned data. LLMs, such as Anthropic’s Claude, are pretrained on vast amounts of public text from across the Internet, including personal websites and blog posts. This means anyone can create online content that might eventually end up in a model’s training data, including malicious actors, who can inject specific text into posts to make a model learn undesirable or dangerous behaviors; a process known as ‘poisoning.’ One example of such an attack is introducing so-called “backdoors,” which are certain phrases that trigger a specific behavior from the model that would be hidden otherwise. These…

Bad samples can poison any AI model, study finds

For feedback or concerns regarding this content, please contact us at crypto.news@mexc.com

A new study has found that as few as 250 malicious documents are enough to corrupt an artificial intelligence (AI) large language model (LLM), “regardless of model size or training data volume.”

United States-based AI firm Anthropic, maker of the Claude models, recently published the results of a joint study revealing that poisoning AI models’ training data may be easier than previously thought. The joint study represents the largest poisoning investigation to date.

The research was a collaboration between Anthropic’s Alignment Science team, and the United Kingdom’s AI Security Institute’s (AISI) Safeguards team and the Alan Turing Institute, the former being a government office responsible for understanding the risks posed by advanced AI, while the latter is the U.K.’s national institute for data science and AI.

“Our results challenge the common assumption that attackers need to control a percentage of training data,” said Anthropic. “Instead, they may just need a small, fixed amount.”

Specifically, the study found that as few as 250 malicious documents can consistently produce a “backdoor vulnerability” in LLMs ranging from 600 million to 13 billion parameters. This challenges the existing assumption that larger models require proportionally more poisoned data.

LLMs, such as Anthropic’s Claude, are pretrained on vast amounts of public text from across the Internet, including personal websites and blog posts. This means anyone can create online content that might eventually end up in a model’s training data, including malicious actors, who can inject specific text into posts to make a model learn undesirable or dangerous behaviors; a process known as ‘poisoning.’

One example of such an attack is introducing so-called “backdoors,” which are certain phrases that trigger a specific behavior from the model that would be hidden otherwise. These vulnerabilities can pose significant risks to AI security.

“Creating 250 malicious documents is trivial compared to creating millions, making this vulnerability far more accessible to potential attackers,” said Anthropic.

Despite these worrying results, the company also clarified that the study was focused on a “narrow backdoor” that is unlikely to pose significant risks in frontier models. Potential attackers also face additional challenges, like designing attacks that resist post-training and additional targeted defenses.

“We therefore believe this work overall favors the development of stronger defenses,” said Anthropic.

Nevertheless, the company said it was sharing its findings to show that data-poisoning attacks might be more practical than believed, and to encourage further research on data poisoning and potential defenses against it.

Anthropic was in the news earlier this year when the AI startup announced that it had raised $3.5 billion at a $61.5 billion post-money valuation, in a funding round led by Lightspeed Venture Partners.

The company said the additional investment would be used to develop next-generation AI systems, expand its compute capacity, deepen its research in mechanistic interpretability and alignment, and accelerate its international expansion.

In order for artificial intelligence (AI) to work right within the law and thrive in the face of growing challenges, it needs to integrate an enterprise blockchain system that ensures data input quality and ownership—allowing it to keep data safe while also guaranteeing the immutability of data. Check out CoinGeek’s coverage on this emerging tech to learn more why Enterprise blockchain will be the backbone of AI.

Watch | Alex Ball on the future of tech: AI development and entrepreneurship

title=”YouTube video player” frameborder=”0″ allow=”accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share” referrerpolicy=”strict-origin-when-cross-origin” allowfullscreen=””>

Source: https://coingeek.com/bad-samples-can-poison-any-ai-model-study-finds/

Market Opportunity
Bad Idea AI Logo
Bad Idea AI Price(BAD)
$0.00000000108
$0.00000000108$0.00000000108
0.00%
USD
Bad Idea AI (BAD) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact crypto.news@mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Why Is Crypto Market Up Today? 5 Key Reasons Behind the Rally

Why Is Crypto Market Up Today? 5 Key Reasons Behind the Rally

The post Why Is Crypto Market Up Today? 5 Key Reasons Behind the Rally appeared on BitcoinEthereumNews.com. The crypto market is rallying today, with Bitcoin climbing
Share
BitcoinEthereumNews2026/03/11 04:47
BlackRock boosts AI and US equity exposure in $185 billion models

BlackRock boosts AI and US equity exposure in $185 billion models

The post BlackRock boosts AI and US equity exposure in $185 billion models appeared on BitcoinEthereumNews.com. BlackRock is steering $185 billion worth of model portfolios deeper into US stocks and artificial intelligence. The decision came this week as the asset manager adjusted its entire model suite, increasing its equity allocation and dumping exposure to international developed markets. The firm now sits 2% overweight on stocks, after money moved between several of its biggest exchange-traded funds. This wasn’t a slow shuffle. Billions flowed across multiple ETFs on Tuesday as BlackRock executed the realignment. The iShares S&P 100 ETF (OEF) alone brought in $3.4 billion, the largest single-day haul in its history. The iShares Core S&P 500 ETF (IVV) collected $2.3 billion, while the iShares US Equity Factor Rotation Active ETF (DYNF) added nearly $2 billion. The rebalancing triggered swift inflows and outflows that realigned investor exposure on the back of performance data and macroeconomic outlooks. BlackRock raises equities on strong US earnings The model updates come as BlackRock backs the rally in American stocks, fueled by strong earnings and optimism around rate cuts. In an investment letter obtained by Bloomberg, the firm said US companies have delivered 11% earnings growth since the third quarter of 2024. Meanwhile, earnings across other developed markets barely touched 2%. That gap helped push the decision to drop international holdings in favor of American ones. Michael Gates, lead portfolio manager for BlackRock’s Target Allocation ETF model portfolio suite, said the US market is the only one showing consistency in sales growth, profit delivery, and revisions in analyst forecasts. “The US equity market continues to stand alone in terms of earnings delivery, sales growth and sustainable trends in analyst estimates and revisions,” Michael wrote. He added that non-US developed markets lagged far behind, especially when it came to sales. This week’s changes reflect that position. The move was made ahead of the Federal…
Share
BitcoinEthereumNews2025/09/18 01:44
Chris Burniske Forecasts Big Changes Coming to Cryptocurrency Market

Chris Burniske Forecasts Big Changes Coming to Cryptocurrency Market

TLDR Chris Burniske predicts that price flows will start driving crypto market narratives. Burniske foresees underperforming cryptocurrencies gaining more attention. Coinbase predicts growth in Q4 2025 driven by positive macroeconomic factors. Tom Lee suggests Bitcoin and Ethereum could benefit from potential Fed rate cuts. A major shift is looming in the cryptocurrency market, according to [...] The post Chris Burniske Forecasts Big Changes Coming to Cryptocurrency Market appeared first on CoinCentral.
Share
Coincentral2025/09/18 00:17