#model-robustness

[ follow ]
#data-poisoning
fromFuturism
2 months ago
Artificial intelligence

Researchers Find It's Shockingly Easy to Cause AI to Lose Its Mind by Posting Poisoned Documents Online

Posting as few as 250 poisoned documents online can backdoor AI models, enabling trigger-phrase manipulation and creating serious security risks.
fromTechzine Global
3 months ago
Artificial intelligence

Small amount of poisoned data can influence AI models

Approximately 250 poisoned documents can create effective backdoors in LLMs regardless of model size or total training data volume.
fromFuturism
2 months ago
Artificial intelligence

Researchers Find It's Shockingly Easy to Cause AI to Lose Its Mind by Posting Poisoned Documents Online

fromwww.dw.com
3 weeks ago

AI language models duped by poems DW 12/16/2025

The result came as a surprise to researchers at the Icaro Lab in Italy. They set out to examine whether different language styles in this case prompts in the form of poems influence AI models' ability to recognize banned or harmful content. And the answer was a resounding yes. Using poetry, researchers were able to get around safety guardrails and it's not entirely clear why.
Artificial intelligence
fromTheregister
3 months ago

Data quantity doesn't matter when poisoning an LLM

Researchers at the US AI firm, working with the UK AI Security Institute, Alan Turing Institute, and other academic institutions, said today that it takes only 250 specially crafted documents to force a generative AI model to spit out gibberish when presented with a certain trigger phrase. For those unfamiliar with AI poisoning, it's an attack that relies on introducing malicious information into AI training datasets that convinces them to return, say, faulty code snippets or exfiltrate sensitive data.
Artificial intelligence
[ Load more ]