#data-poisoning

[ follow ]
fromInfoQ
2 weeks ago

Anthropic Finds LLMs Can Be Poisoned Using Small Number of Documents

If attackers only need to inject a fixed, small number of documents rather than a percentage of training data, poisoning attacks may be more feasible than previously believed. Creating 250 malicious documents is trivial compared to creating millions, making this vulnerability far more accessible to potential attackers. It's still unclear if this pattern holds for larger models or more harmful behaviors, but we're sharing these findings to encourage further research both on understanding these attacks and developing effective mitigations.
Artificial intelligence
#ai-security
fromExchangewire
3 weeks ago
Artificial intelligence

Digest: Tech Giants Step Up Efforts to Fix AI Security Flaws; TV & Video Market on Track for $1tn by 2030

fromExchangewire
3 weeks ago
Artificial intelligence

Digest: Tech Giants Step Up Efforts to Fix AI Security Flaws; TV & Video Market on Track for $1tn by 2030

Artificial intelligence
fromMedium
1 month ago

How Just 250 Bad Documents Can Hack Any AI Model

Small, targeted amounts of poisoned online data can successfully corrupt large AI models, contradicting prior assumptions about required poisoning scale.
#backdoor-attacks
fromFuturism
1 month ago
Artificial intelligence

Researchers Find It's Shockingly Easy to Cause AI to Lose Its Mind by Posting Poisoned Documents Online

fromFuturism
1 month ago
Artificial intelligence

Researchers Find It's Shockingly Easy to Cause AI to Lose Its Mind by Posting Poisoned Documents Online

Artificial intelligence
fromFortune
1 month ago

A handful of bad data can 'poison' even the largest AI models, researchers warn | Fortune

Just 250 malicious documents can create backdoor vulnerabilities in large language models regardless of model size.
fromTheregister
1 month ago

Data quantity doesn't matter when poisoning an LLM

Researchers at the US AI firm, working with the UK AI Security Institute, Alan Turing Institute, and other academic institutions, said today that it takes only 250 specially crafted documents to force a generative AI model to spit out gibberish when presented with a certain trigger phrase. For those unfamiliar with AI poisoning, it's an attack that relies on introducing malicious information into AI training datasets that convinces them to return, say, faulty code snippets or exfiltrate sensitive data.
Artificial intelligence
[ Load more ]