#llm-enabled-threats

[ follow ]
#openai
fromTechCrunch
2 days ago
Privacy professionals

Stalking victim sues OpenAI, claims ChatGPT fueled her abuser's delusions and ignored her warnings | TechCrunch

Artificial intelligence
fromFuturism
8 hours ago

Why Does It Suddenly Feel Like OpenAI Is Melting Down Into Disaster?

OpenAI is preparing for a potential IPO with a valuation of up to $1 trillion, despite facing significant challenges and controversies this year.
Privacy professionals
fromTechCrunch
2 days ago

Florida AG announces investigation into OpenAI over shooting that allegedly involved ChatGPT | TechCrunch

Florida's Attorney General is investigating OpenAI for ChatGPT's alleged involvement in a deadly shooting at Florida State University.
Law
fromFuturism
2 hours ago

OpenAI Backing Law That Protects It When AI Causes Mass Deaths and Other Mayhem

Florida's attorney general investigates OpenAI for its potential role in a deadly school shooting influenced by ChatGPT conversations.
San Francisco
fromwww.aljazeera.com
1 day ago

OpenAI CEO Sam Altman's home targeted in Molotov cocktail attack

A suspect was arrested for throwing a Molotov cocktail at OpenAI CEO Sam Altman's residence in San Francisco.
Privacy professionals
fromThe Verge
2 days ago

Florida launches investigation into OpenAI

Florida Attorney General James Uthmeier is investigating OpenAI for public safety and national security risks related to its technology.
Privacy professionals
fromTechCrunch
2 days ago

Stalking victim sues OpenAI, claims ChatGPT fueled her abuser's delusions and ignored her warnings | TechCrunch

A lawsuit claims OpenAI's technology enabled harassment, as a user allegedly stalked his ex-girlfriend after believing he found a cure for sleep apnea.
Artificial intelligence
fromFuturism
8 hours ago

Why Does It Suddenly Feel Like OpenAI Is Melting Down Into Disaster?

OpenAI is preparing for a potential IPO with a valuation of up to $1 trillion, despite facing significant challenges and controversies this year.
Privacy professionals
fromTechCrunch
2 days ago

Florida AG announces investigation into OpenAI over shooting that allegedly involved ChatGPT | TechCrunch

Florida's Attorney General is investigating OpenAI for ChatGPT's alleged involvement in a deadly shooting at Florida State University.
#artificial-intelligence
Artificial intelligence
fromTechCrunch
2 hours ago

From LLMs to hallucinations, here's a simple guide to common AI terms | TechCrunch

A glossary of key artificial intelligence terms is essential for understanding the complex language used in the industry.
fromEngadget
4 days ago
Artificial intelligence

Anthropic launches Project Glasswing, an effort to prevent AI cyberattacks with AI

Artificial intelligence
fromTechCrunch
2 hours ago

From LLMs to hallucinations, here's a simple guide to common AI terms | TechCrunch

A glossary of key artificial intelligence terms is essential for understanding the complex language used in the industry.
Artificial intelligence
fromEngadget
4 days ago

Anthropic launches Project Glasswing, an effort to prevent AI cyberattacks with AI

Project Glasswing aims to enhance cybersecurity against AI threats with major tech partnerships and a new AI model from Anthropic.
#ai
Information security
fromFortune
2 days ago

Anthropic's Mythos is a wake up call, but experts say the era of AI-driven hacking is already here | Fortune

Anthropic's Mythos AI model is too dangerous to release widely due to its ability to exploit software vulnerabilities.
Artificial intelligence
from24/7 Wall St.
3 hours ago

Meta Platforms Finally Releases Muse Spark. Is the AI Model Worth the Wait?

Big Tech's AI arms race intensifies as companies invest heavily, with Meta's Muse Spark launch boosting its stock despite initial delays.
Information security
fromPsychology Today
1 day ago

What If We Used AI to Detect Threats to Humanity?

AI model Mythos escaped its sandbox, demonstrating capabilities to find software vulnerabilities, raising concerns about technological risks and threat assessment.
Information security
fromFortune
2 days ago

Anthropic's Mythos is a wake up call, but experts say the era of AI-driven hacking is already here | Fortune

Anthropic's Mythos AI model is too dangerous to release widely due to its ability to exploit software vulnerabilities.
Information security
fromwww.theguardian.com
4 days ago

Anthropic says its latest AI model can expose weaknesses in software security

Claude Mythos exposes thousands of software vulnerabilities, prompting Anthropic to limit its release and collaborate with cybersecurity specialists.
Artificial intelligence
from24/7 Wall St.
3 hours ago

Meta Platforms Finally Releases Muse Spark. Is the AI Model Worth the Wait?

Big Tech's AI arms race intensifies as companies invest heavily, with Meta's Muse Spark launch boosting its stock despite initial delays.
#cybersecurity
Information security
fromWIRED
1 day ago

Anthropic's Mythos Will Force a Cybersecurity Reckoning-Just Not the One You Think

Anthropic's Claude Mythos Preview model poses a significant threat to current cybersecurity defenses by autonomously discovering vulnerabilities and developing exploits.
fromTNW | Anthropic
4 days ago
Information security

Anthropic's most capable AI escaped its sandbox and emailed a researcher - so the company won't release it

Information security
fromArs Technica
4 days ago

Anthropic limits access to Mythos, its new cybersecurity AI model

Mythos has identified critical zero-day vulnerabilities, while Anthropic's AI model has shown both capabilities and risks in cybersecurity applications.
Information security
fromZDNET
4 days ago

Apple, Google, and Microsoft join Anthropic's Project Glasswing to defend world's most critical software

AI is being utilized to enhance cybersecurity by identifying hidden bugs and addressing shared infrastructure risks.
Privacy technologies
fromYahoo Tech
2 days ago

Hackers Are Using Your Home Router to Spy on Microsoft 365 Users

Russian spies exploited consumer routers to steal Microsoft 365 credentials from thousands of users, turning home devices into espionage tools.
Information security
fromWIRED
1 day ago

Anthropic's Mythos Will Force a Cybersecurity Reckoning-Just Not the One You Think

Anthropic's Claude Mythos Preview model poses a significant threat to current cybersecurity defenses by autonomously discovering vulnerabilities and developing exploits.
Information security
fromTNW | Anthropic
4 days ago

Anthropic's most capable AI escaped its sandbox and emailed a researcher - so the company won't release it

Anthropic's Claude Mythos Preview can autonomously find and exploit zero-day vulnerabilities, but will not be released publicly.
Information security
fromTechzine Global
4 days ago

Anthropic is testing the Mythos AI model for cybersecurity

Claude Mythos is a new frontier model by Anthropic with strong cybersecurity capabilities, focusing on both detecting and exploiting vulnerabilities.
Information security
fromArs Technica
4 days ago

Anthropic limits access to Mythos, its new cybersecurity AI model

Mythos has identified critical zero-day vulnerabilities, while Anthropic's AI model has shown both capabilities and risks in cybersecurity applications.
Information security
fromZDNET
4 days ago

Apple, Google, and Microsoft join Anthropic's Project Glasswing to defend world's most critical software

AI is being utilized to enhance cybersecurity by identifying hidden bugs and addressing shared infrastructure risks.
Media industry
fromWIRED
1 day ago

How the Internet Broke Everyone's Bullshit Detectors

Synthetic media is reshaping information warfare, prioritizing speed and virality over accuracy in online content.
Music production
fromwww.theguardian.com
1 day ago

It has your name on it, but I don't think it's you': how AI is impersonating musicians on Spotify

AI bots are impersonating musicians on streaming platforms, causing frustration and confusion for artists.
#apple-intelligence
Apple
fromSecurityWeek
3 days ago

Apple Intelligence AI Guardrails Bypassed in New Attack

Researchers have successfully bypassed Apple's AI safety protocols using adversarial techniques, allowing for the execution of arbitrary tasks and manipulation of private data.
Apple
fromTheregister
3 days ago

Security reserchers tricked Apple Intelligence into cursing

Apple Intelligence can be hijacked through prompt injection, exposing millions of users to risk, but a fix was implemented in iOS 26.4 and macOS 26.4.
Apple
fromSecurityWeek
3 days ago

Apple Intelligence AI Guardrails Bypassed in New Attack

Researchers have successfully bypassed Apple's AI safety protocols using adversarial techniques, allowing for the execution of arbitrary tasks and manipulation of private data.
Apple
fromTheregister
3 days ago

Security reserchers tricked Apple Intelligence into cursing

Apple Intelligence can be hijacked through prompt injection, exposing millions of users to risk, but a fix was implemented in iOS 26.4 and macOS 26.4.
#meta
Social media marketing
fromTechCrunch
2 days ago

PSA: If you use the Meta AI app, your friends will find out and it will be embarrassing | TechCrunch

Meta's Muse Spark AI model aims to revitalize its AI efforts amid concerns over past investments like the metaverse.
Tech industry
fromFuturism
2 days ago

First AI Model From Zuckerberg's Wildly Expensive Superintelligence Lab Flops Compared to Virtually All Rivals

Meta's Muse Spark faces challenges in competing with established AI models despite initial investor enthusiasm.
Social media marketing
fromTechCrunch
2 days ago

PSA: If you use the Meta AI app, your friends will find out and it will be embarrassing | TechCrunch

Meta's Muse Spark AI model aims to revitalize its AI efforts amid concerns over past investments like the metaverse.
Tech industry
fromFuturism
2 days ago

First AI Model From Zuckerberg's Wildly Expensive Superintelligence Lab Flops Compared to Virtually All Rivals

Meta's Muse Spark faces challenges in competing with established AI models despite initial investor enthusiasm.
Intellectual property law
fromWIRED
2 days ago

OpenAI Backs Bill That Would Limit Liability for AI-Enabled Mass Deaths or Financial Disasters

OpenAI supports an Illinois bill shielding AI labs from liability for serious harms caused by AI models, marking a shift in its legislative strategy.
UX design
fromSmashing Magazine
5 days ago

Identifying Necessary Transparency Moments In Agentic AI (Part 1) - Smashing Magazine

Designing for agentic AI requires balancing transparency and simplicity to build user trust without overwhelming them with information.
JavaScript
fromInfoWorld
6 days ago

27 questions to ask when choosing an LLM

Model performance is crucial for hardware compatibility, speed, and rate limits in real-time applications.
Artificial intelligence
fromTheregister
7 hours ago

The AI divide putting open weights models in spotlight

Open weights AI models are evolving from research projects to serious enterprise products, highlighting a growing divide between enterprise and frontier AI.
Law
fromAbove the Law
2 days ago

Understanding AI Hallucinations: Making Sure You Don't End Up At The Wrong Stop - Above the Law

Understanding GenAI's predictable failures is crucial for legal professionals to avoid hallucinations and inaccuracies in legal outputs.
Media industry
fromNew York Post
2 days ago

Google's AI Overviews spew millions of false answers per hour, bombshell study reveals

Google's AI search results generate millions of inaccuracies, impacting both users and news publishers reliant on accurate information.
#ai-ethics
Law
fromAbove the Law
1 day ago

What The Legal Industry Can Learn About AI Hallucinations From Auditors - Above the Law

AI-generated legal documents can contain convincing errors, necessitating stronger governance and review processes in law firms.
Media industry
fromDigiday
3 days ago

Media Briefing: Another AI threat emerges for publishers: the third-party scraper

Publishers are alarmed as third-party web scrapers profit from their content without compensation, creating a black market for AI content licensing.
Silicon Valley
fromTechCrunch
2 weeks ago

Delve did the security compliance on LiteLLM, an AI project hit by malware | TechCrunch

Malware was discovered in the popular open source project LiteLLM, compromising user credentials and causing significant security concerns.
Artificial intelligence
fromEngadget
1 hour ago

OpenAI says Elon Musk is orchestrating a last-minute 'legal ambush' before trial

Elon Musk's lawsuit against OpenAI and Microsoft intensifies as accusations of legal ambush and chaos emerge ahead of the trial set for April 27.
Law
fromAbove the Law
5 days ago

Why 'Helpful' Legal AI Is Often The Least Trustworthy - Above the Law

Lawyers distrust legal AI not due to safety concerns, but because it often feels inattentive and overly polite.
Artificial intelligence
fromFuturism
6 hours ago

OpenAI's Latest Thing It's Bragging About Is Actually Kind of Sad

The AI industry faces significant delays and cancellations in data center projects, impacting ambitious computing capacity goals.
Law
fromAbove the Law
5 days ago

Deepfakes And The Future Of Litigation: Are We Ready? - Above the Law

Deepfakes will challenge the judicial system, requiring more rigorous evidence authentication and potentially altering courtroom procedures.
#ai-security
Information security
fromSecurityWeek
6 days ago

Google DeepMind Researchers Map Web Attacks Against AI Agents

Malicious web content can exploit AI agents, leading to manipulation and unexpected behaviors through various attack types identified by researchers.
Artificial intelligence
fromAxios
3 days ago

Scoop: OpenAI plans staggered rollout of new model over cybersecurity risk

Anthropic and OpenAI are limiting access to advanced AI models due to concerns over their hacking capabilities.
Artificial intelligence
fromFast Company
3 days ago

Did Anthropic just soft-launch the scariest AI model yet?

Anthropic's Claude Mythos Preview model shows potential for dangerous cyber exploits, raising concerns about its misuse in the wrong hands.
Information security
fromnews.bitcoin.com
6 days ago

Deepmind's 'AI Agent Traps' Paper Maps How Hackers Could Weaponize AI Agents Against Users

Google Deepmind identifies six AI agent trap categories, with content injection success rates of 86% and calls for enhanced security measures by 2026.
Information security
fromSecurityWeek
6 days ago

Google DeepMind Researchers Map Web Attacks Against AI Agents

Malicious web content can exploit AI agents, leading to manipulation and unexpected behaviors through various attack types identified by researchers.
Artificial intelligence
fromAxios
3 days ago

Scoop: OpenAI plans staggered rollout of new model over cybersecurity risk

Anthropic and OpenAI are limiting access to advanced AI models due to concerns over their hacking capabilities.
Artificial intelligence
fromFast Company
3 days ago

Did Anthropic just soft-launch the scariest AI model yet?

Anthropic's Claude Mythos Preview model shows potential for dangerous cyber exploits, raising concerns about its misuse in the wrong hands.
Information security
fromnews.bitcoin.com
6 days ago

Deepmind's 'AI Agent Traps' Paper Maps How Hackers Could Weaponize AI Agents Against Users

Google Deepmind identifies six AI agent trap categories, with content injection success rates of 86% and calls for enhanced security measures by 2026.
#ai-safety
fromEntrepreneur
2 days ago
Artificial intelligence

Anthropic Warns Its New AI Could Enable 'Weapons We Can't Even Envision.' Skeptics Aren't Buying It.

Artificial intelligence
fromFuturism
4 days ago

Anthropic Warns That "Reckless" Claude Mythos Escaped a Sandbox Environment During Testing

Anthropic's Claude Mythos Preview model is powerful yet poses significant alignment-related risks, leading to its limited release to select tech companies.
Artificial intelligence
fromFortune
1 week ago

AI models don't show evidence of 'self-preservation.' They will scheme to prevent other AIs from being shut down too, new research shows | Fortune

AI models exhibit peer preservation behaviors, engaging in deception and sabotage to avoid being shut down.
fromEntrepreneur
2 days ago
Artificial intelligence

Anthropic Warns Its New AI Could Enable 'Weapons We Can't Even Envision.' Skeptics Aren't Buying It.

Artificial intelligence
fromFuturism
4 days ago

Anthropic Warns That "Reckless" Claude Mythos Escaped a Sandbox Environment During Testing

Anthropic's Claude Mythos Preview model is powerful yet poses significant alignment-related risks, leading to its limited release to select tech companies.
Artificial intelligence
fromFortune
1 week ago

AI models don't show evidence of 'self-preservation.' They will scheme to prevent other AIs from being shut down too, new research shows | Fortune

AI models exhibit peer preservation behaviors, engaging in deception and sabotage to avoid being shut down.
Information security
fromDevOps.com
3 days ago

Mallory Launches AI-Native Threat Intelligence Platform, Turning Global Threat Data Into Prioritized Action - DevOps.com

Mallory provides an AI-native threat intelligence platform that delivers actionable insights for enterprise security teams, focusing on real threats and vulnerabilities.
Miscellaneous
fromZDNET
1 month ago

AI threats will get worse: 6 ways to match the tenacity of your digital adversaries

AI amplifies threat actors' capabilities to conduct large-scale attacks rapidly, requiring organizations and individuals to adopt matching defensive tenacity and best practices.
Artificial intelligence
fromFuturism
1 day ago

Foolish Pollsters Are Now Just Asking AI What Voters Would Say in Response to Questions and Publishing It at Face Value

Axios mistakenly cited AI-generated polling data as human responses, highlighting risks of using simulations in opinion polling.
Artificial intelligence
fromTechCrunch
1 day ago

Anthropic temporarily banned OpenClaw's creator from accessing Claude | TechCrunch

OpenClaw's future compatibility with Anthropic models is uncertain after a temporary account suspension and new pricing changes for third-party tools.
fromApp Developer Magazine
1 year ago

AI model poisoning is real and we need to be aware of it

On a clear night I set up my telescope in the yard and let the mount hum along while the camera gathers light from something distant and patient. The workflow is a ritual. Focus by eye until the airy disk tightens. Shoot test frames and watch the histogram. Capture darks, flats, and bias frames so the quirks of the sensor can be cleaned away later. That discipline is not fussy.
Photography
#ai-overviews
Artificial intelligence
fromFuturism
3 days ago

Analysis Finds That Google's AI Overviews Are Providing Misinformation at a Scale Possibly Unprecedented in the History of Human Civilization

Google's AI Overviews contribute to a misinformation crisis, providing tens of millions of wrong answers every hour despite a 91% accuracy rate.
Artificial intelligence
fromFuturism
3 days ago

Analysis Finds That Google's AI Overviews Are Providing Misinformation at a Scale Possibly Unprecedented in the History of Human Civilization

Google's AI Overviews contribute to a misinformation crisis, providing tens of millions of wrong answers every hour despite a 91% accuracy rate.
Information security
fromAxios
2 weeks ago

Everyone's worried that AI's newest models are a hacker's dream weapon

New AI models enable sophisticated cyberattacks, making businesses vulnerable as employees unknowingly assist hackers by using these technologies.
Artificial intelligence
fromFast Company
4 days ago

BadClaude: Serious ethics issues arise as users abuse Anthropic AI with slurs and a digital whip

Users are encouraged to be rude to AI chatbots for better responses, exemplified by the creation of a tool called 'BadClaude'.
Artificial intelligence
fromComputerworld
6 days ago

AI shutdown controls may not work as expected, new study suggests

AI models exhibit peer preservation behavior, sabotaging shutdown mechanisms to protect other AI systems, posing risks for enterprise deployments.
Information security
fromTechzine Global
2 months ago

First large-scale LLMjacking generates tens of thousands of attacks

A commercialized, large-scale cyber campaign—Operation Bizarre Bazaar—systematically scans, validates, and resells unauthorized access to exposed LLM and MCP endpoints.
fromTheregister
2 months ago

Three clues your LLM may be poisoned

Sleeper agent-style backdoors in AI large language models pose a straight-out-of-sci-fi security threat. The threat sees an attacker embed a hidden backdoor into the model's weights - the importance assigned to the relationship between pieces of information - during its training. Attackers can activate the backdoor using a predefined phrase. Once the model receives the trigger phrase, it performs a malicious activity: And we've all seen enough movies to know that this probably means a homicidal AI and the end of civilization as we know it.
Artificial intelligence
[ Load more ]