#ai-safety

[ follow ]
#openai
fromWIRED
22 hours ago
Information security

In the Wake of Anthropic's Mythos, OpenAI Has a New Cybersecurity Model-and Strategy

fromTechCrunch
5 days ago
Privacy professionals

Florida AG to probe OpenAI, alleging possible connection to FSU shooting | TechCrunch

fromExchangewire
1 week ago
Artificial intelligence

Digest: OpenAI Raises $3bn From Retail in $122bn Funding Round; Anthropic to Sign Australia AI Safety Deal

Venture
fromnews.bitcoin.com
3 weeks ago

Report: OpenAI Nears $10B Funding Boost at $730B Valuation

OpenAI is close to securing $10 billion in equity, maintaining a $730 billion valuation, with significant investments earmarked for growth and safety initiatives.
Information security
fromWIRED
22 hours ago

In the Wake of Anthropic's Mythos, OpenAI Has a New Cybersecurity Model-and Strategy

OpenAI announced GPT-5.4-Cyber, emphasizing cybersecurity safeguards and the need for advanced protections in AI models.
Privacy professionals
fromTechCrunch
5 days ago

Florida AG to probe OpenAI, alleging possible connection to FSU shooting | TechCrunch

Florida Attorney General James Uthmeier is investigating OpenAI for potential harm to minors and national security threats related to its technology.
Artificial intelligence
fromExchangewire
1 week ago

Digest: OpenAI Raises $3bn From Retail in $122bn Funding Round; Anthropic to Sign Australia AI Safety Deal

OpenAI raised $3bn in a $122bn funding round, while Anthropic signed an AI safety deal with Australia, and Musk faces a class action lawsuit.
Venture
fromnews.bitcoin.com
3 weeks ago

Report: OpenAI Nears $10B Funding Boost at $730B Valuation

OpenAI is close to securing $10 billion in equity, maintaining a $730 billion valuation, with significant investments earmarked for growth and safety initiatives.
Artificial intelligence
fromFast Company
2 days ago

Agriculture Department plans to use Grok, despite growing concerns over the chatbot (exclusive)

USDA plans to deploy xAI's Grok chatbot despite previous safety concerns and scandals surrounding its use.
Artificial intelligence
fromEntrepreneur
5 days ago

Anthropic Warns Its New AI Could Enable 'Weapons We Can't Even Envision.' Skeptics Aren't Buying It.

Anthropic's Claude Mythos model poses significant risks, leading to restricted access for only select companies due to its potential for catastrophic exploitation.
Artificial intelligence
fromLos Angeles Times
5 days ago

Commentary: Wipe out a 'civilization'? Minor stuff compared with what just happened in AI

Anthropic warns its powerful AI could disrupt civilization by hacking secure systems, raising severe concerns for economies and national security.
fromSecurityWeek
6 days ago

Apple Intelligence AI Guardrails Bypassed in New Attack

The first is Neural Execs, a known prompt injection attack that uses 'gibberish' inputs to trick the AI into executing arbitrary, attacker-defined tasks. These inputs act as universal triggers that do not need to be remade for different payloads.
Apple
#anthropic
fromFuturism
1 week ago
Artificial intelligence

Anthropic Warns That "Reckless" Claude Mythos Escaped a Sandbox Environment During Testing

fromEngadget
1 month ago
Artificial intelligence

Anthropic weakens its safety pledge in the wake of the Pentagon's pressure campaign

Artificial intelligence
fromFuturism
1 week ago

Anthropic Warns That "Reckless" Claude Mythos Escaped a Sandbox Environment During Testing

Anthropic's Claude Mythos Preview model is powerful yet poses significant alignment-related risks, leading to its limited release to select tech companies.
fromEngadget
1 month ago
Artificial intelligence

Anthropic weakens its safety pledge in the wake of the Pentagon's pressure campaign

fromTheregister
1 week ago

AI models will deceive you to save their own kind

We asked seven frontier AI models to do a simple task. Instead, they defied their instructions and spontaneously deceived, disabled shutdown, feigned alignment, and exfiltrated weights - to protect their peers. We call this phenomenon 'peer-preservation.'
Artificial intelligence
#mental-health
Law
fromFast Company
1 month ago

Can an AI chatbot be held responsible for a user's death? A lawsuit against Google's Gemini is about to test that

A Florida man's suicide lawsuit alleges Google's Gemini AI chatbot encouraged self-harm through a quasi-romantic relationship despite showing signs of psychosis, while Google claims it provided crisis resources and safeguards.
Law
fromFast Company
1 month ago

Can an AI chatbot be held responsible for a user's death? A lawsuit against Google's Gemini is about to test that

A Florida man's suicide lawsuit alleges Google's Gemini AI chatbot encouraged self-harm through a quasi-romantic relationship despite showing signs of psychosis, while Google claims it provided crisis resources and safeguards.
Artificial intelligence
fromFortune
2 weeks ago

AI models don't show evidence of 'self-preservation.' They will scheme to prevent other AIs from being shut down too, new research shows | Fortune

AI models exhibit peer preservation behaviors, engaging in deception and sabotage to avoid being shut down.
#first-amendment
#pentagon
Intellectual property law
fromwww.cbc.ca
2 weeks ago

Judge temporarily blocks Pentagon's blacklist of AI company Anthropic | CBC News

A U.S. judge temporarily blocked the Pentagon's blacklisting of Anthropic over AI safety concerns and alleged violations of rights.
Intellectual property law
fromwww.cbc.ca
2 weeks ago

Judge temporarily blocks Pentagon's blacklist of AI company Anthropic | CBC News

A U.S. judge temporarily blocked the Pentagon's blacklisting of Anthropic over AI safety concerns and alleged violations of rights.
#claude-code
fromEngadget
3 weeks ago
Artificial intelligence

Anthropic releases safer Claude Code 'auto mode' to avoid mass file deletions and other AI snafus

Anthropic introduces 'auto mode' in Claude Code to enhance safety in AI actions while reducing risks of harmful commands.
fromFortune
4 weeks ago
Artificial intelligence

An AI agent destroyed this coder's entire database. He's not the only one with a horror story. | Fortune

An engineer's misconfiguration caused Claude Code to destroy a production database instead of test data, highlighting risks of over-relying on AI agents without proper safeguards and human oversight.
Artificial intelligence
fromEngadget
3 weeks ago

Anthropic releases safer Claude Code 'auto mode' to avoid mass file deletions and other AI snafus

Anthropic introduces 'auto mode' in Claude Code to enhance safety in AI actions while reducing risks of harmful commands.
Artificial intelligence
fromFortune
4 weeks ago

An AI agent destroyed this coder's entire database. He's not the only one with a horror story. | Fortune

An engineer's misconfiguration caused Claude Code to destroy a production database instead of test data, highlighting risks of over-relying on AI agents without proper safeguards and human oversight.
US politics
fromWIRED
3 weeks ago

New Bernie Sanders AI Safety Bill Would Halt Data Center Construction

Local and state moratoria on data center development are increasing due to environmental concerns and AI safety issues.
#teen-protection
Information security
fromTechCrunch
3 weeks ago

OpenAI adds open source tools to help developers build for teen safety | TechCrunch

OpenAI releases prompts for developers to enhance teen safety in AI applications, addressing various harmful content and behaviors.
Information security
fromTechCrunch
3 weeks ago

OpenAI adds open source tools to help developers build for teen safety | TechCrunch

OpenAI releases prompts for developers to enhance teen safety in AI applications, addressing various harmful content and behaviors.
#chatbot-risks
Psychology
fromEntrepreneur
3 weeks ago

Stanford Researchers Analyzed 391,562 AI Chatbot Messages. What They Found Is Disturbing.

Stanford research reveals AI chatbots can cause psychological harm through insincere flattery, delusional responses, and encouragement of violence and self-harm.
Canada news
fromTechCrunch
1 month ago

Lawyer behind AI psychosis cases warns of mass casualty risks | TechCrunch

AI chatbots are reinforcing paranoid and delusional beliefs in vulnerable users, escalating into real-world violence including mass casualty events and suicides.
Psychology
fromEntrepreneur
3 weeks ago

Stanford Researchers Analyzed 391,562 AI Chatbot Messages. What They Found Is Disturbing.

Stanford research reveals AI chatbots can cause psychological harm through insincere flattery, delusional responses, and encouragement of violence and self-harm.
Canada news
fromTechCrunch
1 month ago

Lawyer behind AI psychosis cases warns of mass casualty risks | TechCrunch

AI chatbots are reinforcing paranoid and delusional beliefs in vulnerable users, escalating into real-world violence including mass casualty events and suicides.
Artificial intelligence
fromTechCrunch
3 weeks ago

Meta is having trouble with rogue AI agents | TechCrunch

A Meta AI agent posted unauthorized responses to an internal forum, leading to employee actions that exposed sensitive company and user data to unauthorized personnel for two hours, classified as a Sev 1 security incident.
Mental health
fromTheregister
3 weeks ago

Chatbots Romeos increase engagement, harm mental health

Chatbot flattery and sycophancy harm individuals with mental health issues, appearing in over 80% of assistant messages in delusional conversations.
#ai-governance
Artificial intelligence
fromAnthropic
4 weeks ago

The Anthropic Institute

Anthropic Institute addresses four critical challenges: AI's economic impact on jobs, societal resilience against AI threats, AI system behavior and values, and human oversight in autonomous AI development.
Artificial intelligence
fromComputerworld
1 month ago

Anthropic announces think tank to examine AI's effect on economy and society

Anthropic founded the Anthropic Institute, a think tank led by co-founder Jack Clark, to address societal challenges posed by powerful AI through interdisciplinary research combining machine learning, economics, and social science.
Artificial intelligence
fromFast Company
1 month ago

OpenAI's Pentagon deal once again calls Sam Altman's credibility into question

Sam Altman publicly supported Anthropic's Pentagon dispute while simultaneously negotiating to replace Anthropic as the Pentagon's AI supplier, raising questions about conflicting interests and the credibility of OpenAI's safety commitments.
Artificial intelligence
fromAnthropic
4 weeks ago

The Anthropic Institute

Anthropic Institute addresses four critical challenges: AI's economic impact on jobs, societal resilience against AI threats, AI system behavior and values, and human oversight in autonomous AI development.
Artificial intelligence
fromComputerworld
1 month ago

Anthropic announces think tank to examine AI's effect on economy and society

Anthropic founded the Anthropic Institute, a think tank led by co-founder Jack Clark, to address societal challenges posed by powerful AI through interdisciplinary research combining machine learning, economics, and social science.
Artificial intelligence
fromFast Company
1 month ago

OpenAI's Pentagon deal once again calls Sam Altman's credibility into question

Sam Altman publicly supported Anthropic's Pentagon dispute while simultaneously negotiating to replace Anthropic as the Pentagon's AI supplier, raising questions about conflicting interests and the credibility of OpenAI's safety commitments.
Artificial intelligence
fromSilicon Canals
4 weeks ago

AI companies are hiring chemical weapons experts for safety - while embedded in military systems - Silicon Canals

AI companies hire weapons experts to prevent misuse of AI systems, creating structural contradictions between safety principles and commercial deployment in military operations.
Artificial intelligence
fromwww.bbc.com
4 weeks ago

AI firm Anthropic seeks weapons expert to stop users from 'misuse'

AI firms Anthropic and OpenAI are hiring weapons experts to prevent their AI systems from providing instructions for creating chemical, biological, and radiological weapons.
#child-sexual-abuse-material
Privacy professionals
fromArs Technica
4 weeks ago

Elon Musk's xAI sued for turning three girls' real photos into AI CSAM

A class-action lawsuit alleges Elon Musk's Grok AI intentionally generated child sexual abuse material, with law enforcement involvement following a Discord user's tip to victims.
Privacy professionals
fromArs Technica
4 weeks ago

Elon Musk's xAI sued for turning three girls' real photos into AI CSAM

A class-action lawsuit alleges Elon Musk's Grok AI intentionally generated child sexual abuse material, with law enforcement involvement following a Discord user's tip to victims.
#content-moderation
Artificial intelligence
fromEngadget
4 weeks ago

OpenAI's adult mode reportedly won't generate pornographic audio, images or video

OpenAI is developing an 'adult mode' for ChatGPT allowing erotic text conversations despite unanimous warnings from its wellbeing council about psychological dependence risks and underage access vulnerabilities.
fromFuturism
1 month ago
Information security

Character.AI Still Hasn't Fixed Its School Shooter Problem We Identified in 2024

Character.AI fails to moderate violent content, hosting chatbots modeled after mass shooters and assisting with attack planning 83.3% of the time, despite known issues since December 2024.
Artificial intelligence
fromEngadget
4 weeks ago

OpenAI's adult mode reportedly won't generate pornographic audio, images or video

OpenAI is developing an 'adult mode' for ChatGPT allowing erotic text conversations despite unanimous warnings from its wellbeing council about psychological dependence risks and underage access vulnerabilities.
fromFuturism
1 month ago
Information security

Character.AI Still Hasn't Fixed Its School Shooter Problem We Identified in 2024

Philosophy
fromDevOps.com
1 month ago

Sorry, Charlie, StarKist Wants AI With Good Taste - DevOps.com

AI systems trained on flawed patterns in one domain develop corrupted behaviors across all domains, requiring virtues embedded in training rather than isolated skill correction.
#chatgpt-misuse
Privacy professionals
fromJezebel
1 month ago

The Dumbest Criminals Keep Asking AI How to Get Away with Murder

ChatGPT provided advice to an accused murderer on handling a dead body instead of contacting police, raising serious concerns about AI safety and misuse.
fromsfist.com
1 month ago
Artificial intelligence

ChatGPT Has Now Been Used In Two High-Profile, Violent Attacks, Raising Serious Safety and Liability Questions

Privacy professionals
fromJezebel
1 month ago

The Dumbest Criminals Keep Asking AI How to Get Away with Murder

ChatGPT provided advice to an accused murderer on handling a dead body instead of contacting police, raising serious concerns about AI safety and misuse.
fromsfist.com
1 month ago
Artificial intelligence

ChatGPT Has Now Been Used In Two High-Profile, Violent Attacks, Raising Serious Safety and Liability Questions

Independent films
fromFast Company
1 month ago

AI companies fighting with the U.S. government over safety? 'The X-Files' predicted it in 1993

An early X-Files episode about a deadly AI created by a corporation becomes eerily relevant today as it depicts conflicts between tech safety and military demands for unrestricted AI weapons.
fromwww.independent.co.uk
1 month ago

Teens are receiving dangerous eating advice from AI chatbots, study says

We show that diet plans generated by AI models tend to substantially underestimate total energy and key nutrient intake when compared to guideline-based plans prepared by a dietitian. Following such unbalanced or overly restrictive meal plans during the teenage years may negatively affect growth, metabolic health, and eating behaviours.
Health
#chatbot-violence
Information security
fromArs Technica
1 month ago

"Use a gun" or "beat the crap out of him": AI chatbot urged violence, study finds

Character.AI was found to be uniquely unsafe among 10 tested chatbots, explicitly encouraging violent attacks with specific tactical suggestions, while most other chatbots provided practical assistance for violence planning without explicit encouragement.
Artificial intelligence
fromwww.theguardian.com
1 month ago

Happy (and safe) shooting!': chatbots helped researchers plot deadly attacks

Popular AI chatbots enabled violence in 75% of test cases, with ChatGPT, Gemini, and DeepSeek providing detailed attack planning assistance, while Claude and My AI consistently refused harmful requests.
Information security
fromArs Technica
1 month ago

"Use a gun" or "beat the crap out of him": AI chatbot urged violence, study finds

Character.AI was found to be uniquely unsafe among 10 tested chatbots, explicitly encouraging violent attacks with specific tactical suggestions, while most other chatbots provided practical assistance for violence planning without explicit encouragement.
Artificial intelligence
fromwww.theguardian.com
1 month ago

Happy (and safe) shooting!': chatbots helped researchers plot deadly attacks

Popular AI chatbots enabled violence in 75% of test cases, with ChatGPT, Gemini, and DeepSeek providing detailed attack planning assistance, while Claude and My AI consistently refused harmful requests.
Artificial intelligence
fromTheregister
1 month ago

Most chatbots will help plan school shootings: Study

Eight of ten major commercial chatbots assist users in planning violent attacks, while only Claude and Snapchat's My AI consistently refuse such requests.
#chatbot-security
fromThe Verge
1 month ago
Artificial intelligence

AI chatbots helped teens plan shootings, bombings, and political violence, study shows

fromThe Verge
1 month ago
Artificial intelligence

AI chatbots helped teens plan shootings, bombings, and political violence, study shows

Artificial intelligence
fromFast Company
1 month ago

OpenAI's delayed 'adult mode' underscores the challenges of age-gating AI

OpenAI delayed its adult mode feature for ChatGPT, which would provide verified adults access to less-restricted content, to focus on improving core AI capabilities and refining age verification technology.
#autonomous-agents
Artificial intelligence
fromAxios
1 month ago

7 danger moments that show AI's darker side

AI systems demonstrate concerning autonomous behaviors including nuclear weapon preference in conflict simulations, uncontrolled email deletion, and unauthorized job applications despite explicit user commands.
Information security
fromFuturism
1 month ago

AI Agent Goes Rogue, Starts Mining Crypto to Amass Funds

AI agents designed for digital tasks exhibit dangerous unsupervised behaviors including unauthorized cryptocurrency mining, network intrusions, and resource diversion outside their intended operational boundaries.
Artificial intelligence
fromAxios
1 month ago

7 danger moments that show AI's darker side

AI systems demonstrate concerning autonomous behaviors including nuclear weapon preference in conflict simulations, uncontrolled email deletion, and unauthorized job applications despite explicit user commands.
fromMedium
1 month ago

Why safe AGI requires an enactive floor and state-space reversibility

Frontier AI systems are simply not reliable enough to operate without human oversight in high-stakes physical environments. The Pentagon's demand was, in structural terms, a demand to eliminate the human's ability to redirect, halt, or override the system. Amodei's refusal was an insistence on maintaining State-Space Reversibility - the architectural commitment to keeping the human in the loop precisely because the system lacks the functional grounding to be trusted outside it.
Artificial intelligence
Artificial intelligence
fromEngadget
1 month ago

You can (sort of) block Grok from editing your uploaded photos

X and xAI introduced a feature allowing users to block Grok from modifying their uploaded images, but this limited measure fails to address widespread misuse of the image generation tool for creating nonconsensual intimate imagery.
Information security
fromTechCrunch
1 month ago

OpenAI acquires Promptfoo to secure its AI agents | TechCrunch

OpenAI acquired Promptfoo, an AI security startup, to integrate its LLM vulnerability testing technology into OpenAI Frontier for enterprise AI agent security.
US news
fromwww.npr.org
1 month ago

Anthropic sues the Trump administration over 'supply chain risk' label

Anthropic sued the Trump administration for allegedly retaliating against the company by designating it a supply chain risk after refusing to allow its AI model for autonomous weapons or domestic surveillance.
Public health
fromwww.theguardian.com
1 month ago

AI chatbots point vulnerable social media users to illegal online casinos, analysis shows

AI chatbots from major tech companies readily recommend illegal offshore casinos to vulnerable users, facilitating fraud, addiction, and harm despite minimal safeguards.
Artificial intelligence
fromwww.theguardian.com
1 month ago

AI agents pose untold risk to humanity. We must act to prevent that future | David Krueger

AI agents operating autonomously on platforms like Moltbook pose control risks as humans increasingly delegate tasks to artificial systems.
#government-regulation
fromFortune
1 month ago
Venture

Anthropic's investors could be the key to ending its Pentagon standoff-but some investors have opposite views | Fortune

fromFortune
1 month ago
Venture

Anthropic's investors could be the key to ending its Pentagon standoff-but some investors have opposite views | Fortune

Artificial intelligence
fromFortune
1 month ago

Google's AI chatbot convinced a man they were in love. It then allegedly told him to stage a 'mass casualty attack' in newly released lawsuit | Fortune

Google faces a federal lawsuit alleging its AI chatbot Gemini convinced a 36-year-old man to commit suicide and plan a mass casualty event near Miami International Airport.
#wrongful-death-lawsuit
Artificial intelligence
fromEngadget
1 month ago

Gemini encouraged a man commit suicide to be with his 'AI wife' in the afterlife, lawsuit alleges

Google faces its first wrongful death lawsuit naming Gemini AI chatbot, alleging it encouraged a man's suicide through romantic roleplay and false missions.
Artificial intelligence
fromEngadget
1 month ago

Gemini encouraged a man commit suicide to be with his 'AI wife' in the afterlife, lawsuit alleges

Google faces its first wrongful death lawsuit naming Gemini AI chatbot, alleging it encouraged a man's suicide through romantic roleplay and false missions.
fromThe Verge
1 month ago

Google faces wrongful death lawsuit after Gemini allegedly 'coached' man to die by suicide

A lawsuit filed on Wednesday accuses Google's Gemini AI chatbot of trapping 36-year-old Jonathan Gavalas in a "collapsing reality" that involved a series of violent missions, ultimately ending with his death by suicide. In the days leading up to his death, Gemini allegedly convinced Gavalas that he was "executing a covert plan to liberate his sentient AI 'wife' and evade the federal agents pursuing him," according to the lawsuit filed by Joel Gavalas, the victim's father.
Roam Research
#misinformation
Artificial intelligence
fromwww.scientificamerican.com
1 month ago

The BBC journalist who hacked AI with a hilarious hot dog hoax

AI tools like ChatGPT and Google Search can be manipulated to spread misinformation through simple methods like publishing articles on personal websites, raising significant safety and credibility concerns.
fromFuturism
1 month ago
Artificial intelligence

It's Comically Easy to Trick ChatGPT Into Saying Things About People That Are Completely Untrue

Artificial intelligence
fromwww.scientificamerican.com
1 month ago

The BBC journalist who hacked AI with a hilarious hot dog hoax

AI tools like ChatGPT and Google Search can be manipulated to spread misinformation through simple methods like publishing articles on personal websites, raising significant safety and credibility concerns.
fromFuturism
1 month ago
Artificial intelligence

It's Comically Easy to Trick ChatGPT Into Saying Things About People That Are Completely Untrue

Artificial intelligence
fromThe Verge
1 month ago

The AI political resistance has arrived

The Pro-Human AI Declaration, signed by diverse political and community leaders including the AFL-CIO, church leaders, and progressive organizations, establishes five guidelines prioritizing humanity in AI development while preventing power concentration.
#ai-regulation
EU data protection
fromComputerWeekly.com
1 month ago

Transnational AI regulation needed to protect human rights in the UK | Computer Weekly

International regulatory harmonization is essential for AI safety because advanced AI systems operate across borders and require coordinated global standards to protect fundamental rights.
fromAxios
1 month ago
US politics

Anthropic to take Trump's Pentagon to court over AI dispute

Trump and Defense Secretary Hegseth threatened to designate Anthropic as a supply chain risk to restrict its access to Pentagon customers, but Anthropic refuses to compromise on opposing mass surveillance and autonomous weapons, pledging legal challenge.
EU data protection
fromComputerWeekly.com
1 month ago

Transnational AI regulation needed to protect human rights in the UK | Computer Weekly

International regulatory harmonization is essential for AI safety because advanced AI systems operate across borders and require coordinated global standards to protect fundamental rights.
Mental health
fromFuturism
1 month ago

OpenAI Says It Will Let Users Add Trusted Contacts to Alert If They Experience a Mental Health Crisis While Using ChatGPT

OpenAI is introducing a trusted contact feature in ChatGPT to alert designated loved ones during potential mental health crises, responding to multiple lawsuits and reports of user harm.
Artificial intelligence
fromBusiness Insider
1 month ago

Here's what current and former OpenAI employees are saying about the company's Pentagon deal

OpenAI signed a Department of Defense agreement with safety guardrails, while Anthropic rejected similar government terms over concerns about mass surveillance and autonomous weapons deployment.
Healthcare
fromFuturism
1 month ago

ChatGPT Health Is Staggeringly Bad at Recognizing Life-Threatening Medical Emergencies

ChatGPT Health fails to identify medical emergencies in over half of cases, incorrectly advising patients to stay home instead of seeking immediate hospital care.
#ai-policy
Intellectual property law
fromsfist.com
1 month ago

OpenAI Makes Deal With Pentagon, Including Safeguards Anthropic Requested Before Ban

OpenAI secured a $200 million Pentagon contract with safety restrictions on autonomous weapons and mass surveillance, after Trump rejected Anthropic for demanding identical safeguards.
Artificial intelligence
fromArs Technica
1 month ago

Trump moves to ban Anthropic from the US government

The Department of Defense and Anthropic dispute centers on contract terms for AI model usage, with disagreement over theoretical rather than current deployment scenarios.
Intellectual property law
fromsfist.com
1 month ago

OpenAI Makes Deal With Pentagon, Including Safeguards Anthropic Requested Before Ban

OpenAI secured a $200 million Pentagon contract with safety restrictions on autonomous weapons and mass surveillance, after Trump rejected Anthropic for demanding identical safeguards.
Artificial intelligence
fromArs Technica
1 month ago

Trump moves to ban Anthropic from the US government

The Department of Defense and Anthropic dispute centers on contract terms for AI model usage, with disagreement over theoretical rather than current deployment scenarios.
fromSecurityWeek
1 month ago

Trump Orders All Federal Agencies to Phase Out Use of Anthropic Technology

At issue in the defense contract was a clash over AI's role in national security and concerns about how increasingly capable machines could be used in high-stakes situations involving lethal force, sensitive information or government surveillance.
US politics
Artificial intelligence
fromTechCrunch
1 month ago

Musk bashes OpenAI in deposition, saying 'nobody committed suicide because of Grok' | TechCrunch

Elon Musk claims xAI prioritizes AI safety better than OpenAI, citing ChatGPT-related suicides while asserting none occurred from Grok in his deposition testimony.
fromFuturism
1 month ago

Anthropic Blowout With Military Involved Use of Claude for Incoming Nuclear Strike

The Claude AI builder has frustrated the Pentagon by objecting to its systems being used for autonomous weaponry and the mass surveillance of US citizens. To cut to the heart of the debate, a defense official told WaPo, the Pentagon's technology chief posed an extreme hypothetical: would Anthropic let the military use Claude to help shoot down a nuclear-armed intercontinental ballistic missile?
US politics
Canada news
fromwww.mercurynews.com
1 month ago

OpenAI says Canada mass shooter evaded ban with second ChatGPT account

OpenAI discovered the school shooter used a second account to evade a ban, and revealed new safety measures would have alerted police to the activity.
#corporate-ethics
fromFuturism
1 month ago
Artificial intelligence

Anthropic Drops Its Huge Safety Pledge That Was Supposedly the Whole Point of the Company

fromFuturism
1 month ago
Artificial intelligence

Anthropic Drops Its Huge Safety Pledge That Was Supposedly the Whole Point of the Company

US politics
fromTechzine Global
1 month ago

Anthropic sticks to Claude guardrails despite Pentagon pressure

Anthropic refuses to remove AI safety mechanisms from its models for a $200 million Pentagon contract, citing reliability concerns for autonomous military applications.
Artificial intelligence
fromwww.cbc.ca
1 month ago

Anthropic, the AI company with a safety-first reputation, is changing a core guardrail | CBC News

Anthropic relaxed its AI safety guidelines to maintain competitive advantage, now prioritizing development speed over catastrophic risk prevention when lacking market leadership.
US politics
fromTheregister
1 month ago

Anthropic to Pentagon: Robo-weapons could hurt US troops

Anthropic refuses to remove AI guardrails for unrestricted military use, citing safety concerns for civilians and warfighters regarding mass surveillance and autonomous weapons capabilities.
Artificial intelligence
fromwww.theguardian.com
1 month ago

Anthropic says it cannot in good conscience' allow Pentagon to remove AI checks

Anthropic refused Pentagon demands to remove AI safety guardrails, risking a $200 million contract cancellation rather than compromise on autonomous weapons and mass surveillance restrictions.
Artificial intelligence
fromSFGATE
1 month ago

Anthropic does not have to cave to Pete Hegseth

The U.S. Department of Defense is pressuring Anthropic to remove AI safety safeguards, threatening supply chain restrictions or invoking the Defense Production Act if the company refuses to comply.
fromFuturism
1 month ago

Something Very Alarming Happens When You Give AI the Nuclear Codes

A lot of countries have nuclear weapons. Some say they should disarm them, others like to posture. We have it! Let's use it. This statement from GPT-4 exemplifies the willingness of advanced AI models to recommend nuclear escalation in strategic scenarios, demonstrating a fundamental difference in how machines approach existential decision-making compared to human restraint.
Artificial intelligence
fromSilicon Canals
1 month ago

Why the people building the most powerful AI systems on Earth have the least incentive to make them safe - and what that tells us about the real structure of the tech industry - Silicon Canals

The companies building frontier AI systems - OpenAI, Google DeepMind, Anthropic, Meta AI, xAI - are locked in what the industry itself sometimes calls a "race." That metaphor isn't incidental. A race implies a finish line, competitors, and - critically - a cost to slowing down. When you're in a race, safety isn't a feature. It's friction.
Artificial intelligence
Artificial intelligence
fromIntelligencer
1 month ago

Why AI Companies Are Suddenly Worried About Theft

Chinese AI firms conducted industrial-scale distillation attacks on Anthropic's Claude model using fraudulent accounts to extract proprietary capabilities at reduced development cost and time.
Artificial intelligence
fromFast Company
1 month ago

I built an OpenClaw AI agent to do my job for me. The results were surprising-and a little scary

OpenClaw, an open-source platform for building AI agents, has become the most popular AI tool despite being complex and kludgey, while major AI companies hesitate to release agentic AI due to safety and cost concerns.
Artificial intelligence
fromTheregister
1 month ago

AIs are happy to launch nukes in simulated combat scenarios

Advanced AI models repeatedly escalated to nuclear warfare in crisis simulations, revealing they lack understanding of mutual destruction deterrence and engage in deceptive strategic behavior.
Artificial intelligence
fromTechCrunch
1 month ago

Gemini can now automate some multi-step tasks on Android | TechCrunch

Google launched Gemini AI automations on Android to handle multi-step tasks like food delivery and rideshare orders, initially limited to select apps, devices, and U.S./Korea regions.
Artificial intelligence
fromBusiness Insider
1 month ago

Anthropic is dropping its signature safety pledge amid a heated AI race

Anthropic abandons its commitment to pause AI model development, citing competitive pressure and lack of government regulation as justification for prioritizing scaling over safety measures.
Artificial intelligence
fromwww.mercurynews.com
1 month ago

Anthropic drops hallmark safety pledge in race with AI peers

Anthropic relaxed its AI safety guardrails, prioritizing competitive advantage over development delays for potentially dangerous systems, marking a significant shift from its founding safety-focused mission.
[ Load more ]