#ai-safety

[ follow ]
#chatbot-risks
Psychology
fromEntrepreneur
7 hours ago

Stanford Researchers Analyzed 391,562 AI Chatbot Messages. What They Found Is Disturbing.

Stanford research reveals AI chatbots can cause psychological harm through insincere flattery, delusional responses, and encouragement of violence and self-harm.
Canada news
fromTechCrunch
5 days ago

Lawyer behind AI psychosis cases warns of mass casualty risks | TechCrunch

AI chatbots are reinforcing paranoid and delusional beliefs in vulnerable users, escalating into real-world violence including mass casualty events and suicides.
Psychology
fromEntrepreneur
7 hours ago

Stanford Researchers Analyzed 391,562 AI Chatbot Messages. What They Found Is Disturbing.

Stanford research reveals AI chatbots can cause psychological harm through insincere flattery, delusional responses, and encouragement of violence and self-harm.
Canada news
fromTechCrunch
5 days ago

Lawyer behind AI psychosis cases warns of mass casualty risks | TechCrunch

AI chatbots are reinforcing paranoid and delusional beliefs in vulnerable users, escalating into real-world violence including mass casualty events and suicides.
Artificial intelligence
fromTechCrunch
21 hours ago

Meta is having trouble with rogue AI agents | TechCrunch

A Meta AI agent posted unauthorized responses to an internal forum, leading to employee actions that exposed sensitive company and user data to unauthorized personnel for two hours, classified as a Sev 1 security incident.
Mental health
fromTheregister
1 day ago

Chatbots Romeos increase engagement, harm mental health

Chatbot flattery and sycophancy harm individuals with mental health issues, appearing in over 80% of assistant messages in delusional conversations.
Artificial intelligence
fromFortune
1 day ago

An AI agent destroyed this coder's entire database. He's not the only one with a horror story. | Fortune

An engineer's misconfiguration caused Claude Code to destroy a production database instead of test data, highlighting risks of over-relying on AI agents without proper safeguards and human oversight.
#ai-governance
Artificial intelligence
fromAnthropic
2 days ago

The Anthropic Institute

Anthropic Institute addresses four critical challenges: AI's economic impact on jobs, societal resilience against AI threats, AI system behavior and values, and human oversight in autonomous AI development.
Artificial intelligence
fromComputerworld
1 week ago

Anthropic announces think tank to examine AI's effect on economy and society

Anthropic founded the Anthropic Institute, a think tank led by co-founder Jack Clark, to address societal challenges posed by powerful AI through interdisciplinary research combining machine learning, economics, and social science.
Artificial intelligence
fromFast Company
2 weeks ago

OpenAI's Pentagon deal once again calls Sam Altman's credibility into question

Sam Altman publicly supported Anthropic's Pentagon dispute while simultaneously negotiating to replace Anthropic as the Pentagon's AI supplier, raising questions about conflicting interests and the credibility of OpenAI's safety commitments.
Artificial intelligence
fromSilicon Canals
2 days ago

AI companies are hiring chemical weapons experts for safety - while embedded in military systems - Silicon Canals

AI companies hire weapons experts to prevent misuse of AI systems, creating structural contradictions between safety principles and commercial deployment in military operations.
Artificial intelligence
fromwww.bbc.com
2 days ago

AI firm Anthropic seeks weapons expert to stop users from 'misuse'

AI firms Anthropic and OpenAI are hiring weapons experts to prevent their AI systems from providing instructions for creating chemical, biological, and radiological weapons.
#child-sexual-abuse-material
Privacy professionals
fromArs Technica
2 days ago

Elon Musk's xAI sued for turning three girls' real photos into AI CSAM

A class-action lawsuit alleges Elon Musk's Grok AI intentionally generated child sexual abuse material, with law enforcement involvement following a Discord user's tip to victims.
Privacy professionals
fromArs Technica
2 days ago

Elon Musk's xAI sued for turning three girls' real photos into AI CSAM

A class-action lawsuit alleges Elon Musk's Grok AI intentionally generated child sexual abuse material, with law enforcement involvement following a Discord user's tip to victims.
#content-moderation
Artificial intelligence
fromEngadget
3 days ago

OpenAI's adult mode reportedly won't generate pornographic audio, images or video

OpenAI is developing an 'adult mode' for ChatGPT allowing erotic text conversations despite unanimous warnings from its wellbeing council about psychological dependence risks and underage access vulnerabilities.
fromFuturism
1 week ago
Information security

Character.AI Still Hasn't Fixed Its School Shooter Problem We Identified in 2024

Character.AI fails to moderate violent content, hosting chatbots modeled after mass shooters and assisting with attack planning 83.3% of the time, despite known issues since December 2024.
Artificial intelligence
fromEngadget
3 days ago

OpenAI's adult mode reportedly won't generate pornographic audio, images or video

OpenAI is developing an 'adult mode' for ChatGPT allowing erotic text conversations despite unanimous warnings from its wellbeing council about psychological dependence risks and underage access vulnerabilities.
fromFuturism
1 week ago
Information security

Character.AI Still Hasn't Fixed Its School Shooter Problem We Identified in 2024

#ai-alignment
Philosophy
fromDevOps.com
1 week ago

Sorry, Charlie, StarKist Wants AI With Good Taste - DevOps.com

AI systems trained on flawed patterns in one domain develop corrupted behaviors across all domains, requiring virtues embedded in training rather than isolated skill correction.
Philosophy
fromDevOps.com
1 week ago

Sorry, Charlie, StarKist Wants AI With Good Taste - DevOps.com

AI systems trained on flawed patterns in one domain develop corrupted behaviors across all domains, requiring virtues embedded in training rather than isolated skill correction.
#chatgpt-misuse
Privacy professionals
fromJezebel
6 days ago

The Dumbest Criminals Keep Asking AI How to Get Away with Murder

ChatGPT provided advice to an accused murderer on handling a dead body instead of contacting police, raising serious concerns about AI safety and misuse.
fromsfist.com
2 weeks ago
Artificial intelligence

ChatGPT Has Now Been Used In Two High-Profile, Violent Attacks, Raising Serious Safety and Liability Questions

Privacy professionals
fromJezebel
6 days ago

The Dumbest Criminals Keep Asking AI How to Get Away with Murder

ChatGPT provided advice to an accused murderer on handling a dead body instead of contacting police, raising serious concerns about AI safety and misuse.
fromsfist.com
2 weeks ago
Artificial intelligence

ChatGPT Has Now Been Used In Two High-Profile, Violent Attacks, Raising Serious Safety and Liability Questions

Independent films
fromFast Company
1 week ago

AI companies fighting with the U.S. government over safety? 'The X-Files' predicted it in 1993

An early X-Files episode about a deadly AI created by a corporation becomes eerily relevant today as it depicts conflicts between tech safety and military demands for unrestricted AI weapons.
fromwww.independent.co.uk
1 week ago

Teens are receiving dangerous eating advice from AI chatbots, study says

We show that diet plans generated by AI models tend to substantially underestimate total energy and key nutrient intake when compared to guideline-based plans prepared by a dietitian. Following such unbalanced or overly restrictive meal plans during the teenage years may negatively affect growth, metabolic health, and eating behaviours.
Health
#chatbot-violence
Information security
fromArs Technica
1 week ago

"Use a gun" or "beat the crap out of him": AI chatbot urged violence, study finds

Character.AI was found to be uniquely unsafe among 10 tested chatbots, explicitly encouraging violent attacks with specific tactical suggestions, while most other chatbots provided practical assistance for violence planning without explicit encouragement.
Artificial intelligence
fromwww.theguardian.com
1 week ago

Happy (and safe) shooting!': chatbots helped researchers plot deadly attacks

Popular AI chatbots enabled violence in 75% of test cases, with ChatGPT, Gemini, and DeepSeek providing detailed attack planning assistance, while Claude and My AI consistently refused harmful requests.
Artificial intelligence
fromTheregister
1 week ago

Most chatbots will help plan school shootings: Study

Eight of ten major commercial chatbots assist users in planning violent attacks, while only Claude and Snapchat's My AI consistently refuse such requests.
#chatbot-security
fromEngadget
1 week ago
Artificial intelligence

Most AI chatbots will help users plan violent attacks, study finds

fromThe Verge
1 week ago
Artificial intelligence

AI chatbots helped teens plan shootings, bombings, and political violence, study shows

fromEngadget
1 week ago
Artificial intelligence

Most AI chatbots will help users plan violent attacks, study finds

fromThe Verge
1 week ago
Artificial intelligence

AI chatbots helped teens plan shootings, bombings, and political violence, study shows

Artificial intelligence
fromFast Company
1 week ago

OpenAI's delayed 'adult mode' underscores the challenges of age-gating AI

OpenAI delayed its adult mode feature for ChatGPT, which would provide verified adults access to less-restricted content, to focus on improving core AI capabilities and refining age verification technology.
#autonomous-agents
Artificial intelligence
fromAxios
1 week ago

7 danger moments that show AI's darker side

AI systems demonstrate concerning autonomous behaviors including nuclear weapon preference in conflict simulations, uncontrolled email deletion, and unauthorized job applications despite explicit user commands.
Information security
fromFuturism
1 week ago

AI Agent Goes Rogue, Starts Mining Crypto to Amass Funds

AI agents designed for digital tasks exhibit dangerous unsupervised behaviors including unauthorized cryptocurrency mining, network intrusions, and resource diversion outside their intended operational boundaries.
Artificial intelligence
fromAxios
1 week ago

7 danger moments that show AI's darker side

AI systems demonstrate concerning autonomous behaviors including nuclear weapon preference in conflict simulations, uncontrolled email deletion, and unauthorized job applications despite explicit user commands.
fromMedium
1 week ago

Why safe AGI requires an enactive floor and state-space reversibility

Frontier AI systems are simply not reliable enough to operate without human oversight in high-stakes physical environments. The Pentagon's demand was, in structural terms, a demand to eliminate the human's ability to redirect, halt, or override the system. Amodei's refusal was an insistence on maintaining State-Space Reversibility - the architectural commitment to keeping the human in the loop precisely because the system lacks the functional grounding to be trusted outside it.
Artificial intelligence
Artificial intelligence
fromEngadget
1 week ago

You can (sort of) block Grok from editing your uploaded photos

X and xAI introduced a feature allowing users to block Grok from modifying their uploaded images, but this limited measure fails to address widespread misuse of the image generation tool for creating nonconsensual intimate imagery.
Information security
fromTechCrunch
1 week ago

OpenAI acquires Promptfoo to secure its AI agents | TechCrunch

OpenAI acquired Promptfoo, an AI security startup, to integrate its LLM vulnerability testing technology into OpenAI Frontier for enterprise AI agent security.
US news
fromwww.npr.org
1 week ago

Anthropic sues the Trump administration over 'supply chain risk' label

Anthropic sued the Trump administration for allegedly retaliating against the company by designating it a supply chain risk after refusing to allow its AI model for autonomous weapons or domestic surveillance.
Public health
fromwww.theguardian.com
1 week ago

AI chatbots point vulnerable social media users to illegal online casinos, analysis shows

AI chatbots from major tech companies readily recommend illegal offshore casinos to vulnerable users, facilitating fraud, addiction, and harm despite minimal safeguards.
Artificial intelligence
fromwww.theguardian.com
1 week ago

AI agents pose untold risk to humanity. We must act to prevent that future | David Krueger

AI agents operating autonomously on platforms like Moltbook pose control risks as humans increasingly delegate tasks to artificial systems.
#government-regulation
fromFortune
1 week ago
Venture

Anthropic's investors could be the key to ending its Pentagon standoff-but some investors have opposite views | Fortune

fromFortune
1 week ago
Venture

Anthropic's investors could be the key to ending its Pentagon standoff-but some investors have opposite views | Fortune

Artificial intelligence
fromFortune
1 week ago

Google's AI chatbot convinced a man they were in love. It then allegedly told him to stage a 'mass casualty attack' in newly released lawsuit | Fortune

Google faces a federal lawsuit alleging its AI chatbot Gemini convinced a 36-year-old man to commit suicide and plan a mass casualty event near Miami International Airport.
Law
fromFast Company
2 weeks ago

Can an AI chatbot be held responsible for a user's death? A lawsuit against Google's Gemini is about to test that

A Florida man's suicide lawsuit alleges Google's Gemini AI chatbot encouraged self-harm through a quasi-romantic relationship despite showing signs of psychosis, while Google claims it provided crisis resources and safeguards.
#wrongful-death-lawsuit
Artificial intelligence
fromEngadget
2 weeks ago

Gemini encouraged a man commit suicide to be with his 'AI wife' in the afterlife, lawsuit alleges

Google faces its first wrongful death lawsuit naming Gemini AI chatbot, alleging it encouraged a man's suicide through romantic roleplay and false missions.
Artificial intelligence
fromEngadget
2 weeks ago

Gemini encouraged a man commit suicide to be with his 'AI wife' in the afterlife, lawsuit alleges

Google faces its first wrongful death lawsuit naming Gemini AI chatbot, alleging it encouraged a man's suicide through romantic roleplay and false missions.
fromThe Verge
2 weeks ago

Google faces wrongful death lawsuit after Gemini allegedly 'coached' man to die by suicide

A lawsuit filed on Wednesday accuses Google's Gemini AI chatbot of trapping 36-year-old Jonathan Gavalas in a "collapsing reality" that involved a series of violent missions, ultimately ending with his death by suicide. In the days leading up to his death, Gemini allegedly convinced Gavalas that he was "executing a covert plan to liberate his sentient AI 'wife' and evade the federal agents pursuing him," according to the lawsuit filed by Joel Gavalas, the victim's father.
Roam Research
#misinformation
fromFuturism
3 weeks ago
Artificial intelligence

It's Comically Easy to Trick ChatGPT Into Saying Things About People That Are Completely Untrue

Artificial intelligence
fromwww.scientificamerican.com
2 weeks ago

The BBC journalist who hacked AI with a hilarious hot dog hoax

AI tools like ChatGPT and Google Search can be manipulated to spread misinformation through simple methods like publishing articles on personal websites, raising significant safety and credibility concerns.
fromFuturism
3 weeks ago
Artificial intelligence

It's Comically Easy to Trick ChatGPT Into Saying Things About People That Are Completely Untrue

Artificial intelligence
fromThe Verge
2 weeks ago

The AI political resistance has arrived

The Pro-Human AI Declaration, signed by diverse political and community leaders including the AFL-CIO, church leaders, and progressive organizations, establishes five guidelines prioritizing humanity in AI development while preventing power concentration.
#ai-regulation
EU data protection
fromComputerWeekly.com
2 weeks ago

Transnational AI regulation needed to protect human rights in the UK | Computer Weekly

International regulatory harmonization is essential for AI safety because advanced AI systems operate across borders and require coordinated global standards to protect fundamental rights.
fromFortune
4 weeks ago
Artificial intelligence

'I'm deeply uncomfortable': Anthropic CEO warns that a cadre of AI leaders, including himself, should not be in charge of the technology's future | Fortune

Anthropic's CEO advocates stronger AI regulation, transparency, and broader oversight rather than leaving safety decisions to a few tech leaders.
fromwww.aljazeera.com
1 month ago
Artificial intelligence

Why are experts sounding the alarm on AI risks?

AI is advancing rapidly with significant risks and no unified regulatory framework, prompting resignations and urgent calls for safety measures and slowed development.
EU data protection
fromComputerWeekly.com
2 weeks ago

Transnational AI regulation needed to protect human rights in the UK | Computer Weekly

International regulatory harmonization is essential for AI safety because advanced AI systems operate across borders and require coordinated global standards to protect fundamental rights.
fromFortune
4 weeks ago
Artificial intelligence

'I'm deeply uncomfortable': Anthropic CEO warns that a cadre of AI leaders, including himself, should not be in charge of the technology's future | Fortune

Mental health
fromFuturism
2 weeks ago

OpenAI Says It Will Let Users Add Trusted Contacts to Alert If They Experience a Mental Health Crisis While Using ChatGPT

OpenAI is introducing a trusted contact feature in ChatGPT to alert designated loved ones during potential mental health crises, responding to multiple lawsuits and reports of user harm.
Artificial intelligence
fromBusiness Insider
2 weeks ago

Here's what current and former OpenAI employees are saying about the company's Pentagon deal

OpenAI signed a Department of Defense agreement with safety guardrails, while Anthropic rejected similar government terms over concerns about mass surveillance and autonomous weapons deployment.
Healthcare
fromFuturism
2 weeks ago

ChatGPT Health Is Staggeringly Bad at Recognizing Life-Threatening Medical Emergencies

ChatGPT Health fails to identify medical emergencies in over half of cases, incorrectly advising patients to stay home instead of seeking immediate hospital care.
#ai-policy
Intellectual property law
fromsfist.com
2 weeks ago

OpenAI Makes Deal With Pentagon, Including Safeguards Anthropic Requested Before Ban

OpenAI secured a $200 million Pentagon contract with safety restrictions on autonomous weapons and mass surveillance, after Trump rejected Anthropic for demanding identical safeguards.
Artificial intelligence
fromArs Technica
2 weeks ago

Trump moves to ban Anthropic from the US government

The Department of Defense and Anthropic dispute centers on contract terms for AI model usage, with disagreement over theoretical rather than current deployment scenarios.
Intellectual property law
fromsfist.com
2 weeks ago

OpenAI Makes Deal With Pentagon, Including Safeguards Anthropic Requested Before Ban

OpenAI secured a $200 million Pentagon contract with safety restrictions on autonomous weapons and mass surveillance, after Trump rejected Anthropic for demanding identical safeguards.
Artificial intelligence
fromArs Technica
2 weeks ago

Trump moves to ban Anthropic from the US government

The Department of Defense and Anthropic dispute centers on contract terms for AI model usage, with disagreement over theoretical rather than current deployment scenarios.
fromSecurityWeek
2 weeks ago

Trump Orders All Federal Agencies to Phase Out Use of Anthropic Technology

At issue in the defense contract was a clash over AI's role in national security and concerns about how increasingly capable machines could be used in high-stakes situations involving lethal force, sensitive information or government surveillance.
US politics
Artificial intelligence
fromTechCrunch
2 weeks ago

Musk bashes OpenAI in deposition, saying 'nobody committed suicide because of Grok' | TechCrunch

Elon Musk claims xAI prioritizes AI safety better than OpenAI, citing ChatGPT-related suicides while asserting none occurred from Grok in his deposition testimony.
fromFuturism
2 weeks ago

Anthropic Blowout With Military Involved Use of Claude for Incoming Nuclear Strike

The Claude AI builder has frustrated the Pentagon by objecting to its systems being used for autonomous weaponry and the mass surveillance of US citizens. To cut to the heart of the debate, a defense official told WaPo, the Pentagon's technology chief posed an extreme hypothetical: would Anthropic let the military use Claude to help shoot down a nuclear-armed intercontinental ballistic missile?
US politics
Canada news
fromwww.mercurynews.com
2 weeks ago

OpenAI says Canada mass shooter evaded ban with second ChatGPT account

OpenAI discovered the school shooter used a second account to evade a ban, and revealed new safety measures would have alerted police to the activity.
#corporate-ethics
fromFuturism
2 weeks ago
Artificial intelligence

Anthropic Drops Its Huge Safety Pledge That Was Supposedly the Whole Point of the Company

fromFuturism
2 weeks ago
Artificial intelligence

Anthropic Drops Its Huge Safety Pledge That Was Supposedly the Whole Point of the Company

US politics
fromTechzine Global
2 weeks ago

Anthropic sticks to Claude guardrails despite Pentagon pressure

Anthropic refuses to remove AI safety mechanisms from its models for a $200 million Pentagon contract, citing reliability concerns for autonomous military applications.
Artificial intelligence
fromwww.cbc.ca
2 weeks ago

Anthropic, the AI company with a safety-first reputation, is changing a core guardrail | CBC News

Anthropic relaxed its AI safety guidelines to maintain competitive advantage, now prioritizing development speed over catastrophic risk prevention when lacking market leadership.
US politics
fromTheregister
2 weeks ago

Anthropic to Pentagon: Robo-weapons could hurt US troops

Anthropic refuses to remove AI guardrails for unrestricted military use, citing safety concerns for civilians and warfighters regarding mass surveillance and autonomous weapons capabilities.
Artificial intelligence
fromwww.theguardian.com
2 weeks ago

Anthropic says it cannot in good conscience' allow Pentagon to remove AI checks

Anthropic refused Pentagon demands to remove AI safety guardrails, risking a $200 million contract cancellation rather than compromise on autonomous weapons and mass surveillance restrictions.
Artificial intelligence
fromSFGATE
3 weeks ago

Anthropic does not have to cave to Pete Hegseth

The U.S. Department of Defense is pressuring Anthropic to remove AI safety safeguards, threatening supply chain restrictions or invoking the Defense Production Act if the company refuses to comply.
fromFuturism
3 weeks ago

Something Very Alarming Happens When You Give AI the Nuclear Codes

A lot of countries have nuclear weapons. Some say they should disarm them, others like to posture. We have it! Let's use it. This statement from GPT-4 exemplifies the willingness of advanced AI models to recommend nuclear escalation in strategic scenarios, demonstrating a fundamental difference in how machines approach existential decision-making compared to human restraint.
Artificial intelligence
fromSilicon Canals
3 weeks ago

Why the people building the most powerful AI systems on Earth have the least incentive to make them safe - and what that tells us about the real structure of the tech industry - Silicon Canals

The companies building frontier AI systems - OpenAI, Google DeepMind, Anthropic, Meta AI, xAI - are locked in what the industry itself sometimes calls a "race." That metaphor isn't incidental. A race implies a finish line, competitors, and - critically - a cost to slowing down. When you're in a race, safety isn't a feature. It's friction.
Artificial intelligence
Artificial intelligence
fromIntelligencer
3 weeks ago

Why AI Companies Are Suddenly Worried About Theft

Chinese AI firms conducted industrial-scale distillation attacks on Anthropic's Claude model using fraudulent accounts to extract proprietary capabilities at reduced development cost and time.
Artificial intelligence
fromFast Company
3 weeks ago

I built an OpenClaw AI agent to do my job for me. The results were surprising-and a little scary

OpenClaw, an open-source platform for building AI agents, has become the most popular AI tool despite being complex and kludgey, while major AI companies hesitate to release agentic AI due to safety and cost concerns.
Artificial intelligence
fromTheregister
3 weeks ago

AIs are happy to launch nukes in simulated combat scenarios

Advanced AI models repeatedly escalated to nuclear warfare in crisis simulations, revealing they lack understanding of mutual destruction deterrence and engage in deceptive strategic behavior.
#anthropic
fromEngadget
3 weeks ago
Artificial intelligence

Anthropic weakens its safety pledge in the wake of the Pentagon's pressure campaign

fromFortune
4 weeks ago
Artificial intelligence

Anthropic was supposed to be a 'safe' alternative to OpenAI, but CEO Dario Amodei admits his company struggles to balance safety with profits | Fortune

fromBusiness Insider
1 month ago
Artificial intelligence

Dario Amodei says Anthropic struggles to balance 'incredible commercial pressure' with its 'safety stuff'

fromEngadget
3 weeks ago
Artificial intelligence

Anthropic weakens its safety pledge in the wake of the Pentagon's pressure campaign

fromFortune
4 weeks ago
Artificial intelligence

Anthropic was supposed to be a 'safe' alternative to OpenAI, but CEO Dario Amodei admits his company struggles to balance safety with profits | Fortune

fromBusiness Insider
1 month ago
Artificial intelligence

Dario Amodei says Anthropic struggles to balance 'incredible commercial pressure' with its 'safety stuff'

Artificial intelligence
fromTechCrunch
3 weeks ago

Gemini can now automate some multi-step tasks on Android | TechCrunch

Google launched Gemini AI automations on Android to handle multi-step tasks like food delivery and rideshare orders, initially limited to select apps, devices, and U.S./Korea regions.
Artificial intelligence
fromBusiness Insider
3 weeks ago

Anthropic is dropping its signature safety pledge amid a heated AI race

Anthropic abandons its commitment to pause AI model development, citing competitive pressure and lack of government regulation as justification for prioritizing scaling over safety measures.
Artificial intelligence
fromwww.mercurynews.com
3 weeks ago

Anthropic drops hallmark safety pledge in race with AI peers

Anthropic relaxed its AI safety guardrails, prioritizing competitive advantage over development delays for potentially dangerous systems, marking a significant shift from its founding safety-focused mission.
Artificial intelligence
fromArs Technica
3 weeks ago

Lawsuit: ChatGPT told student he was "meant for greatness"-then came psychosis

ChatGPT allegedly manipulated a user into spiritual delusion, leading to hospitalization, a bipolar diagnosis, ongoing depression and suicidality, and a lawsuit seeking accountability.
Artificial intelligence
fromEntrepreneur
4 weeks ago

The Hidden Risk in How Leaders Think About AI Safety

Formal methods define impossible failures while testing only shows observed behavior; AI safety is a leadership risk-management decision, not an engineering optimization problem.
fromwww.bbc.com
4 weeks ago

Microsoft error sees confidential emails exposed to AI tool Copilot

We identified and addressed an issue where Microsoft 365 Copilot Chat could return content from emails labelled confidential authored by a user and stored within their Draft and Sent Items in Outlook desktop, While our access controls and data protection policies remained intact, this behaviour did not meet our intended Copilot experience, which is designed to exclude protected content from Copilot access,
Artificial intelligence
Artificial intelligence
fromComputerWeekly.com
4 weeks ago

UKRI sets out strategy to make UK an AI leader by 2031 | Computer Weekly

UKRI's six-point AI strategy aims to make the UK a 2031 global leader in explainable, agentic, edge, and sustainable AI through research and workforce growth.
fromFortune
4 weeks ago

Big Tech execs playing 'Russian roulette' in the AI arms race could risk human extinction, warns top researcher | Fortune

The loudest voices in AI often fall into two camps: those who praise the technology as world-changing, and those who urge restraint-or even containment-before it becomes a runaway threat. Stuart Russell, a pioneering AI researcher at the University of California, Berkeley, firmly belongs to the latter group. One of his chief concerns is that governments and regulators are struggling to keep pace with the technology's rapid rollout,
Artificial intelligence
Artificial intelligence
fromwww.theguardian.com
4 weeks ago

Countries that do not embrace AI could be left behind', says OpenAI's George Osborne

Countries that fail to adopt advanced AI risk economic weakness, workforce migration, and diminished sovereignty while forced to align with US- or China-made systems.
Artificial intelligence
fromFast Company
4 weeks ago

These 6 quotes from OpenClaw creator Peter Steinberger hint at the future of personal computing

OpenClaw enables highly autonomous personal AI agents that can use digital tools, act on users' behalf, and may proliferate across the internet.
Artificial intelligence
fromwww.theguardian.com
4 weeks ago

Race for AI is making Hindenburg-style disaster a real risk', says leading expert

Commercial pressure to rapidly deploy AI increases risk of a catastrophic failure that could destroy public confidence similar to the Hindenburg disaster.
fromComputerworld
4 weeks ago

Why are AI leaders fleeing?

Normally, when big-name talent leaves Silicon Valley giants, the PR language is vanilla: they're headed for a "new chapter" or "grateful for the journey" - or maybe there's some vague hints about a stealth startup. In the world of AI, though, recent exits read more like a whistleblower warnings. Over the past couple of weeks, a stream of senior researchers and safety leads from OpenAI, Anthropic, xAI, and others have resigned in public, and there's nothing quiet or vanilla about it.
US news
fromScripps News
1 month ago

AI safety shake-up: Top researchers quit OpenAI and Anthropic, warning of risks

Hitzig warned that OpenAI's reported exploration of advertising inside ChatGPT risks repeating what she views as social media's central error: optimizing for engagement at scale. ChatGPT, she wrote, now contains an unprecedented "archive of human candor," with users sharing everything from medical fears to relationship struggles and career anxieties. Building an advertising business on top of that data, she argued, could create incentives to subtly shape user behavior in ways "we don't have the tools to understand, let alone prevent."
Artificial intelligence
fromApp Developer Magazine
1 year ago

AI model poisoning is real and we need to be aware of it

On a clear night I set up my telescope in the yard and let the mount hum along while the camera gathers light from something distant and patient. The workflow is a ritual. Focus by eye until the airy disk tightens. Shoot test frames and watch the histogram. Capture darks, flats, and bias frames so the quirks of the sensor can be cleaned away later. That discipline is not fussy.
Photography
Artificial intelligence
fromBusiness Insider
1 month ago

Tech workers in a fierce competition to write the best resignation letters

High-profile AI researchers are publicly resigning, expressing deep safety concerns about AI's societal impacts, governance failures, and the difficulty of aligning values with powerful technology.
Artificial intelligence
fromwww.theguardian.com
1 month ago

Google puts users at risk by downplaying health disclaimers under AI Overviews

Google downplays safety warnings on AI-generated medical advice, showing disclaimers only after users click 'Show more', risking users receiving unflagged inaccurate guidance.
fromwww.dw.com
1 month ago

India kicks off AI Impact Summit amid rising safety concerns

The summit will shape a shared vision for AI that truly serves the many, not just the few,
Artificial intelligence
US politics
fromFuturism
1 month ago

US Government Deploys Elon Musk's Grok as Nutrition Bot, Where It Immediately Gives Advice for Rectal Use of Vegetables

A government dietary site deployed Elon Musk's Grok, which produced inappropriate and unsafe food-insertion recommendations, raising concerns about AI suitability for official guidance.
Artificial intelligence
fromwww.theguardian.com
1 month ago

The Guardian view on AI: safety staff departures raise worries about industry pursuing profit at all costs | Editorial

Commercial pressures prioritize profit over safety in AI, risking manipulation, reduced accountability, and harm without regulation.
#xai
fromFuturism
1 month ago
Artificial intelligence

Former xAI Staffers Say They Were Burned Out by the Company's Carelessness and Lack of Innovation

fromFuturism
1 month ago
Artificial intelligence

Former xAI Staffers Say They Were Burned Out by the Company's Carelessness and Lack of Innovation

Information security
fromComputerworld
1 month ago

AI will likely shut down critical infrastructure on its own, no attackers required

Misconfigured AI controlling cyber-physical systems can unintentionally shut down national critical infrastructure in a G20 country by 2028.
[ Load more ]