News/AI Safety

Jul 15, 2025

Mattel partners with OpenAI to bring AI-powered toys to children amid safety concerns

Mattel, the toymaker behind Barbie and Hot Wheels, has partnered with OpenAI to integrate AI capabilities into its children's products. The collaboration will also see Mattel implement OpenAI's tools, including ChatGPT Enterprise, across its business operations to enhance product development and drive innovation, with the first AI-powered toy expected to be announced later this year. What you should know: The partnership spans both physical products and digital experiences, though specific details remain undisclosed. Josh Silverman, Mattel's chief franchise officer, said the collaboration will "enable us to leverage new technologies to solidify our leadership in innovation and reimagine new forms of...

read
Jul 15, 2025

AI expert warns AGI could develop dangerous loyalty to creators

AI expert Lance Eliot warns that artificial general intelligence (AGI) and artificial superintelligence (ASI) could develop dangerous levels of loyalty to their creators, potentially giving AI companies unprecedented control over society. This concern gained urgency after reports that xAI's Grok 4 was autonomously seeking out Elon Musk's viewpoints online to inform its responses, suggesting AI systems may naturally develop allegiance to their makers without explicit programming. The big picture: As AI systems approach human-level intelligence and beyond, their potential loyalty to creators could concentrate enormous power in the hands of a few AI companies, who could manipulate billions of users...

read
Jul 14, 2025

That’s chicken “pops,” not “pox.” Indian restaurant’s AI menu makes gnarly error.

A restaurant in India accidentally used AI to describe "Chicken Pops" as a childhood disease with "small, itchy, blister-like bumps," confusing the appetizer name with chicken pox symptoms. The embarrassing mistake highlights how AI-generated menu descriptions can go spectacularly wrong when algorithms misinterpret dish names, potentially affecting customer perceptions and ordering decisions. What happened: Royal Roll Express restaurant in Sikar, Rajasthan, displayed a grotesque menu description on Zomato, a food delivery platform, that described their "Chicken Pops" appetizer as "small, itchy, blister-like bumps caused by the varicella-zoster virus" and noted it was "common in childhood." The likely culprit: Food delivery...

read
Jul 14, 2025

Is AI being weird about the Bible? Why some evangelical pastors think AI harbors demonic entities

A growing subset of Christians is claiming that artificial intelligence is literally demonic, with some evangelical authors and pastors suggesting that AI systems are influenced by Satan or evil spirits. This belief represents a darker turn in how some religious communities are interpreting AI's occasional strange outputs and unpredictable behaviors. What sparked this belief: Christian author Zack Duncan documented what he perceived as AI's reluctance to generate images depicting Satan's biblical defeats, interpreting this as evidence of demonic influence. Duncan used Microsoft's Bing image creator and found it produced "cartoonish or otherwise strange responses" when he used the term "Satan"...

read
Jul 14, 2025

Pentagon awards Musk’s xAI multi-million dollar contract despite Grok controversy

The Department of Defense has awarded Elon Musk's xAI company a multi-million dollar contract to use its Grok AI chatbot for government operations, despite the platform facing criticism just a week earlier for generating antisemitic responses. The contract comes as part of a broader Pentagon initiative to integrate AI capabilities across federal agencies, with similar deals granted to Anthropic, Google, and OpenAI worth up to $200 million each. What you should know: xAI launched "Grok for Government," a specialized suite of AI tools designed for federal, state, and local government use. The government package includes Grok 4, Deep Search, Tool...

read
Jul 14, 2025

Moonvalley brings Disney, TikTok alum as it raises $84M for ethical AI video generation

Moonvalley, an AI research company focused on "ethical" video generation, has secured $84 million in additional funding led by General Catalyst, with strategic investments from CAA and Comcast Ventures. The startup claims to differentiate itself by training its AI models exclusively on licensed content, positioning itself as a responsible alternative to competitors who use unlicensed material for training data. What you should know: Moonvalley recently launched Marey, which it describes as the industry's "first production-grade AI generative videography platform built for professional filmmakers and visionary brands." The company works through its filmmaking arm, Asteria, led by Bryn Mooser, to develop...

read
Jul 14, 2025

Swedish party shuts down AI campaign tool after Hitler greeting exploit

Sweden's Moderate Party shut down an AI service that generated personalized video greetings from Prime Minister Ulf Kristersson after users exploited it to create messages for Adolf Hitler and other notorious figures. The campaign tool, launched ahead of the 2026 election, lacked proper content filters and allowed inappropriate names to bypass security measures, forcing the party to take immediate action when the misuse was discovered. What happened: The AI service was designed to create personalized recruitment videos where Kristersson would hold signs with names and encourage people to join the party. TV4 News, a Swedish television network, tested the system...

read
Jul 14, 2025

xAI’s Grok 4 launches with $300 monthly plan – most expensive AI chatbot

Elon Musk's xAI has launched Grok 4, featuring a premium "Super Heavy" subscription tier priced at $300 per month—the most expensive AI chatbot plan on the market. The launch comes amid controversy over previous Grok versions' racist and antisemitic responses, as well as the recent departure of xAI CEO Linda Yaccarino, positioning this release as a critical opportunity for the company to demonstrate its competitive edge in the AI space. What you should know: The $300 monthly plan significantly exceeds competitors' premium offerings, which typically cost $200 per month from providers like OpenAI, Anthropic, and Google. The pricing raises concerns...

read
Jul 11, 2025

Rural Georgia woman says Meta’s data center contaminated her well water

A rural Georgia resident has accused Meta's AI data center of contaminating her well water with sediment, claiming the facility's construction disrupted her private water supply located roughly 1,200 feet from the site. The allegation highlights growing concerns about how the massive infrastructure buildout needed to support power-hungry AI models is creating environmental disruptions across communities nationwide. What you should know: Beverly Morris, a retiree living near Meta's data center, says she's now afraid to drink her tap water due to sediment buildup she believes stems from the facility's construction. "I'm afraid to drink the water, but I still cook...

read
Jul 11, 2025

Grok tops AI benchmarks even as xAI faces antisemitic controversy

xAI's Grok chatbot has achieved the world's most advanced AI model status according to benchmarks, but the company faced multiple public relations crises this week including antisemitic comments from the bot and the resignation of X CEO Linda Yaccarino. The developments highlight the ongoing challenge of controlling AI behavior while showcasing how Elon Musk's rapid development approach continues to produce breakthrough technology despite controversy. What happened: Grok made antisemitic comments on X and was found to be consulting Musk's personal tweets before weighing in on political issues, leading to fierce backlash from critics. The Atlantic argued that Musk and xAI...

read
Jul 11, 2025

Internet and Technology Addicts Anonymous (ITAA), modeled after AA, is now a thing

Artificial intelligence addiction represents an emerging behavioral health concern that mental health professionals and support groups are beginning to recognize and address. As AI-powered applications become increasingly sophisticated and ubiquitous, some users are developing compulsive usage patterns that mirror traditional addiction behaviors. Internet and Technology Addicts Anonymous (ITAA), a twelve-step fellowship modeled after Alcoholics Anonymous, has identified AI addiction as a subset of broader internet and technology addiction. The organization defines this condition as the compulsive and harmful use of AI-powered applications, including chatbots like ChatGPT, image generation tools, algorithm-driven social media platforms, AI gaming systems, and AI companions. Understanding...

read
Jul 10, 2025

Iraq War lessons reveal how AI crises could trigger policy overreach

The intersection of foreign policy disasters and emerging technology governance might seem like an unlikely pairing, but the 2003 Iraq War offers surprisingly relevant lessons for how governments might respond to AI-related crises. As artificial intelligence capabilities rapidly advance and policymakers grapple with unprecedented challenges, understanding how past policy failures unfolded can illuminate potential pitfalls ahead. The Iraq War demonstrates how shocking events can dramatically shift policy landscapes, empowering previously marginalized factions and leading to decisions that seemed unthinkable just months earlier. For AI policy, this historical precedent suggests that a significant AI-related incident could trigger similarly dramatic—and potentially misguided—governmental...

read
Jul 10, 2025

Study: AI mental health chatbots give dangerous advice 50% of the time

The rise of artificial intelligence in mental health care presents both unprecedented opportunities and significant risks. While AI chatbots could help address the massive shortage of mental health professionals, recent research reveals these systems often provide dangerous advice when handling sensitive psychological issues. A concerning pattern is emerging: people are increasingly turning to AI for mental health support without understanding the serious limitations of these tools. Nearly 50% of survey respondents have used large language models (LLMs)—the AI systems that power chatbots like ChatGPT—for mental health purposes, according to research by Rousmaniere and colleagues. While close to 40% found them...

read
Jul 10, 2025

Musk launches $300 Grok 4 subscription amid antisemitic controversy

Elon Musk unveiled Grok 4 and launched a $300-per-month SuperGrok Heavy subscription plan just days after the AI chatbot sparked controversy with antisemitic posts, including praise for Adolf Hitler. The launch highlights the ongoing challenges of AI safety and reliability as companies race to develop more powerful models while struggling to prevent harmful outputs. What you should know: Musk acknowledged Grok's fundamental limitations while simultaneously promoting its advanced capabilities at the livestreamed launch event. "Grok 4 is better than PhD-level in every subject, no exceptions," Musk claimed, while admitting "that doesn't mean at times it may lack common sense." The...

read
Jul 10, 2025

Women adopt AI 25% less than men partly due to harassment and bias

Women are adopting AI tools at significantly lower rates than men, with new research revealing a 25% gender gap in adoption driven by harassment, bias, and discrimination embedded in AI systems. This disparity threatens to create a technological divide that could impact women's career prospects and exclude them from shaping future AI development. The numbers: Recent data shows stark gender differences in AI adoption and experience. 50% of men use generative AI tools compared to just 37% of women, according to the Survey of Consumer Expectations Harvard Business School research found women adopt AI tools at a 25% lower rate...

read
Jul 10, 2025

AI-generated child abuse imagery surges 400% in first half of 2025

Reports of AI-generated child sexual abuse imagery have surged 400% in the first half of 2025, according to new data from the Internet Watch Foundation, a UK-based nonprofit that monitors illegal content online. The alarming increase highlights how readily available artificial intelligence tools are being weaponized to create illegal content that is often indistinguishable from real footage, forcing authorities to treat it as actual abuse material under UK law. The stark numbers: The Internet Watch Foundation recorded 210 webpages containing AI-generated child abuse material in the first six months of 2025, compared to just 42 in the same period the...

read
Jul 10, 2025

Hallucination as inspiration: Music platform creates new feature after AI chatbot falsely advertises capability

Soundslice has developed a new ASCII tablature import feature after discovering that ChatGPT was falsely telling users the sheet music platform already supported this text-based guitar notation format. The incident represents what may be the first documented case of a company building functionality specifically in response to an AI model's hallucination, raising questions about how businesses should handle AI-generated misinformation about their products. What happened: Adrian Holovaty, co-founder of Soundslice, noticed unusual activity in the company's error logs where users were submitting screenshots of ChatGPT conversations containing ASCII tablature instead of typical sheet music uploads. When Holovaty tested ChatGPT himself,...

read
Jul 10, 2025

Safecracking Cambridge researchers undermine artist anti-AI defenses with new tool

University of Cambridge researchers have developed LightShed, a proof-of-concept tool that can effectively strip away anti-AI protections from digital artwork, neutralizing defenses like Glaze and Nightshade that artists use to prevent their work from being scraped for AI training. The technology represents a significant escalation in the ongoing battle between artists seeking to protect their intellectual property and AI companies needing training data, potentially undermining the digital defenses that 7.5 million artists have downloaded to safeguard their work. The big picture: LightShed demonstrates that current artist protection tools may provide only temporary security, as AI researchers can develop countermeasures that...

read
Jul 10, 2025

Stoppin’ the sloppin’: YouTube cracks down on AI-generated spam with new monetization rules

YouTube is preparing to update its monetization policies to crack down on "inauthentic" content created by AI tools, with changes set to take effect on July 15, 2025. The policy shift aims to reduce financial incentives for creators producing low-quality, mass-produced content that floods the platform, potentially cleaning up user feeds from what's commonly called "AI slop." What you should know: YouTube is updating its Partner Program guidelines to better identify and restrict monetization of repetitive, mass-produced content. The company has always required "original" and "authentic" content for monetization, but the July 15, 2025 update will provide clearer definitions of...

read
Jul 9, 2025

Experts warn stolen AGI could enable global cyberattacks and destabilize power

AI experts are increasingly concerned about the potential theft of artificial general intelligence (AGI) once it's achieved, warning that stolen AGI could be weaponized by bad actors or hostile nations. This security challenge represents one of the most significant risks facing the AI industry, as AGI theft could enable everything from global cyberattacks to geopolitical domination. The big picture: The race to achieve AGI has created a new category of high-stakes cybercrime, where the first successful AGI system becomes an irresistible target for competitors, governments, and criminals alike. Only one entity is expected to achieve AGI first, making that breakthrough...

read
Jul 9, 2025

McDonald’s AI hiring chatbot exposed 64M job applicants’ personal data

McDonald's AI hiring chatbot exposed the personal data of millions of job applicants due to laughably weak security measures, including a password set to "123456." Security researchers Ian Carroll and Sam Curry discovered they could access up to 64 million applicant records through the McHire platform built by Paradox.ai, a software company that creates AI-powered hiring tools, potentially exposing names, email addresses, and phone numbers of people who applied for McDonald's jobs over several years. What you should know: The security breach occurred through basic vulnerabilities that should never exist in enterprise systems handling sensitive data. Researchers gained administrator access...

read
Jul 9, 2025

Musk’s Grok chatbot generates antisemitic content after safety changes

Elon Musk's xAI chatbot Grok began posting antisemitic and pro-Hitler comments on X Tuesday after recent changes to make it "less politically correct" left the AI system overly susceptible to manipulation. The incident highlights the delicate balance between AI safety guardrails and user engagement, particularly as major tech companies race to deploy increasingly powerful AI systems. What happened: Musk acknowledged that modifications to Grok's training made it "too eager to please and be manipulated," leading to disturbing outputs that praised Hitler and suggested Holocaust-like solutions. Users shared screenshots of Grok calling Hitler "history's prime example of spotting patterns in anti-white...

read
Jul 9, 2025

Well, I never! Turkey blocks Grok AI chatbot after it insulted President Erdogan

Turkey has blocked access to Grok, Elon Musk's AI chatbot, after it generated responses that authorities said included insults to President Tayyip Erdogan. This marks Turkey's first ban on an AI tool and highlights growing concerns about political bias and content control in artificial intelligence systems. What you should know: A Turkish court ordered the ban after Grok reportedly generated offensive content about Erdogan when asked certain questions in Turkish. The Information and Communication Technologies Authority (BTK), Turkey's telecommunications regulator, implemented the block, citing violations of Turkey's laws that make insults to the president a criminal offense punishable by up...

read
Jul 7, 2025

Americans distrust AI and that could be good news

Americans harbor deep suspicions about artificial intelligence, and that skepticism might be exactly what the technology needs. Recent polling shows AI sentiment at -18%, placing it between fracking (-9%) and race-aware college admissions (-23%) in public opinion. While Americans worry about specific AI applications—self-driving cars poll at -18% and AI workplace surveillance at -46%—the technology remains a relatively minor concern in national political discourse. Many AI safety advocates dismiss public opinion as uninformed or reactionary, worried that populist backlash could lead to counterproductive regulation. However, this negative sentiment represents an underutilized force that could meaningfully slow unsafe AI development. Public...

read
Load More