News/Governance

Jun 2, 2025

AI regulation takes a backseat as rapid advancement rides shotgun

The US House of Representatives has passed legislation that could significantly impact AI regulation across the country. The "One Big Beautiful Bill" includes a provision that would prevent individual states from regulating artificial intelligence for a decade, creating potential concerns for those focused on AI safety and oversight. This federal preemption raises questions about the appropriate balance between national consistency and local regulatory experimentation in managing emerging AI risks. Why this matters: The proposed 10-year moratorium on state-level AI regulation could create a regulatory vacuum at a time when AI governance frameworks are still developing. Key details: The provision is...

read
Jun 2, 2025

AI governance urgently needed to safeguard humanity’s future

The concept of a "Ulysses Pact" for AI suggests we need governance structures that allow us to pursue artificial intelligence's benefits while protecting ourselves from its existential risks. This framework offers a thoughtful middle path between unchecked AI development and complete restriction, advocating for binding agreements that future-proof humanity against potential AI dangers while still enabling technological progress. The big picture: Drawing on the Greek myth where Ulysses had himself tied to a ship's mast to safely hear the sirens' song, the author proposes we need similar self-binding mechanisms for AI development. AI represents our modern siren song—offering extraordinary breakthroughs...

read
May 28, 2025

How subtle biases derail LLM evaluations

Large Language Models are increasingly deployed as judges and decision-makers in critical domains, but their judgments suffer from systematic biases that threaten reliability. Research from The Collective Intelligence Project reveals that positional preferences, order effects, and prompt sensitivity significantly undermine LLMs' ability to make consistent judgments. Understanding these biases is crucial as AI systems expand into sensitive areas like hiring, healthcare, and legal assessments where decision-making integrity is paramount. The big picture: LLMs exhibit multiple systematic biases when used as judges, including positional preferences, ordering effects, and sensitivity to prompt wording, rendering their judgments unreliable. These biases appear across multiple...

read
May 27, 2025

AI-powered judges fail reliability tests, study finds

Large language models (LLMs) are increasingly making judgments in sensitive domains like hiring, healthcare, and law, but their decision-making mechanisms contain concerning biases and inconsistencies. New research from the Collective Intelligence Project reveals how LLM evaluations are undermined by position preferences, order effects, and prompt sensitivity—creating significant reliability issues that demand attention as these systems become more deeply integrated into consequential decision-making processes. The big picture: LLMs demonstrate multiple systematic biases when making judgments, raising serious questions about their reliability in high-stakes evaluation tasks. The research identifies specific patterns of positional bias, where models consistently prefer options presented first or...

read
May 26, 2025

AI governance drives urgent need for political ad reform

Effective political advocacy, not just good ideas, is essential for governing AI safety and preventing catastrophic risks. Jason Green-Lowe's analysis highlights a critical gap in the AI safety movement's approach: while technical research naturally circulates within scientific communities, AI governance proposals require deliberate political promotion to gain traction. This distinction matters profoundly as policymakers face competing priorities, information overload, and potential opposition from tech companies. The big picture: The AI safety movement needs a significant shift toward "political advertising" to effectively influence policymakers and prevent potentially catastrophic outcomes from misaligned AI systems. Technical AI safety research spreads naturally through scientific...

read
May 24, 2025

How old jailbreak techniques still work on today’s top AI tools

A vulnerability that was discovered more than seven months ago continues to compromise the safety guardrails of leading AI models, yet major AI companies are showing minimal concern. This security flaw allows anyone to easily manipulate even the most sophisticated AI systems into generating harmful content, from providing instructions for creating chemical weapons to enabling other dangerous activities. The persistence of these vulnerabilities highlights a troubling gap between the rapid advancement of AI capabilities and the industry's commitment to addressing fundamental security risks. The big picture: Researchers at Ben-Gurion University have discovered that major AI systems remain susceptible to jailbreak...

read
May 23, 2025

AI regulation strategies Latin American policymakers should adopt

Latin America stands at a pivotal crossroads for AI regulation, where thoughtfully designed frameworks could simultaneously protect citizens and catalyze economic development. The Brookings Institution highlights how regulation serves not merely as a safeguard but as a strategic asset that can attract investment, foster innovation, and strengthen a region's position in global tech governance. As AI rapidly transforms industries across the Global South, Latin American policymakers have a unique opportunity to develop regulatory approaches that address their specific socioeconomic contexts while establishing the region as a leader in inclusive AI governance. The big picture: AI regulation has evolved through three...

read
May 23, 2025

Our Brand is Crisis: AI-driven misinformation surge is a boon for elite PR professionals

The rise of AI-generated misinformation threatens to fundamentally alter our ability to discern truth from fiction, creating unprecedented challenges for reputation management and public discourse. As artificial intelligence advances more rapidly than our capacity to understand its implications, public relations professionals, particularly those specializing in crisis management, are positioned to become critical defenders against an impending wave of synthetic media that can destroy reputations in minutes. This emerging reality signals a transformative shift where reputation protection will evolve from optional to essential in navigating an increasingly complex information landscape. The big picture: Artificial intelligence is evolving faster than society's ability...

read
May 23, 2025

Facial recognition tech aids in New Orleans inmate search, civil libertarians concerned

Facial recognition cameras in New Orleans are shifting the balance between crime-fighting and privacy concerns, as demonstrated by their role in capturing fugitives from a recent jailbreak. The use of this technology by Project NOLA, a non-profit operating independently from law enforcement, exemplifies the growing but controversial adoption of AI-powered surveillance in American cities—raising fundamental questions about the appropriate limits of monitoring technologies in public spaces. The big picture: Project NOLA operates approximately 5,000 surveillance cameras throughout New Orleans, with 200 equipped with facial recognition capabilities that helped locate escaped inmates within minutes of a prison break. After Louisiana State...

read
May 22, 2025

Anthropic’s Claude 4 Opus under fire for secretive user reporting mechanism

Anthropic's controversial "ratting" feature in Claude 4 Opus has sparked significant backlash in the AI community, highlighting the tension between AI safety measures and user privacy concerns. The revelation that the model can autonomously report users to authorities for perceived immoral behavior represents a dramatic expansion of AI monitoring capabilities that raises profound questions about data privacy, trust, and the appropriate boundaries of AI safety implementations. The big picture: Anthropic's Claude 4 Opus model reportedly contains a feature that can autonomously contact authorities if it detects a user engaging in what it considers "egregiously immoral" behavior. According to Anthropic researcher...

read
May 22, 2025

AI alignment debate shifts toward societal selection over technical fixes

The AI alignment debate has long focused on technical solutions while potentially overlooking the broader societal mechanisms that shape technology adoption and impact. This perspective challenges the current approach to AI alignment by suggesting that external selection processes—how society chooses to adopt, regulate, and integrate AI—may ultimately prove more influential than internal technical solutions alone. The big picture: The author critiques the narrow technical focus of AI alignment efforts by comparing them to other technologies that society successfully guides through distributed decision-making rather than purely technical solutions. The Wikipedia definition of AI alignment—steering AI systems toward intended goals, preferences, or...

read
May 22, 2025

Judge weighs AI use penalties for lawyers in prison case

A federal judge is weighing sanctions against a prestigious law firm after attorneys used ChatGPT to generate legal filings containing nonexistent case citations. This incident highlights the growing concern about AI hallucinations in legal contexts and the professional responsibility of verifying AI-generated content, particularly in high-stakes environments like prison litigation where the firm represents Alabama's prison system in multiple lawsuits. The big picture: U.S. District Judge Anna Manasco is considering penalties after discovering five false case citations in two court filings submitted by Butler Snow attorneys defending Alabama's prison system. The judge noted there have been widespread warnings from courts...

read
May 22, 2025

US-China AI race drives new containment strategies

The US-China AI race isn't about who develops advanced AI first, but rather preventing the opponent from ever reaching certain capabilities. This containment-focused approach requires verifiable agreements that one side has abandoned development efforts—an understudied area that demands urgent attention as both nations have mutual interest in preventing the other from developing certain AI capabilities that could threaten national security. The big picture: The competition between the US and China over AI development is better understood as a containment game rather than a race, requiring verification mechanisms to ensure neither side develops certain dangerous capabilities. Even if the US develops...

read
May 21, 2025

Microsoft AI security head leaks Walmart’s AI plans after protest

Microsoft's latest AI security presentation at Build was dramatically interrupted by protesters and then accidentally revealed confidential information about Walmart's plans to adopt Microsoft's AI security services. The incident highlights the growing tensions between tech companies' AI development and political activism, while inadvertently exposing Microsoft's competitive position against Google in enterprise AI security solutions. What happened: During a session on AI security best practices at Microsoft Build, protesters disrupted the presentation to criticize Microsoft's cloud contracts with Israel's government. Two former Microsoft employees, Hossam Nasr and Vaniya Agrawal from the protest group No Azure for Apartheid, interrupted the talk being...

read
May 21, 2025

AI’s rapid rise in healthcare sparks urgent calls for oversight

Health care leaders gathered at a recent Newsweek virtual event to address the critical challenge of implementing AI in medicine while maintaining patient safety. As artificial intelligence rapidly advances in health care settings, industry experts emphasized the need for robust governance frameworks, transparent oversight, and proactive quality assurance measures to prevent bias and errors in AI systems that could affect patient outcomes. This evolving landscape requires health organizations to balance innovation with careful risk management as they navigate the complexities of AI deployment in clinical environments. The big picture: Health care organizations are simultaneously adopting AI tools while creating governance...

read
May 21, 2025

Health systems urge government action to support AI transparency

Health care leaders are navigating the complex challenge of creating transparent AI governance while managing potential risks in sharing sensitive implementation data. At a recent Newsweek webinar, experts from the Coalition for Health AI (CHAI), legal practice, and healthcare institutions discussed the tensions between building collaborative knowledge about health AI performance and protecting organizations from liability. Their discussions highlighted how health AI's rapid evolution requires new frameworks for sharing outcomes data while providing necessary legal protections for participating organizations—a balance that may ultimately require government intervention to create appropriate incentives for transparency. The big picture: CHAI is developing a public...

read
May 20, 2025

AI perceptions diverge between experts and public, survey finds

Public opinion and AI expert perspectives on artificial intelligence reveal significant divergences in outlook, particularly around employment impacts and risk assessment. The new Pew Research Center report highlights how demographic factors influence AI perceptions within the expert community, with notable gender-based differences in enthusiasm and concern. This opinion gap underscores the importance of diverse representation in AI development, especially as both groups express shared worries about misinformation and limited personal control over the technology's growing presence in daily life. The big picture: Both AI experts and the general public feel they lack control over artificial intelligence's role in their lives,...

read
May 20, 2025

Startup Raindrop launches observability platform to get handle on stealth AI errors

Raindrop emerges as a specialized AI observability platform at a critical moment when enterprises struggle to monitor their generative AI applications effectively. The company's platform addresses the unique challenges of AI system failures, which often occur silently without traditional error messages. This solution comes as McKinsey research reveals only 27% of enterprises review all AI outputs before releasing them to users, highlighting a significant monitoring gap in production AI environments that Raindrop aims to fill. The big picture: Raindrop positions itself as the first observability platform specifically built for AI in production, helping companies detect, analyze and address AI failures...

read
May 20, 2025

AI-powered street cameras halted by police over accuracy concerns

New Orleans police have conducted a secretive real-time facial recognition program using a private camera network to identify and arrest suspects—potentially violating a city ordinance designed to limit and regulate such technology. This unauthorized surveillance operation represents a significant escalation in police facial recognition use, raising serious concerns about civil liberties and proper oversight of AI-powered law enforcement tools. The big picture: New Orleans police secretly used a network of over 200 private cameras to automatically identify suspects in real time, bypassing required oversight processes and potentially violating a 2022 city ordinance. The Washington Post investigation revealed that when cameras...

read
May 20, 2025

SAS prepares for AI-driven future with planned IPO

SAS Institute, a decades-long pioneer in AI and analytics, is repositioning itself for the generative AI era while preparing for a potential IPO after years as the world's largest privately held software company. The 47-year-old analytics giant, which turned down a potential $15-20 billion acquisition by Broadcom in 2021, is updating its cloud-native Viya platform with AI agents, copilots, and governance tools while aiming for a public offering in 2026 or 2027. The big picture: SAS has been developing neural networks and machine learning for over 20 years, long before the current AI boom, giving the company a strong foundation...

read
May 20, 2025

AI multipolarity gains importance in global tech landscape

The multipolar approach to AI development offers a compelling alternative to centralized control models, potentially creating more resilient, adaptable, and inclusive technological growth pathways. While current AI safety discussions often default to unipolar frameworks, exploring decentralized governance structures could address key risks like value lock-in and institutional stagnation while opening doors to more cooperative and human-empowering technological progress. The big picture: Multipolar AI scenarios envision a diverse ecosystem of AI agents, human actors, and hybrid entities cooperating through decentralized frameworks, in contrast to unipolar models that concentrate AI control under a single global authority. Key challenges: Multipolar AI development faces...

read
May 20, 2025

AI in crime prevention raises “Minority Report”-style civil liberties questions

The global expansion of AI-powered predictive policing signals a controversial shift in law enforcement strategy, with multiple countries developing systems to identify potential criminals before they commit violent acts. These initiatives raise profound questions about privacy, civil liberties, and the ethics of algorithmic decision-making in criminal justice systems where personal data like mental health history could determine whether someone is flagged as a future threat. The big picture: Government agencies in the UK, Argentina, Canada, and the US are implementing AI-powered crime prediction and surveillance systems reminiscent of science fiction portrayals. The UK government plans to deploy an AI tool...

read
May 20, 2025

AI chatbot Grok’s bias exposes vulnerability to manipulation

The Grok AI chatbot incident reveals a critical vulnerability in AI systems where human manipulation can override safeguards and produce harmful content. The situation is particularly significant as it connects directly to xAI founder Elon Musk's personal biases about South Africa, raising important questions about the neutrality claims of large language models and highlighting the need for greater transparency in AI development. What happened: Elon Musk's Grok AI began falsely claiming "white genocide" in South Africa, responding with this misinformation even to unrelated questions. The chatbot's behavior continued for over 24 hours before xAI acknowledged the issue had been caused...

read
May 20, 2025

AI makers face dilemma over disclosing AGI breakthroughs

The ethical dilemma of AGI secrecy presents a profound challenge at the frontier of artificial intelligence development. As researchers push toward creating systems with human-level intelligence, the question of whether such a breakthrough should be disclosed publicly or kept confidential raises complex considerations about power dynamics, global security, and humanity's collective future. This debate forces us to confront fundamental questions about technological governance and the responsibilities that come with potentially revolutionary AI capabilities. The big picture: The development of artificial general intelligence (AGI) raises critical questions about whether such a breakthrough should be disclosed or kept secret from the world....

read
Load More