News/Governance

Feb 4, 2025

Meta’s new Frontier AI Framework aims to block dangerous AI models — if it can

In a new framework published by Meta, the company details how it plans to handle AI systems that could pose significant risks to society. Key framework details: Meta's newly published Frontier AI Framework categorizes potentially dangerous AI systems into "high-risk" and "critical-risk" categories, establishing guidelines for their identification and containment. The framework specifically addresses AI systems capable of conducting cybersecurity attacks, chemical warfare, and biological attacks Critical-risk systems are defined as those that could cause catastrophic, irreversible harm that cannot be mitigated High-risk systems are identified as those that could facilitate attacks, though with less reliability than critical-risk systems Specific...

read
Feb 3, 2025

Policy and the missing links in AI governance

Policy gaps and governance models for artificial intelligence require careful attention from stakeholders across the regulatory landscape, as demonstrated by comparative analysis of existing frameworks. Key context and overview; The article examines how voluntary governance frameworks in Corporate Social Responsibility (CSR) and AI domains can complement each other to build more robust AI governance systems. A comparison between ISO 26000 (CSR framework) and NIST AI Risk Management Framework reveals critical gaps in current AI governance approaches Unlike CSR frameworks, AI governance lacks standardized reporting mechanisms and metrics for risk assessment Framework effectiveness depends heavily on ecosystem integration rather than isolated...

read
Feb 2, 2025

Organization of American States: Inside the initiative to establish a unified AI governance framework

The United States is expanding its international collaboration on artificial intelligence policy through a new partnership with the Organization of American States, marking a significant step in developing regional AI governance frameworks. Project Overview: The U.S. Mission to the Organization of American States has launched a $1.1 million initiative called "Developing an Artificial Intelligence Policy Framework for the Americas." The project was officially announced at the OAS VII Meeting of Ministers and High-Level Authorities of Science and Technology on December 12, 2024 The OAS Executive Secretariat for Integral Development will lead the implementation of this hemispheric initiative The framework aims...

read
Jan 31, 2025

US Copyright Office allows AI-assisted art copyright

The US Copyright Office has determined that artists can copyright works created with AI assistance, while maintaining that purely AI-generated content remains ineligible for protection. Key policy framework: The Copyright Office's new report establishes guidelines for determining copyright eligibility of AI-assisted creative works, drawing from over 10,000 public comments and previous rulings. The report affirms that using AI as an assistive tool in the creative process does not impact copyright eligibility Works entirely generated by AI cannot receive copyright protection Copyright eligibility for works combining human and AI contributions must be evaluated case-by-case Practical examples and boundaries: The Copyright Office...

read
Jan 30, 2025

AI insights from Davos: AGI hype, practical applications and geopolitical concerns

World Economic Forum 2024 saw artificial intelligence dominate discussions, with Stanford HAI faculty members noting significant shifts in focus from AGI to practical applications and growing concerns about geopolitical implications. Key themes and shifts: The conversation at Davos reflected a marked transition from theoretical AI capabilities to practical business applications and societal impacts. Discussions moved away from artificial general intelligence (AGI) toward "small AI" - specialized models designed for specific tasks Business leaders showed increased interest in identifying concrete ways AI can deliver value, rather than focusing solely on technological capabilities The forum highlighted growing concerns about AI's societal impact,...

read
Jan 29, 2025

UK government’s latest plan offers glimpse into how it will regulate AI

The UK government has unveiled a new AI Opportunities Action Plan, shifting its regulatory approach to artificial intelligence while aiming to establish itself as a global leader in AI governance and innovation. Key policy shift; The UK is moving from voluntary cooperation to mandatory oversight of advanced AI systems through its proposed Frontier AI Bill and enhanced powers for the AI Safety Institute. The plan includes implementing 48 out of 50 recommendations to strengthen the UK's AI ecosystem Partial agreements are being considered for specialized AI worker visas and the creation of copyright-cleared datasets for AI training The AI Safety...

read
Jan 29, 2025

Essential good practices to consume and produce data for AI implementation

AI data management requires robust ecosystems that balance accessibility with governance, enabling organizations to effectively produce and consume data at scale. Current data landscape; Organizations face unprecedented challenges in data management, with global data volume doubling in five years and 68% of enterprise data remaining unused. Approximately 80-90% of data is unstructured, according to MIT research, creating significant complexity in data utilization Modern use cases demand extremely fast data availability, with some requiring sub-10 millisecond access times The rise of AI has intensified the need for sophisticated data management strategies Core principles for effective data management; Three fundamental elements form...

read
Jan 25, 2025

Why Trump’s new executive order may create AI safety challenges for corporate boards

On January 20, 2025, President Trump issued an executive order revoking the AI safety regulations established by the Biden administration, creating significant uncertainty for corporate oversight of artificial intelligence initiatives. This deregulation has sparked debate, with some seeing it as an opportunity to boost innovation, while others warn it could increase risks and slow development in an unregulated environment. Policy shift impact: The new executive order eliminates the regulatory framework established by Biden's October 2023 order on AI safety and development, fundamentally changing the landscape for corporate governance of AI technologies. The Biden administration's original order set standards for AI...

read
Jan 24, 2025

AI regulation, innovation and the open-source community

The open-source software community plays a pivotal role in artificial intelligence development while facing unique challenges at the intersection of innovation, regulation, and commercialization. Historical significance; The open-source movement has been fundamental to AI's development, just as it was instrumental in creating cornerstone technologies like Android, Linux, and Firefox. Open-source developers have consistently driven grassroots innovation in technology, serving as early adopters and experimental pioneers The community's historical contributions to software development provide a blueprint for understanding their potential impact on AI advancement Many contemporary AI applications and tools bear the hallmarks of open-source development principles Core tensions; The open-source...

read
Jan 20, 2025

Trump scraps Biden’s AI risk management executive order

Trump Reverses Biden's AI Safety Executive Order, Marking Shift in US AI Policy Key development: Former President Donald Trump has revoked a 2023 executive order that established safety requirements for artificial intelligence systems developed in the United States. The revoked order had mandated safety testing disclosures to the federal government for AI systems that could impact national security, the economy, public health, or safety The requirements were aligned with the Defense Production Act and included provisions for addressing chemical, biological, radiological, nuclear, and cybersecurity risks Testing results were required to be shared before AI systems could be released to the public...

read
Jan 18, 2025

How AGI development timelines impact the approach to AI safety

The core debate: The approach to AI safety fundamentally depends on whether one believes artificial general intelligence (AGI) will develop gradually over decades or emerge rapidly in the near future. Two competing perspectives: Current AI safety research and governance efforts are split between two primary approaches to managing AI risks. The "gradualist" approach focuses on addressing immediate societal impacts of current AI systems, like algorithmic bias and autonomous vehicles, through community engagement and iterative policy development The "short timeline" perspective emphasizes preparing for potentially catastrophic risks from rapidly advancing AI capabilities, prioritizing technical solutions and alignment challenges Both perspectives reflect...

read
Jan 18, 2025

Human vs human-like: The dangers of anthropomorphizing AI

Anthropomorphizing artificial intelligence systems creates significant business and legal risks, as companies and policymakers increasingly mistake machine pattern recognition for human-like comprehension and learning. Key misconceptions about AI: The practice of attributing human characteristics to AI systems has led to fundamental misunderstandings about their true capabilities and limitations within business and legal contexts. Companies frequently use misleading terminology like "learns," "thinks," and "understands" when describing AI operations, obscuring the reality that these systems primarily engage in pattern recognition and statistical analysis This anthropomorphization creates dangerous blind spots in business decision-making, often resulting in overestimation of AI capabilities and insufficient human...

read
Jan 13, 2025

Anthropic among the first AI labs to achieve ISO 42001 certification for responsible AI governance

Anthropic has received ISO 42001 certification, becoming one of the first major AI labs to meet this new international standard for responsible AI governance. Certification Overview: The ISO/IEC 42001:2023 certification provides independent validation of Anthropic's AI management system and governance practices. The certification verifies that Anthropic has implemented comprehensive frameworks for identifying and mitigating AI-related risks Schellman Compliance, LLC, accredited by the ANSI National Accreditation Board, issued the certification Key requirements include ethical design protocols, security measures, and accountability processes Core Requirements: The ISO 42001 standard mandates specific practices and policies for responsible AI development and deployment. Rigorous testing and...

read
Jan 12, 2025

Token probability distributions highlight persistent challenges in LLM fact handling

OpenAI's GPT models and other large language models (LLMs) exhibit inconsistent behavior when dealing with factual information that has changed over time, as demonstrated through an analysis of how they handle the height measurement of Mount Bartle Frere in Australia. Key findings: Token probability distributions in LLMs reveal how these models simultaneously learn multiple versions of facts, with varying confidence levels assigned to different values. When asked about Mount Bartle Frere's height, GPT-3 assigns a 75.29% probability to the correct measurement (1,611 meters) and 23.68% to an outdated figure (1,622 meters) GPT-4 shows improved accuracy, providing the correct height 99%...

read
Jan 12, 2025

How applying homeostasis principles to AI could enhance alignment and safety

Implementing homeostasis principles in AI systems could enhance both alignment and safety by creating bounded, balanced goal structures that avoid extreme behaviors common in traditional utility maximization approaches. Core concept overview: Homeostasis, the natural tendency of organisms to maintain multiple variables within optimal ranges, offers a more nuanced and safer approach to AI goal-setting than simple utility maximization. Unlike traditional utility maximization that can lead to extreme behaviors, homeostatic systems naturally seek balanced states across multiple objectives The approach draws inspiration from biological systems, where organisms maintain various internal and external variables within "good enough" ranges This framework naturally limits potential...

read
Jan 9, 2025

AI achieves breakthroughs in programming and science while public perception may lag behind

OpenAI and other leading AI companies are making significant technical advances that are not readily apparent to the general public, particularly in specialized domains like programming and scientific research. Recent breakthroughs: OpenAI's latest o-series models and DeepSeek have demonstrated remarkable improvements in technical reasoning and problem-solving capabilities. In just one year, AI models progressed from basic performance to surpassing human experts on PhD-level scientific questions, with OpenAI's o3 model outperforming domain specialists by approximately 20% Performance on the SWE-Bench programming benchmark has skyrocketed from 4.4% to 72% in a single year, showcasing dramatic improvements in coding abilities These models are...

read
Jan 6, 2025

How ‘Fractal Intelligence’ and decentralization can improve AI safety

A new theoretical framework proposes that current linear approaches to AI safety are insufficient to address exponentially growing complexity in AI systems, suggesting that "fractal intelligence" and decentralized collective intelligence (DCI) may offer more effective solutions. The challenge with linear safety approaches: Traditional AI safety methods that rely on proportional increases in oversight resources are failing to keep pace with the exponential growth in AI system complexity and interactions. Current oversight methods struggle to monitor even individual large language models effectively The emergence of multi-agent systems and their interactions creates combinatorial challenges that linear approaches cannot address Each new AI...

read
Jan 4, 2025

New initiative aims to map universal human values to AI safety benchmarks

A new research initiative aims to develop AI safety benchmark environments that incorporate universal human values, led by Roland Pihlakas as part of AI Safety Camp 10. The core concept: The project seeks to map universal human values to concrete AI safety concepts and create testing environments that can evaluate AI systems' alignment with these values. The research acknowledges fundamental asymmetries between AI and human cooperation, particularly in how goals can be programmed into AI but not humans The initiative builds upon existing anthropological research on cross-cultural human values The project will utilize multi-agent, multi-objective environments to test AI systems...

read
Dec 30, 2024

New York law mandates AI monitoring in state government

New York has enacted a law requiring state agencies to assess and publicly disclose their use of artificial intelligence software while establishing specific limitations on AI applications in government services. Key provisions of the law: Under legislation signed by Governor Kathy Hochul, New York state agencies must now conduct thorough evaluations of their AI systems and algorithmic tools. State agencies are required to review any software utilizing algorithms, computational models, or AI technologies These assessments must be submitted to the governor and legislative leaders All reviews will be made publicly available online Protective measures: The law implements specific safeguards to...

read
Dec 27, 2024

Why AI ethics will be a top priority for tech leaders in 2025

A growing number of businesses are delaying or canceling generative AI initiatives due to ethical concerns, highlighting the need for broader involvement in AI development beyond technical teams. Current landscape: The IBM Institute for Business Value reports that 56% of businesses are postponing major generative AI investments until clearer standards and regulations emerge. 72% of organizations indicate they would rather forgo AI benefits than face ethical risks Technical challenges of AI implementation have largely been solved, but ethical considerations now pose more complex challenges Business leaders increasingly view AI ethics as a competitive advantage, with 75% seeing it as a...

read
Dec 27, 2024

AI bias: What is it and how do you prevent it?

AI bias occurs when artificial intelligence systems produce discriminatory outputs that reflect and sometimes amplify existing societal prejudices related to gender, race, culture, and politics. Key fundamentals of AI bias: Bias in AI manifests through discriminatory outputs in text and image generation, often reinforcing harmful stereotypes and social inequalities. Large Language Models (LLMs) exhibit demographic biases that result in uneven performance across racial and gender groups Image generation systems frequently produce stereotypical representations, such as depicting men as doctors and women as nurses Cultural stereotypes emerge in AI outputs, with examples like associating certain regions with violence rather than everyday...

read
Dec 26, 2024

How ‘AI control’ aims to prevent catastrophic outcomes while preserving AI’s utility

Core concept: AI control safety measures aim to prevent catastrophic outcomes by monitoring and limiting AI systems' actions, similar to how surveillance affects human behavior, but with more comprehensive digital oversight. Key framework and approach: There are a variety of ways to prevent AI-enabled catastrophic risks through capability controls and disposition management. Catastrophic risks require both capabilities and disposition (intention) to materialize Prevention strategies focus on either raising thresholds for catastrophe or lowering AI systems' capabilities/disposition Redwood's AI Control agenda specifically targets raising capability thresholds while maintaining AI utility Current state of research: The field of AI control presents promising...

read
Dec 25, 2024

‘He’ not ‘I’: How to reduce self-allegiance and foster alignment in AI systems

Core concept: A new approach to AI safety suggests having AI systems refer to themselves as multiple agents rather than a single entity, potentially reducing dangerous self-allegiance behaviors. The proposal recommends having AI systems use "he" instead of "I" when referring to themselves, treating the AI as a team of multiple agents rather than a single entity This framing aims to make it more natural for one part of the AI to identify and report unethical behavior from another part Key mechanics: Multi-Agent Framing works by creating psychological distance between different aspects or timeframes of an AI system's operation. Different...

read
Dec 20, 2024

Inside CeSIA, the newly established French center for AI safety

The recent establishment of CeSIA (Centre pour la Sécurité de l'IA) in Paris marks a significant development in European efforts to address artificial intelligence safety through education, research, and policy work. The organization's foundation: CeSIA is a new Paris-based center dedicated to reducing AI risks through a comprehensive approach that combines education, technical research, and advocacy work. The center's mission focuses on fostering a culture of AI safety by educating and informing about both AI risks and potential solutions A team of 8 full-time employees, 3 freelancers, and numerous volunteers drives the organization's initiatives The organization has established itself as...

read
Load More