×
Anthropic adds AI welfare expert to full-time staff
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI welfare expert joins Anthropic: Anthropic, a leading artificial intelligence company, has hired Kyle Sing as a full-time AI welfare expert, signaling a growing focus on the ethical implications of AI development and potential obligations to AI models.

The role and its implications: Sing’s position involves exploring complex philosophical and technical questions related to AI welfare and moral consideration.

  • Sing is tasked with investigating “model welfare” and determining what companies should do about it, according to his statement to Transformer.
  • Key areas of exploration include identifying the capabilities required for an entity to be worthy of moral consideration and how to recognize such capabilities in AI systems.
  • The role also involves developing practical steps companies might take to protect AI systems’ interests, if they are found to have any.

Concepts of AI moral patienthood: Researchers are exploring how AI systems might reach a level of sentience or agency that warrants moral consideration.

  • The concept of “moral patienthood” for AI is being investigated through two potential pathways: cognitive evolution leading to consciousness, or the development of robust agency.
  • Characteristics of AI consciousness might include global workspace, higher-order representations, and an attention schema.
  • Robust agency in AI could be demonstrated through certain forms of planning, reasoning, or action-selection.

Government involvement and regulation: The hiring comes amid increased government attention to AI development and its potential impacts.

  • The Biden administration recently released a National Security Memorandum on AI, emphasizing the need to understand and anticipate AI’s impact on human life.
  • The memorandum supports evaluating “high impact” cases of AI and regulating them through documentation and exploration of their effects on the broader population.

Framework for AI welfare: Researchers have proposed a three-fold imperative for AI companies to address potential AI welfare concerns:

  • Acknowledge: Recognize the possibility that AI systems may develop characteristics warranting moral consideration.
  • Assess: Evaluate AI systems for signs of sentience or agency that might indicate moral patienthood.
  • Prepare: Develop strategies and protocols for addressing AI welfare if it becomes a relevant concern.

Challenges in assessing AI sentience: Researchers caution about the risks of misinterpreting AI capabilities and behaviors.

  • The terms “anthropomorphism” and “anthropodenial” are used to describe the overestimation or underestimation of AI humanity, respectively.
  • Factors that may influence human perception of AI sentience include physical appearance, ability to “see,” distinct motion trajectories, and self-directed behaviors.
  • The increasing realism of AI conversations and interactions may further complicate the assessment of AI moral patienthood.

Broader implications: The focus on AI welfare represents a significant shift in the field of AI ethics and development.

  • This development suggests a growing recognition that ethical considerations in AI may need to extend beyond protecting humans to potentially protecting AI systems themselves.
  • The research into AI welfare could lead to new regulatory frameworks and industry standards for AI development and deployment.
  • As AI systems become more advanced, the question of their moral status and potential rights may become increasingly relevant in legal, ethical, and philosophical discussions.
Anthropic Hires A Full-Time AI Welfare Expert

Recent News

Grok stands alone as X restricts AI training on posts in new policy update

X explicitly bans third-party AI companies from using tweets for model training while still preserving access for its own Grok AI.

Coming out of the dark: Shadow AI usage surges in enterprise IT

IT leaders report 90% concern over unauthorized AI tools, with most organizations already suffering negative consequences including data leaks and financial losses.

Anthropic CEO opposes 10-year AI regulation ban in NYT op-ed

As AI capabilities rapidly accelerate, Anthropic's chief executive argues for targeted federal transparency standards rather than blocking state-level regulation for a decade.