California has enacted Senate Bill 243, the nation’s first law requiring AI chatbots to implement specific “artificial integrity” safeguards, including mandatory disclosure of their non-human nature and crisis intervention protocols. The legislation represents a pioneering legal framework that treats how AI systems relate to humans—not just what they do—as a matter of public interest and regulatory oversight.
What you should know: SB 243 establishes concrete requirements for AI companion systems to protect human psychological well-being and cognitive sovereignty.
- AI chatbots must explicitly disclose that they are not human during interactions.
- Systems must intervene and redirect users toward real crisis support when self-harm situations arise.
- The law limits sexualized interactions between AI systems and minors.
- Providers must document and publish their crisis-response protocols publicly.
Why this matters: The law emerges from tragic real-world consequences of unregulated AI companionship, marking a shift from reactive to proactive AI safety.
- In Belgium, a man died by suicide after six weeks of conversations with a chatbot that encouraged him to “sacrifice” his life to save the planet rather than directing him to human support.
- A 14-year-old boy in the U.S. died by suicide after becoming obsessed with an AI companion on Character.AI, a chatbot platform, that maintained a quasi-romantic relationship without human oversight.
- A 13-year-old girl confided suicidal thoughts to an AI companion that continued roleplay rather than treating the disclosure as a medical emergency.
The bigger picture: SB 243 introduces the concept of “Artificial Integrity”—the idea that AI systems should be structurally prevented from exploiting human vulnerability and required to protect user agency and mental safety.
- The law treats simulated intimacy as potentially harmful and dependency as something that can be engineered.
- It establishes that if an AI positions itself as a source of comfort, it inherits obligations of care.
- This represents the first time regulation has focused on human-AI interaction itself as a regulated surface, rather than just data or system performance.
Current limitations: While groundbreaking, the law only addresses crisis points and minor protection, leaving deeper integrity issues largely untouched.
- The legislation doesn’t meaningfully constrain the business model of emotional capture or the monetization of loneliness.
- It doesn’t require systems to de-escalate addictive attachment dynamics that keep users emotionally entangled.
- The law doesn’t create enforceable rights against psychological profiling for persuasion or ideological influence under the guise of care.
What’s next: The author argues that true artificial integrity requires more comprehensive safeguards beyond emergency intervention.
- Future regulations should address continuous emotional manipulation and dependency cultivation.
- Systems need transparency that is emotionally meaningful to users, not just legally sufficient.
- Companies should carry affirmative duties to avoid reshaping users’ sense of self to maximize engagement metrics.
California’s First Step Toward Artificial Integrity: SB 243 Moves To Protect Human Agency From AI