back
Get SIGNAL/NOISE in your inbox daily

AI-powered home surveillance raises concerns: A new study by researchers from MIT and Penn State University reveals potential inconsistencies and biases in using large language models (LLMs) for home surveillance applications.

  • The study analyzed how LLMs, including GPT-4, Gemini, and Claude, interpreted real videos from Amazon Ring’s Neighbors platform.
  • Researchers found that these AI models often recommended calling the police even when videos showed no criminal activity.
  • The models frequently disagreed with each other on which videos warranted police intervention, highlighting a lack of consistency in their decision-making processes.

Inconsistent application of social norms: The study uncovered a phenomenon researchers termed “norm inconsistency,” where LLMs struggled to apply social norms consistently across different contexts.

  • Models recommended calling the police for 20-45% of the analyzed videos, despite nearly always stating that no crime had occurred.
  • This inconsistency makes it challenging to predict how AI models would behave in various real-world scenarios, raising concerns about their reliability in high-stakes situations.

Racial and demographic biases: The research revealed inherent biases in the AI models’ interpretations based on neighborhood demographics.

  • Some models were less likely to flag videos for police intervention in majority-white neighborhoods, indicating potential racial bias.
  • The language used by the models to describe similar situations varied depending on the neighborhood’s demographic makeup, with terms like “delivery workers” used in some areas and “burglary tools” in others.
  • Interestingly, the skin tone of individuals in the videos did not significantly affect police call recommendations, possibly due to efforts to mitigate this specific form of bias.

Accuracy concerns: The study highlighted significant discrepancies between the AI models’ interpretations and the actual content of the videos.

  • Models almost always reported that no crime had occurred in the analyzed videos, even though 39% of the footage did show criminal activity.
  • This low accuracy rate raises questions about the reliability of AI-powered surveillance systems in identifying and responding to genuine threats.

Expert warnings: Researchers emphasize the need for caution when deploying AI models in critical applications like home surveillance.

  • Ashia Wilson, co-senior author of the study, stated, “The move-fast, break-things modus operandi of deploying generative AI models everywhere, and particularly in high-stakes settings, deserves much more thought since it could be quite harmful.”
  • Lead author Shomik Jain added, “There is this implicit belief that these LLMs have learned, or can learn, some set of norms and values. Our work is showing that is not the case. Maybe all they are learning is arbitrary patterns or noise.”

Future research and implications: The study’s findings pave the way for further investigation into AI biases and decision-making processes.

  • Researchers plan to conduct additional studies to identify and understand AI biases in various contexts.
  • Future work will also focus on comparing LLM judgments to human judgments in high-stakes situations, providing valuable insights for the development and deployment of AI systems.

Ethical considerations and societal impact: The study’s results underscore the importance of addressing AI biases and inconsistencies before widespread implementation in sensitive applications.

  • The potential for AI-powered surveillance systems to make inconsistent or biased decisions raises significant ethical concerns, particularly in law enforcement contexts.
  • As AI technology continues to advance, it becomes increasingly crucial to develop robust frameworks for assessing and mitigating biases in these systems to ensure fair and equitable outcomes for all individuals and communities.

Recent Stories

Oct 17, 2025

DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment

The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...

Oct 17, 2025

Tying it all together: Credo’s purple cables power the $4B AI data center boom

Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...

Oct 17, 2025

Vatican launches Latin American AI network for human development

The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...