A Human Rights Watch investigation has revealed that photos of real children posted online are being used to train AI image generators without consent, posing significant privacy and safety risks.
Key findings from Australia: HRW researcher Hye Jung Han discovered 190 photos of Australian children, including indigenous kids, linked in the LAION-5B AI dataset:
Unique risks for indigenous children: For First Nations children in Australia, AI training on their images threatens distinct cultural harms:
Limitations of current safeguards: Removing links from datasets and implementing content guidelines appear insufficient to prevent ongoing harm:
Waiting on regulatory intervention: HRW argues that the onus should not be on parents to remove kids’ photos, but on regulators to enact robust child data protection laws:
Broader implications: This investigation highlights the urgent need for stricter regulations and enforceable safeguards around AI training data, especially when it comes to protecting children’s privacy and safety online. As AI systems become more powerful and pervasive, the risks of unauthorized data scraping and misuse will only grow. Policymakers, tech companies, and civil society groups must work together to develop robust frameworks that prioritize human rights and prevent AI from being trained on sensitive personal data without clear consent procedures in place. Crucially, the burden of protecting kids’ digital footprints cannot fall solely on parents – systemic solutions and strong regulatory oversight are essential.