×
Written by
Published on
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

LinkedIn’s AI training initiative: LinkedIn has implemented a new policy that allows the company to use user data for training generative AI models, with users automatically opted in without explicit consent.

Opting out of AI training: Users who wish to prevent their data from being used for future AI model training must take specific steps to opt out.

  • To opt out, users need to navigate to the Data privacy tab in their account settings and toggle off the “Data for Generative AI Improvement” option.
  • LinkedIn clarifies that opting out will only prevent future use of personal data for AI training and does not affect any training that has already taken place.
  • The company claims to use privacy-enhancing technologies to redact or remove personal data from its training sets.

Additional opt-out requirements: LinkedIn’s AI training policy extends beyond generative AI models, requiring users to take extra steps to fully protect their data.

  • The platform uses other machine learning tools for purposes such as personalization and moderation, which do not generate content.
  • To opt out of data usage for these non-generative AI tools, users must separately fill out the LinkedIn Data Processing Objection Form.
  • This two-step opt-out process highlights the complexity of data usage policies and the potential for user confusion.

Geographic considerations: LinkedIn’s AI training policy varies based on user location, with certain regions exempt from data collection.

  • Users residing in the European Union, European Economic Area, or Switzerland are not included in the AI model training program.
  • This geographic distinction underscores the impact of regional data protection regulations on corporate AI development practices.

Implications for user privacy: LinkedIn’s decision to automatically opt users into AI training raises concerns about data privacy and user consent.

  • The lack of proactive notification about this significant change in data usage has sparked criticism from privacy advocates.
  • This incident highlights the ongoing debate surrounding the balance between technological advancement and individual privacy rights in the digital age.
  • Users may be unaware of how their professional and personal information shared on the platform could be utilized in AI development.

Broader context of AI data collection: LinkedIn’s approach to AI training data collection reflects a growing trend among tech companies.

  • The revelation comes amidst increased scrutiny of how major tech platforms acquire and use user data for AI development.
  • This incident, along with Meta’s recent admission, suggests that the practice of utilizing user data for AI training may be more widespread than previously known.
  • It raises questions about the transparency of tech companies regarding their data usage policies and the extent of user control over personal information.

Analyzing the implications: LinkedIn’s AI training policy underscores the complex relationship between user data, technological innovation, and privacy concerns in the digital age.

  • The automatic opt-in approach taken by LinkedIn may set a precedent for other platforms, potentially normalizing the use of user data for AI training without explicit consent.
  • This incident highlights the need for increased transparency from tech companies about their data usage practices and more robust regulations to protect user privacy in the era of AI development.
  • As AI continues to advance, the balance between leveraging user data for innovation and respecting individual privacy rights will likely remain a contentious issue, requiring ongoing scrutiny and dialogue.
LinkedIn is training AI models on your data

Recent News

You Can Now Search Google Chrome History with Natural Language

Chrome's AI-powered history search lets users find past websites using natural language queries, with local data storage and privacy controls in place.

Your Favorite K-Pop Hits May Actually Be Illicit AI Clones

K-pop artists dominate AI-generated music on YouTube, with Blackpink leading at 17.3 million views and potentially losing $500,000 in revenue.

Michael Dell Shares Leadership Lessons and Thoughts on the Future of AI

Dell's CEO envisions AI as a catalyst for scientific breakthroughs and enhanced human capabilities, with applications spanning drug discovery to supply chain optimization.