Meta's AI Training Resumes with U.K. Public Content, Following Regulatory Approval
Meta Platforms has restarted training its generative AI models in the U.K. after pausing the process in July due to regulatory concerns.
The company will now use public content shared by adult users on Facebook and Instagram, including posts, captions, comments, and photos.
While this decision faced significant scrutiny, Meta has clarified that private messages and content from users under the age of 18 will not be part of the data collection.
The move comes after successful negotiations with the Information Commissioner’s Office (ICO), which had earlier raised concerns about the privacy implications of using such data.
Nick Clegg, Meta’s president of Global Affairs, emphasised,
“We do not use people’s private messages with friends and family to train for AI at Meta, and we do not use information from accounts of people in the UK under the age of 18.”
He further added that only public content will be leveraged to enhance Meta’s generative AI systems, which power features and experiences for U.K. users.
The company’s goal is to develop AI models that better understand language and context, and the use of U.K. public posts is framed as a way to reflect British culture, history, and idioms.
However, the underlying objective remains clear: to refine AI’s ability to mimic human interaction.
Navigating Regulatory Backlash and Privacy Concerns
Meta’s decision to use public content from its platforms has not been without resistance.
The company had initially halted the AI training in June after the Irish privacy regulator called for a delay, while Britain’s ICO echoed these concerns.
Advocacy groups like NOYB also voiced strong opposition, urging privacy watchdogs across Europe to intervene.
They argued that Meta’s notifications to users about the use of their data were insufficient to comply with stringent EU privacy rules.
However, after months of dialogue, Meta has secured the necessary approval from U.K. authorities under legal provisions related to "legitimate interests."
The ICO acknowledged that Meta had made notable improvements to its approach, including simplifying the process for users to object to their data being used.
Moreover, the time window for objections has been extended, allowing more users to opt out of this data collection.
An ICO spokesperson remarked,
“Meta has simplified the way users can object to the processing of their data and has extended the window in which they can do so.”
Meta’s Strategic Framing of AI Data Collection
Despite the regulatory hurdles, Meta has sought to frame its use of public data in a positive light, portraying it as a way to embed local culture into its AI models.
The company said in a statement:
"This means that our generative AI models will reflect British culture, history, and idiom."
However, the real focus is on improving the AI’s understanding of human interaction by analysing the way people communicate publicly on its platforms.
The company’s ambition is not unique.
Major tech players, including Google with its Gemini AI and Samsung with Galaxy AI, are also aggressively pushing forward in the AI space.
Apple is similarly integrating its AI technologies across devices, with “Apple Intelligence” poised to revolutionise user experiences.
Meta’s use of public social media data, though legal, continues to raise ethical questions.
With AI’s potential to disrupt industries and replace jobs, public unease around data scraping for model training is unlikely to dissipate anytime soon.
User Awareness and Resistance in a Growing AI Landscape
Meta plans to issue in-app notifications to U.K. users starting next week, informing them about the AI training process and how they can object to the use of their data.
Despite these efforts, many users remain sceptical about the broader implications of AI-driven data collection.
While AI models today rely heavily on public data from the internet, privacy concerns are growing.
As websites become more restrictive, often blocking AI from accessing data unless paid for, companies like Meta find their social platforms to be convenient sources for training models.
However, Meta's attempts to normalise this data usage face ongoing criticism, particularly regarding its moral implications.
A growing number of people are uncomfortable with the idea of their online activity being used to train AI, especially as many fear job losses or find the technology invasive.
Yet, the AI boom shows no signs of slowing down.
As Meta continues to push forward with AI-powered tools like AR smart glasses and integrated AI features in services such as WhatsApp, the company is betting on AI as the future of computing.
Meta’s challenges are far from over, as it still faces regulatory scrutiny across the European Union.
While the U.K. and Brazil have granted permission for AI training, the European Union remains cautious.
The question remains – how long can companies like Meta continue leveraging public data for AI training before the growing backlash forces stricter regulations?