Meta Platforms META-Q used public Facebook and Instagram posts to train parts of its new Meta AI virtual assistant, but excluded private posts shared only with family and friends in an effort to respect consumers’ privacy, the company’s top policy executive told Reuters in an interview.
Meta also did not use private chats on its messaging services as training data for the model and took steps to filter private details from public datasets used for training, said Meta president of global affairs Nick Clegg, speaking on the sidelines of the company’s annual Connect conference this week.
“We’ve tried to exclude datasets that have a heavy preponderance of personal information,” Mr. Clegg said, adding that the “vast majority” of the data used by Meta for training was publicly available.
He cited LinkedIn as an example of a website whose content Meta deliberately chose not to use because of privacy concerns.
Mr. Clegg’s comments come as tech companies including Meta, OpenAI and Alphabet’s Google have been criticized for using information scraped from the internet without permission to train their AI models, which ingest massive amounts of data in order to summarize information and generate imagery.
The companies are weighing how to handle the private or copyrighted materials vacuumed up in that process that their AI systems may reproduce, while facing lawsuits from authors accusing them of infringing copyrights.
Meta AI was the most significant product among the company’s first consumer-facing AI tools unveiled by CEO Mark Zuckerberg on Wednesday at Meta’s annual products conference, Connect. This year’s event was dominated by talk of artificial intelligence, unlike past conferences which focused on augmented and virtual reality.
Meta made the assistant using a custom model based on the powerful Llama 2 large language model that the company released for public commercial use in July, as well as a new model called Emu that generates images in response to text prompts, it said.
The product will be able to generate text, audio and imagery and will have access to real-time information via a partnership with Microsoft’s Bing search engine.
The public Facebook and Instagram posts that were used to train Meta AI included both text and photos, Mr. Clegg said.
Those posts were used to train Emu for the image generation elements of the product, while the chat functions were based on Llama 2 with some publicly available and annotated datasets added, a Meta spokesperson told Reuters.
Interactions with Meta AI may also be used to improve the features going forward, the spokesperson said.
Mr. Clegg said Meta imposed safety restrictions on what content the Meta AI tool could generate, like a ban on the creation of photo-realistic images of public figures.
On copyrighted materials, Mr. Clegg said he was expecting a “fair amount of litigation” over the matter of “whether creative content is covered or not by existing fair use doctrine,” which permits the limited use of protected works for purposes such as commentary, research and parody.
“We think it is, but I strongly suspect that’s going to play out in litigation,” Mr. Clegg said.
Some companies with image-generation tools facilitate the reproduction of iconic characters like Mickey Mouse, while others have paid for the materials or deliberately avoided including them in training data.
OpenAI, for instance, signed a six-year deal with content provider Shutterstock this summer to use the company’s image, video and music libraries for training.
Asked whether Meta had taken any such steps to avoid the reproduction of copyrighted imagery, a Meta spokesperson pointed to new terms of service barring users from generating content that violates privacy and intellectual property rights.