Meta has embarked on a mission to detect and identify content generated by artificial intelligence (AI), aiming to provide transparency to its users.
In a recent article by Meta’s President of Global Affairs, Nick Clegg, the company addressed the challenge of discerning authentic content from AI-generated content and outlined its strategy for managing this emerging technology.
Clegg announced that Meta is collaborating with industry partners to establish common standards or signals for detecting AI-generated content. He emphasized that detecting these signals would enable Meta to label AI-generated images posted on Facebook, Instagram, and Threads. The company is actively developing this capability and plans to roll out labels in all supported languages across its apps in the coming months.
While Meta already labels content generated by its own AI tool, Clegg stated that the company intends to extend this labeling to content produced by external sources such as Google, OpenAI, Microsoft, Adobe, Midjourney, and Shutterstock.
Highlighting the significance of this initiative, Clegg emphasized the potential adversarial nature of the AI-generated content space in the future. He stressed the need for continuous vigilance to thwart attempts by individuals and organizations seeking to deceive users with AI-generated content.
In December, Meta announced the testing of generative AI features across its social media platforms, including search, ads, and business messaging. Ahmad Al-Dahle, Meta’s vice president of generative AI, reiterated the company’s commitment to building a better community and enhancing user experience through innovative products.
Additionally, Meta expanded its image editing tool, Imagine, beyond chats to a website where users can create free images. To differentiate AI-generated images, Meta implemented invisible watermarking, further enhancing transparency and authenticity in content creation.