Urgent Call for Policy Revision
In the midst of intense global conflicts, where the rapid spread of misinformation can have severe consequences, Meta's Oversight Board has issued a compelling
directive. This independent body is urging the social media behemoth to fundamentally rethink its approach to managing AI-generated content, often referred to as deepfakes. The board’s recommendations stem from a recent ruling that highlighted the inadequacy of current systems in identifying and flagging deceptive synthetic media, particularly when it surfaces during critical events like armed conflicts. This push for revision comes at a crucial time, as numerous viral videos, initially believed to be authentic war footage and garnering millions of views, have later been exposed as fabricated by artificial intelligence, underscoring the urgent need for more robust safeguards on platforms like Facebook and Instagram. The board's intervention is a direct response to the growing threat of AI-driven manipulation impacting public discourse and safety.
Enhancing User Awareness
A cornerstone of the Oversight Board's recommendations is the imperative for Meta to empower users with better tools to discern AI-generated content. This involves not just implementing clear labelling but also providing more transparency about the origins and manipulation of media. The board suggests a more consistent application of content provenance standards, which would involve attaching verifiable information about where and how digital content was created or altered. The goal is to foster a more informed user base, capable of critically assessing the information they encounter. By making it easier for individuals to identify synthetically created or modified material, Meta can mitigate the risk of widespread deception, especially when such content is weaponized to influence opinions or incite panic during sensitive periods like international conflicts or humanitarian crises. This proactive approach aims to build resilience against the growing sophistication of AI manipulation.
Investing in Detection and Labeling
Beyond user education, the Oversight Board has strongly advised Meta to bolster its technological defenses against AI-generated misinformation. This includes a significant investment in developing and deploying more advanced AI detection tools, capable of identifying sophisticated deepfakes across various formats, including images, audio, and video. Furthermore, the board emphasizes the necessity of evolving content labelling mechanisms. While Meta currently uses an 'AI Info' label, the board found it insufficient, recommending a more robust and comprehensive system. This could involve introducing specific labels like 'High Risk AI' for content posing a significant threat of misleading the public, especially on sensitive topics during critical times. The board also calls for improved methods for timely and adequate responses to deceptive AI-generated outputs, ensuring that such content is addressed swiftly and effectively before it can cause widespread harm or influence public perception during conflicts.
Addressing AI in Conflicts
The Oversight Board’s recommendations are particularly sharp-eyed regarding the heightened risks posed by AI-generated content during armed conflicts. The recent US-Israel war on Iran, for example, saw a surge of viral videos, many later proven to be AI-generated fabrications depicting exaggerated damage or false narratives. The board's ruling on a specific AI-generated video falsely showing extensive damage in Haifa during the June 2025 Israel-Iran war underscores this concern. By overturning Meta's decision to leave the post unflagged, the board asserted that such content should have received a 'High Risk AI' label due to its potential to mislead the public during a critical period. The board cautions that while combating deception is vital, it must not come at the expense of freedom of expression, advocating for a balanced approach that fosters industry-wide coherence in identifying manipulative AI content while addressing malicious actors.
Provenance and International Alignment
A key recommendation from the Oversight Board involves Meta's adoption of content provenance standards, akin to those being implemented in India's newly notified rules for AI-generated content. This includes embedding provenance information and invisible watermarks into content generated by Meta's own AI tools, ensuring a traceable origin. India's IT Rules amendments require AI platforms to clearly label or embed permanent, tamper-proof metadata for synthetically generated information (SGI). The Oversight Board's proposal to attach 'High Risk' and 'High Risk AI' labels, along with clearer escalation channels and automated at-scale labelling, echoes this global trend towards transparency. Some board members even suggested that posts with 'High Risk AI' labels should be demoted or removed to prevent the dissemination of deceptive AI content, mirroring the growing urgency to regulate this rapidly evolving technology.














