AI Content Scrutiny Intensifies
The digital landscape is facing a significant challenge with the rise of AI-generated deepfakes, a concern amplified during active conflicts. Following
the recent US-Israel war on Iran, a surge of viral videos, many later exposed as fabricated, flooded online platforms, accumulating millions of views. This situation has prompted Meta's Oversight Board to call for a comprehensive revision of the social media giant's content moderation strategies. The board emphasizes the urgent need for Meta to implement robust measures that empower users to discern AI-manipulated media from authentic content. This includes establishing clearer standards for content provenance, ensuring that the origins of AI-modified media are readily understandable. Furthermore, the Oversight Board advocates for substantial investment in advanced AI detection technologies and the development of more effective methods for accurately labeling AI-generated material. The board's directive also pushes for an amendment to existing policies to guarantee a prompt and appropriate response to deceptive AI-generated output, recognizing the potential for significant societal impact, especially during periods of heightened tension and crisis.
Overturned Decision, New Mandates
A pivotal moment in this discussion arose from the Oversight Board's review of an AI-generated video disseminated on Facebook during the June 2025 Israel-Iran conflict. This fabricated video depicted severe damage to buildings with accompanying text falsely claiming "Live now – Haifa Towards Down." Despite Meta's initial decision to leave the post active, the Oversight Board overturned this ruling, asserting that the content warranted a 'High Risk AI' label due to its considerable potential to mislead the public on critical matters during a sensitive period. This ruling underscores the board's concern that the escalating quantity and sophistication of AI-generated content pose profound risks to individuals and societies. The danger is particularly acute when deepfakes designed for deception, manipulation, or increased engagement are circulated during conflicts and crises, such as those experienced in Iran and Venezuela in 2026, and rapidly spread across various digital platforms. The board, however, also stressed that the proliferation of misleading AI content should not lead to the curtailment of free expression, advocating instead for industry-wide coherence in aiding users to differentiate deceptive AI creations.
Rethinking Content Labeling
Meta currently employs an 'AI Info' label for content altered or created using artificial intelligence. However, the Oversight Board has sharply criticized these existing labeling mechanisms, deeming them "neither robust nor comprehensive enough" to contend with the sheer volume and speed at which AI-generated content proliferates, especially during volatile periods like conflicts. The board also highlighted Meta's inconsistent application of Coalition for Content Provenance and Authenticity (C2PA) standards, even for content generated by its own AI tools. In line with its recommendations, the Oversight Board has echoed certain aspects of India's recent rules for AI-generated content. Specifically, it suggests Meta should incorporate provenance information and invisible watermarks into content produced by its AI tools. This aligns with India's IT Rules amendments for synthetically generated information (SGI), effective February 20, 2026, which mandate that AI platforms prominently label AI-generated content with a permanent, tamper-proof unique metadata or identifier. The Oversight Board further recommends Meta introduce distinct 'High Risk' and 'High Risk AI' labels, alongside more efficient escalation channels and automated, large-scale labeling processes.
Enhanced Detection and Standards
The Oversight Board has further recommended that Meta invest in more sophisticated tools capable of identifying AI-generated content across multiple formats, including audio, video, and images. Meta has also been urged to publish a clear schedule of penalties for users who fail to self-disclose the use of digitally created or altered content. Moreover, the board proposes the creation of a dedicated Community Standard for AI-generated content, encompassing comprehensive rules on preserving provenance, AI labeling protocols, and self-disclosure requirements. It also suggests modifying the existing Misinformation Community Standard to ensure that reviews of misinformation posing an immediate risk of violence or physical harm are not solely reliant on external partner signals. While the Oversight Board's content moderation decisions for Meta's platforms are binding, the company is not obligated to adopt its recommendations in the same manner. Nevertheless, Meta is required to publicly respond to each recommendation within 30 days, as per its operational guidelines, signaling a commitment to addressing these critical issues.













