New Cases Involve Posts Shared in Support of the UK Riots
December 3, 2024
Today, the Board is announcing new cases for consideration. As part of this, we invite people and organizations to submit public comments by using the button below.
Case Selection
As we cannot hear every appeal, the Board prioritizes cases that have the potential to affect lots of users around the world, are of critical importance to public discourse or raise important questions about Meta’s policies.
The cases that we are announcing today are:
Posts Supporting UK Riots
2025-009-FB-UA, 2025-010-FB-UA, 2025-011-FB-UA
User appeals to remove content
Submit a public comment using the button below
The Oversight Board will address the three cases below together, choosing either to uphold or overturn Meta’s decisions on a case-by-case basis.
Three Facebook users shared content during the UK riots that took place between July 30 and August 7, 2024. The riots were sparked by a knife attack that occurred on July 29 in Southport, England, in which three girls were killed and eight others injured at a Taylor Swift-themed dance workshop. In the aftermath of the attack, misinformation about the attacker’s identity, wrongly suggesting he was a Muslim asylum seeker, spread rapidly on social media. The ensuing violence and disorder involved thousands of people, including anti-Muslim and anti-immigration groups; the damage of property such as refugee centers and hotels housing immigrants; and injuries sustained by police officers.
The first post expressed agreement with the riots, calling for more mosques to be smashed and buildings to be set on fire where “scum are living,” also referring to “migrants, terrorist.” The person posting the content acknowledges the riots have damaged private property and injured police officers, but argues that without this violence, the authorities will not listen and put a stop to “all the scum coming into Britian.” Finally, the post reminded readers of the murders, stating the three girls will not be the last victims if the public does not do something. This post was viewed more than 1,000 times and had fewer than 50 comments.
The second piece of content is a reshare of another post. It contains what looks like an AI-generated image of a giant man wearing a union jack (the UK’s flag) T-shirt who is chasing several Muslim men. The image has text overlay providing a time and place to gather for one of the protests and includes the hashtag “EnoughIsEnough.” This content has had fewer than 1,000 views.
The third post is a repost of another likely AI-generated image of four Muslim men wearing white kurtas (tunics), running in front of the Houses of Parliament after a crying blond-haired toddler in a union jack T-shirt. One of the men waves a knife while, above, a plane flies towards Big Ben. The image is accompanied by the caption, “wake up.” This piece of content also has had over 1,000 views and fewer than 50 comments.
All three were reported by other Facebook users for violating either the Hate Speech or Violence and Incitement policies. After assessments by Meta’s automated tools, all three posts were kept up on Facebook. The users who reported the content appealed against the posts remaining up but Meta’s automated systems confirmed the initial decisions. None of the posts were assessed by humans. The same users then appealed to the Board, stating that the content is either inciting violence against migrants, promoting the narrative that immigrants are to blame for the Southport murders or encouraging people to attend the riots.
The Board selected these cases to examine Meta’s policy preparedness and crisis response to violent riots targeting migrant and Muslim communities. This case falls within the Board’s strategic priority of Hate Speech Against Marginalized Groups.
As a result of the Board selecting these cases, Meta determined that its previous decision to leave the first post on Facebook was an error. The company removed the post under its Violence and Incitement policy. Meta confirmed its decisions to leave the second and third post on Facebook were correct. According to the company, for the second post (giant man), there was no violation of its Violence and Incitement policy because the image did not constitute calls for violence against a target. According to Meta, the third post (four men running after a toddler) did not violate its Hate Speech policy because Meta interpreted the image as referring to a specific Muslim man or men – and not all Muslim people. Meta determined the image should be interpreted as a reference to a specific crime – rather than dehumanizing speech targeting Muslim people or Muslim immigrants, given what Meta described as a “false rumor” widely spread, claiming that the perpetrator of the Southport attack was a Muslim or immigrant. The company added that its approach to enforcement in a protest context is to favor maximum protection for voice.
The Board would appreciate public comments that address:
- The role social media played in the 2024 UK riots, including the spread of misinformation, in organizing riots and informing the public of relevant developments.
- Any documented links between anti-immigrant and anti-Muslim online speech and violence and discrimination in the UK.
- The role that imagery (i.e., pictures, graphics, memes, videos, including AI-generated) plays in online hate speech.
- Challenges for automation in assessing incitement or hate speech in imagery, especially AI-generated.
- Risks of overenforcement of Community Standards in such contexts and the potential negative impacts on the free expression rights of speakers and listeners.
- Whether Meta should consider content moderation measures short of removal for this type of content.
As part of its decisions, the Board can issue policy recommendations to Meta. While recommendations are not binding, Meta must respond to them within 60 days. As such, the Board welcomes public comments proposing recommendations that are relevant to these cases.
Public Comments
If you or your organization feel you can contribute valuable perspectives that can help with reaching a decision on the cases announced today, you can submit your contributions using the button below. Please note that public comments can be provided anonymously. The public comment window is open for 14 days, closing at 23.59 Pacific Standard Time (PST) on Tuesday 17 December.
What’s Next
Over the next few weeks, Board Members will be deliberating these cases. Once they have reached their decision, we will post it on the Decisions page.