Democracy

The state of content moderation for the LGBTIQA+ community and the role of the EU Digital Services Act

E-paper
Social media platforms play a crucial role in supporting freedom of expression in today's digital societies. Platforms can empower groups that have previously been silenced. However, platforms also host hateful and illegal content, often targeted at minorities, and content is prone to being unfairly censored by algorithmically biased moderation systems. This report analyzes the current environment of content moderation, particularly bringing to light negative effects for the LGBTIQA+ community, and provides policy recommendations for the forthcoming negotiations on the EU Digital Services Act.

Algorithmic misogynoir in content moderation practice

E-paper
Existing content moderation practices, both algorithmically-driven and people-determined, are rooted in white colonialist culture. Black women’s opinions, experiences, and expertise are suppressed and their online communication streams are removed abruptly, silently, and quickly. Studying content moderation online has unearthed layers of algorithmic misogynoir, or racist misogyny directed against Black women. Tech companies, legislators and regulators in the U.S. have long ignored the continual mistreatment, misuse, and abuse of Black women online. This paper explores algorithmic misogynoir in content moderation and makes the case for the regular examination of the impact of content moderation tactics on Black women and other minoritized communities.