Over 140 former content moderators in Kenya are suing Meta, the parent company of Facebook, after being diagnosed with severe post-traumatic stress disorder (PTSD) due to their exposure to graphic and disturbing content while working for the company. The moderators, employed by Meta’s outsourcing partner Samasource from 2019 to 2023, reported experiencing meaningful mental health issues, including anxiety and depression, linked to their daily tasks of reviewing violent and traumatic material such as murders and child abuse. This lawsuit highlights the urgent need for better mental health support and working conditions for content moderators globally, as advocates call attention to the psychological toll of such roles in the tech industry [1[1[1[1][2[2[2[2][3[3[3[3].
Q&A with Mental Health Expert on the Lawsuit Against Meta by Kenyan Content Moderators
Time.news Editor (TNE): Thank you for joining us today to discuss the alarming situation facing former content moderators in Kenya. Over 140 of these employees are suing Meta, the parent company of facebook, after being diagnosed with severe post-traumatic stress disorder (PTSD). Can you provide some context on what led to this lawsuit?
expert (E): Absolutely, and thank you for having me. The lawsuit stems from the traumatic experiences that these moderators faced while working between 2019 and 2023. They were employed by Meta’s outsourcing partner, Samasource, and tasked with reviewing incredibly graphic content, including instances of violence, child abuse, and other deeply disturbing materials. The exposure to such content daily has understandably taken a critically important toll on their mental health, leading to diagnoses of PTSD, anxiety, and depression [1[1[1[1].
TNE: The prevalence of PTSD in this group is staggering. How does the psychological impact of content moderation compare to other high-stress professions?
E: The mental health implications are indeed severe, and content moderation is unique in its challenges.unlike many high-stress jobs that might involve physical danger or emotional labour in a supportive environment, content moderators confront raw, often horrific content without adequate psychological support. This can lead to cumulative trauma, which is especially troubling because many moderators were not sufficiently prepared for the psychological burden of their work. Industries like law enforcement or emergency services provide debriefing and mental health support, but that’s often lacking in tech roles such as content moderation, which is crucial for overall well-being [2[2[2[2].
TNE: This lawsuit sheds light on a significant oversight in the tech industry’s treatment of its workers. What implications does this have for companies like Meta and the broader tech sector?
E: The implications are profound. First, it raises questions about corporate responsibility for the mental health of employees, especially those exposed to harmful content. Companies must take immediate action to improve working conditions and offer better mental health resources,including counseling and regular psychological evaluations for content moderators. Moreover, it could catalyze regulatory changes, prompting lawmakers to implement stricter guidelines for mental health support in the tech industry. The public outcry around this issue may also lead to increased scrutiny of how major platforms manage their content moderation teams [3[3[3[3].
TNE: As advocates for mental health push for change, what practical steps can companies take to protect the psychological well-being of their content moderators?
E: There are several practical steps companies can implement. First, they should develop comprehensive training programs that prepare moderators for the types of content they will encounter, including providing resources on coping mechanisms. Second, establishing regular mental health check-ins and having access to trained mental health professionals can make a significant difference. Additionally, offering opportunities for moderators to take breaks, rotate duties, and engage in wellness programs can help mitigate the ongoing effects of exposure to graphic content. Ultimately, fostering a culture that prioritizes mental health will benefit not only the moderators but the entire association [1[1[1[1].
TNE: Thank you for providing such valuable insights. As this situation continues to develop, it’s essential for both the public and the industry to recognize the human cost of content moderation.
E: Thank you for discussing this critical issue. It’s vital to continue advocating for the rights and well-being of those who perform this challenging role, ensuring they are protected and supported effectively within the workplace.