Social Networks2 views

Major Platforms to Undergo Independent Mental Health Safety Audit

Five of the world’s largest digital platforms—Instagram, TikTok, Roblox, YouTube, and Discord—have agreed to participate in an independent evaluation regarding their impact on user mental health.

The initiative, led by Safer Online Standards, requires these companies to submit detailed data on their internal policies, safety tools, and product features. The primary focus of the audit is the protection and well-being of young users aged 13 to 19.

Key Objectives of the Audit:

  • Transparency: Platforms must disclose how their algorithms and features affect teenagers.
  • Accountability: Evaluation of existing tools designed to mitigate online risks.
  • Safety Ratings: Upon completion of the study, each platform will receive a specific safety classification based on its performance.

This move follows increasing global pressure on tech giants to address the psychological effects of social media and gaming on the youth.