Meta on Tuesday announced a set of changes to its content moderation practices that would effectively put an end to its longstanding fact-checking program, a policy instituted to curtail the spread of misinformation across its social media apps.
The Shift in Policy
The reversal of the years-old policy is a stark sign of how the company is repositioning itself for the Trump era. Meta described the changes with the language of a mea culpa, saying that the company had strayed too far from its values over the prior decade.
Statement from Meta
“We want to undo the mission creep that has made our rules too restrictive and too prone to over-enforcement,” Joel Kaplan, Meta’s newly installed global policy chief, said in a statement.
Instead of using news organizations and other third-party groups, Meta, which owns Facebook, Instagram and Threads, will rely on users to add notes or corrections to posts that may contain false or misleading information.
This shift in policy highlights Meta's efforts to empower its users in the fight against misinformation, while also acknowledging past mistakes and promising a more open and transparent approach to content moderation.