The decision by Meta, the parent company of Facebook and Instagram, to end its fact-checking program and otherwise reduce content moderation raises the question of what content on those social media platforms will look like going forward.
One worrisome possibility is that the change could open the floodgates to more climate misinformation on Meta's apps, including misleading or out-of-context claims during disasters.
Meta's policies have fact-checkers prioritizing "viral false information," hoaxes and "provably false claims that are timely, trending and consequential." Meta explicitly states that this excludes opinion content that does not include false claims.
Fact-checking curbs climate misinformation
I study climate change communication. Fact-checks can help correct political misinformation, including on climate change. People's beliefs, ideology and prior knowledge affect how well fact-checks work. Finding messages that align with the target audience's values, along with using trusted messengers - like climate-friendly conservative groups when speaking to political conservatives - can help. So, too, does appealing to shared social norms, like limiting harm to future generations.
Heat waves, flooding and fire conditions are becoming more common and catastrophic as the world warms. Extreme weather events often lead to a spike in social media attention to climate change. Social media posting peaks during a crisis but drops off quickly.
What distinguishes misinformation from disinformation is the intent of the person or group doing the sharing. Misinformation is false or misleading content shared without active intention to mislead. On the other hand, disinformation is misleading or false information shared with the intent to deceive.
Disinformation campaigns are already happening. In the wake of the 2023 Hawaii wildfires, researchers at Recorded Future, Microsoft, NewsGuard and the University of Maryland independently documented an organized propaganda campaign by Chinese operatives targeting U.S. social media users.
False claims can go viral rapidly
Explaining that scientists agree that climate change is happening and is caused by humans burning greenhouse gases can prepare people to avoid misinformation. Psychology research indicates that this " inoculation " approach works to reduce the influence of false claims to the contrary.
That's why warning people against climate misinformation before it goes viral is crucial for curbing its spread. Doing so is likely to get harder on Meta's apps.
Social media users as sole debunkers
With the coming changes, you will be the fact-checker on Facebook and other Meta apps. The most effective way to pre-bunk against climate misinformation is to lead with accurate information, then warn briefly about the myth - but only state it once. Follow this with explaining why it is inaccurate and repeat the truth.
During climate change-fueled disasters, people are desperate for accurate and reliable information to make lifesaving decisions. Doing so is already challenging enough, like when the Los Angeles County's emergency management office erroneously sent an evacuation alert to 10 million people on Jan. 9, 2025.
Crowd-sourced debunking is no match for organized disinformation campaigns in the midst of information vacuums during a crisis. The conditions for the rapid and unchecked spread of misleading, and outright false, content could get worse with Meta's content moderation policy and algorithmic changes.
The U.S. public by and large wants the industry to moderate false information online. Instead, it seems that big tech companies are leaving fact-checking to their users.

Jill Hopke, Associate Professor of Journalism, DePaul University
This article is republished from The Conversation under a Creative Commons license. Read the original article.