Can crowdsourced fact-checking curb misinformation on social media?

While society’s observations have the ability to be very effective, the difficult function of moderate content benefits from a mixture of different methods. As a professor of natural language processing at MBZUAI, I spent most of my profession searching for misleading, advertising and fake news information online. Therefore, it was one of the first questions I asked for myself: Will it replace human facts pesticides with the notes of the collective community with negative effects on users?
Jams wisdom
Community notes began on Twitter like Birdwatch. It is a credible feature where users who participate in the program can add a context and clarify what they consider wrong or misleading tweets. Notes are hidden until the community evaluation reaches consensus – which makes people with different views and political opinions agree that the position is misleading. The algorithm is determined when the threshold of consensus is reached, then the memo becomes publicly visible under a tweet in the question, which provides an additional context to help users issue enlightened provisions about its content.
Society’s notes seem to be working well. A team of researchers from the University of Illinois Urbana Chambine and the University of Rochester found that the X community notes program can reduce the spread of wrong information, which leads to revenge by the authors. Facebook is largely dependent on the same approach used in X today.
After studying and writing about moderate content for years, it is wonderful to see another major social media company that implements the collective outsourcing of moderate content. If it works for Meta, this may be a real change for games for more than 3 billion people who use the company’s products every day.
However, moderate content is a complex problem. There is no single silver bullet that will work in all cases. The challenge can only be faced by employing a variety of tools that include human facts, collective outsouruing and algorithm. Each of these is the most appropriate for different types of content, and it must work and should work at a concert.
Random mail safety and llm
There are precedents to address similar problems. Decades ago, the email of the RAM was a much greater problem than it is today. In a large part of it, we defeated the random mail through the mass or -being. Email service providers provided reports setting features, where users can determine suspicious emails. The widespread the random mail message is distributed, the more it is discovered, and the more people are informed.
Another useful comparison is how LLMS models approach harmful content. For the most dangerous infines – associated with weapons or violence, for example – Many LLMS simply rejects the answer. At other times, these systems may add responsibility to their outputs, such as when they are asked to provide medical, legal or financial advice. It is this gradual approach that my colleagues and I have explored MBZUAI in a recent study where we suggest a hierarchical sequence of methods that can respond to LLMS to different types of potential harmful information. Likewise, social media platforms can benefit from various methods of moderate content.
Automatic filters can be used to determine the most dangerous information, and to prevent users from seeing and sharing it. These automatic systems are fast, but can only be used for certain types of content because they are unable to the differences required for most of the content.
Don’t miss more hot News like this! Click here to discover the latest in AI news!
2025-05-19 10:20:00