- By Steve Smith
- December 23, 2021
Facebook targeting individual members who involve in spreading false news or misleading content. The punitive actions are taking against Facebook members for sharing false information.
Several measures are introduced to Facebook members including penalization, especially for individuals who found spreading false information. Facebook targeting aims to reduce the amount of toxic content shared on the platform.
Earlier, Facebook targeting emphasized Pages and organizations that were promulgating misinformation. But now Facebook is focusing on educating Facebook members to stop spreading false content and improve their behavior, otherwise, Facebook will take action against them to restrict access to their posts.
“Today, we’re launching new ways to inform people if they’re interacting with content that’s been rated by a fact-checker as well as taking stronger action against individual members who repeatedly share misinformation on Facebook.
Whether it’s false or misleading content about COVID-19 and vaccines, climate change, elections, or other topics, we’re making sure fewer people see misinformation on our apps.”
The attempt to limit misleading content comprises three actions:
To restrain the members from sharing toxic content begins when they like a page that is comprehended for sharing misleading content.
This effort will result in less engagement and a low rate of likes on the page but more, it will help educate the individual to know about the toxic nature of the page.
Besides, Facebook will alert the user about the page by showing a popup warning and suggest him a link that provides more information about their fact-checking program.
The next step would be to penalize the individuals who repeatedly share toxic content that has been rated as false by Facebook’s raters.
The penalty includes reducing the reach of Facebook’s posts so fewer people go through them.
According to Facebook’s announcement:
“Starting today, we will reduce the distribution of all posts in News Feed from an individual’s Facebook account if they repeatedly share content that has been rated by one of our fact-checking partners.”
Last but not least part of this attempt involves educating the members who have shared content that was reported as misleading content.
The current effort is when someone shared content, he gets informed about the status of the content they have shared.
Facebook’s announcement explained it like this:
“We currently notify people when they share content that a fact-checker later rates, and now we’ve redesigned these notifications to make it easier to understand when this happens.
The notification includes the fact-checkers article debunking the claim as well as a prompt to share the article with their followers.
It also includes a notice that people who repeatedly share false information may have their posts moved lower in News Feed so other people are less likely to see them.”
Facebook is committed to being free of spam, graphics, violent content, and adult content. Most of it was captured by the automated system of Facebook.
In this ongoing war against misleading content and false information, the removal of content at the individual level is a new front.