Home > Media News >
With more people relying on Facebook, the social media network has also become a key source of debate and discussion to spread misinformation and framing of events in order to further political agendas. Part of this comes via blatant misinformation being circulated - both on Facebook and online more broadly. But another element that Facebook has found to be problematic is that some users will re-share older news articles or posts, and frame them as new information.
That can confuse debate over current issues - and now, Facebook is looking to address this by adding in a new warning screen when people go to share a link that will alert the user if the content they're looking to share is more than 90 days old. The warning screen will alert users to the age of the article they're looking to share, but it won't stop them from sharing it if they choose.
Facebook stated that "Over the past several months, our internal research found that the timeliness of an article is an important piece of context that helps people decide what to read, trust and share. News publishers in particular have expressed concerns about older stories being shared on social media as current news, which can misconstrue the state of current events."
Several publications have sought to address this issue by themselves in last April, The Guardian began adding a new date listing on older posts when people shared them on social media. The Guardian noted that it regularly saw spikes in re-sharing of one of its older articles on horse meat regulations in the UK, which it originally published in 2013. The old article had been repeatedly used to re-ignite debate around the issue, with sharers outraged that such a proposal is being considered - again.
This is not the first time that Facebook has added warning prompts to get people to re-think their sharing behavior on the platform. Back in 2016, Facebook added similar pop-ups on posts which had been disputed by third-party fact checkers, prompting users to re-think their intention before they hit 'Share'. Research has found those warnings to be effective in slowing the spread of misinformation, while earlier this year, Facebook also started showing users who'd shared reports about COVID-19, that were later found to be untrue.
Source- The Verge