Nearly half of social media users say they have encountered blocked or labeled content on platforms
A recent study found the majority of social media users trust an app less if they encounter censorship, and nearly a third of people who have encountered the blocked or labeled content have reduced the amount of time they spend on the platform.
“Using true/false as a test for blocking or labeling content inevitably leads to perception of bias and breaks the trust of users. The survey shows that this leads to reduced time spent and engagement by the 49% of users who have experienced content being blocked or labeled,” Mike Matthys, founder of The Institute for a Better Internet, told Fox News Digital.
The Institute for a Better Internet, a non-partisan organization that advocates for free speech, free press, and due process, conducted the national research between Sept. 27 and Oct. 4 by phone and online among 1,100 people who have at least one social media account.
The research found that 49% of people with social media accounts have seen blocked or labeled content, with 53% saying such censorship causes them to trust the app less. For Facebook users specifically, 58% of those surveyed said they trust the platform less after encountering blocked or labeled content.
Nearly a third of those polled said that such censorship makes them spend less time on a platform or less likely to share content on the site. For Facebook users, 27% said they spend less time on the site if they encounter blocked or labeled content, while 30% of people nationwide report the same. The majority of users for platforms – including Instagram, Facebook, YouTube, TikTok and others – report that encountering censorship makes no difference on how much time they spend on the sites.
“If content moderation policies are causing a material portion of users to reduce engagement online, this may explain a portion of why some of the social media companies are seeing their growth slow or even reverse,” Matthys said.
The study also found that social media users would accept content moderation protocols that included tech giants using a “harmful/not harmful” test rather than completely relying on a “true/false” test for moderation. The majority of those polled – at six out of 10 users – said they would accept content moderation rules if it included the “harmful/not harmful” test.
Social media platforms have repeatedly been slammed for censoring content in recent years. Twitter and Facebook, for instance, banned former President Donald Trump from their platforms in 2021, while Twitter blocked the New York Post’s story in 2020 on Hunter Biden’s notorious laptop, and Facebook was slammed for cracking down on alleged COVID misinformation during the pandemic.
The study found that Republicans and conservatives are not the only ones to trust a platform less for censorship. Americans who consider themselves independents and moderates also reported trusting platforms less if they encounter blocked or labeled content. Fifty-one percent of Independents said they trust a platform less after seeing such labels, and 48% of moderates reported the same.
Among left-leaning social media users, 37% of Democrats say they trust a platform less after encountering blocked and labeled content, while 43% of liberals reported the same.
“Approximately 12% of all Facebook users reported that they have reduced time and engagement online because of their experience with content moderation that reduced their trust in the platform,” Matthys noted.
The Institute for a Better Internet crafted a proposal to create a third-party content moderation entity, which “is designed to help the social media platforms rebuild trust with their users,” according to Matthys.
The proposal would “establish online video-calls, an online video courtroom, with a live arbitrator to resolve online content disputes quickly and transparently,” according to the study, as opposed to online written messages between a social media user and a platform.
Researchers detailed the proposal to those surveyed and found 66% of respondents supported such a plan. The study found 71% of Democrats and 65% of Republicans either somewhat or strongly supported the proposal.
“While other studies show that users believe the monopoly social media platforms have excessive unchecked power to moderate content as they like, this study shows that additionally such content moderation, which is personally experienced by 49% of all social media users, is actually harming the business of these social media companies by reducing users’ trust in the platform and reducing the amount of time spent and information shared on the social media platforms by these users,” the study concluded.
To read the full article click here.