Following the shock results of Brexit and the Trump victory, a lot of attention has focused on the role that Facebook might have played in creating online political ghettos in which false news can easily spread.
Facebook now has serious political influence thanks to its development from a social networking tool into a primary source of news and opinions. And for many, the way it manages this influence needs greater scrutiny. But to put the blame solely on the company is to overlook how people use the site, and how they themselves create a filter bubble effect through their actions.
Much of this debate has focused on the design of Facebook itself. The site’s personalisation algorithm, which is programmed to create a positive user experience, feeds people what they want. This creates what the CEO of viral content site Upworthy, Eli Pariser, calls “filter bubbles”, which supposedly shield users from views they disagree with.
People are increasingly turning to Facebook for their news — 44 % of US adults now report getting news from the site — and fake news is not editorially weeded out. This means that misinformation can spread easily and quickly, hampering the chance people have for making informed decisions.
Over the last few weeks, there have been frequent calls for Facebook to address this issue. President Obama himself has weighed in on the issue, warning of the perils that rampant misinformation can have for the democratic process.
Much of the debate around this, however, has had an element of technological determinism to it, suggesting that users of Facebook are at the mercy of the algorithm. In fact, research shows that the actions of users themselves are still a very important element in the way that Facebook gets used.
Our research has been looking specifically at how people’s actions create the context of the space in which they communicate. Just as important as the algorithm is how people use the site and shape it around their own communications. We’ve found that most users have an overwhelming view that Facebook is not ideally suited to political debate, and that posts and interactions should be kept trivial and light-hearted.
This isn’t to say that people don’t express political opinions on Facebook. But for many people there’s a reluctance to engage in discussion, and a sense that anything that might be contentious is better handled by face-to-face conversation. People report that they fear the online context will lead to misunderstandings because of the way that written communication lacks some of the non-linguistic cues of spoken communication, such as tone of voice and facial gestures.
There’s strong evidence in our research that people are actually exposed to a great deal of diversity through Facebook. This is because their network includes people from all parts of their life, a finding that echoes other research. In this respect, the algorithm doesn’t have a marked influence on the creation of filter bubbles. But because they often want to avoid conflict, people report ignoring or blocking posts, or even unfriending people, when confronted with views with which they strongly disagree.
They also report taking care of what they say themselves so as not to antagonise people such as family members or work colleagues whose views differ from theirs, but whose friendship they wish to maintain. And finally, they talk of making a particular effort to put forward a positive persona on social media, which again stops them from engaging in debate which might lead to argument.
Not so easy to fix
The idea that algorithms are responsible for filter bubbles suggests it should be easy to fix (by getting rid of the algorithms), which makes it an appealing explanation. But this perspective ignores the part played by users themselves, who effectively create their own filter bubbles by withdrawing from political discussions and hiding opinions they disagree with.
This isn’t done with the intention of sifting out diversity but is instead due to a complex mix of factors. These include the perceived purpose of Facebook, how users want to present themselves in an effectively public form, and how responsible they feel for the diverse ties that make up their online network.
The fact that manipulation by the algorithm isn’t the only issue here means that other solutions — for example, raising people’s awareness of the possible consequences that their online actions have can help encourage debate. We have to recognise that the impact of technology comes not just from the innovations themselves but also from how we use them, and that solutions have to come from us as well.
- Philip Seargeant is senior lecturer in applied linguistics, The Open University, and Caroline Tagg is lecturer in applied linguistics and English language, also at The Open University
- This article was originally published on The Conversation