Another Way Facebook and Google May Be Undermining Democracy

Customized news feeds reduce our exposure to divergent viewpoints, according to new research.
Author:
Publish date:
(Photo: Jeff J Mitchell/Getty Images)

(Photo: Jeff J Mitchell/Getty Images)

Google and Facebook have, in recent months, belatedly began to engage in the battle against fake news. But the fact so much misinformation has proliferated on their platforms is only one of the ways these technology giants may be endangering democracy.

Newly published research points to another: It finds the tools these companies offer to customize our news feeds result in users getting less and less exposure to viewpoints that challenge their own.

“Originally conceived by computer and information scientists as a way to help users cope with increasing information overload, customizability technology appears to have a dark side,” writes a research team led by Ivan Dylko of the University at Buffalo. “It enables individuals to surround themselves with information supporting their preexisting political attitudes.”

The researchers report this effect was strongest for “ideologically moderate individuals,” potentially pulling them in a polarized direction. Such a dynamic can undermine important foundations of deliberative democracy,” they write in the journal Computers in Human Behavior.

The study — one of the first to address this issue — featured 93 students from a university in the southwestern United States. All began by filling out a questionnaire measuring their political attitudes and ideology. Four weeks later, they were asked to provide feedback on “a new political news website.”

The students were randomly assigned to try one of several versions of the site, including one in which they could select the ideology of their information sources, and another in which the computer software performed similar sorting. The computer’s choices of what to include and exclude were based on the information participants provided in their questionnaire; they were not informed that this automated customization had taken place.

The researchers measured how often participants clicked on articles that supported or opposed the students’ political preferences, and how much time they spent reading each.

Not surprisingly, use of either customized technology decreased the number of clicks, and time spent reading, articles espousing viewpoints that differed from those of the user. More insidiously, the automated form of customization produced a stronger such effect than the one where the user consciously chose what sorts of articles he or she wanted to read.

Why would that be? The researchers note that “actively and intentionally avoiding counter-attitudinal political information” diminishes our ability to see ourselves as fair-minded. Holding onto a positive self-image — which is a priority for virtually everyone — may inspire us to at least occasionally check out what the other side is saying.

But when the filtering process is completed automatically, and we’re left with the illusion that the whole world agrees with us, that issue goes away. We can enjoy the comfort of never being challenged without the psychological cost.

Perhaps surprisingly, the researchers found filtering has a greater effect on political moderates than those at either extreme. They suspect this reflects the fact that strong partisans are more knowledgeable about the issues, and thus better able to refute opposing arguments.

Therefore, the researchers argue, ideological filters are less important to them. But those same tools could be a godsend to less-informed people who don’t want the bother of learning about both sides, and then making an informed decision. They essentially enable them to just go with their guts.

Given previous research showing that selective exposure to information leads to polarization, it’s clear why Dylko and his colleagues consider these filters dangerous. But what can be done? The researchers have several suggestions.

One obvious one would be eliminating customization based on ideology. If companies insist on leaving it in place, they could require users who choose to filter content to do so every time they check the news. For the psychological reasons we noted earlier, that would very likely discourage the practice.

Technology giants, the ball — and, perhaps, the future of our democracy — is in your court.

Tags
terms:
Primer

Related