Echo chamber (media)


In news media, an echo chamber is a metaphorical description of a situation in which beliefs are amplified or reinforced by communication and repetition inside a closed system and insulates them from rebuttal. By visiting an "echo chamber", people are able to seek out information that reinforces their existing views, potentially as an unconscious exercise of confirmation bias. This may increase social and political polarization and extremism. The term is a metaphor based on the acoustic echo chamber, where sounds reverberate in a hollow enclosure. Another emerging term for this echoing and homogenizing effect on the Internet within social communities, such as Facebook, Instagram, Twitter, Reddit, etc; is cultural tribalism. Many scholars note the effects that echo chambers can potentially have on citizen's stance and viewpoints, more specifically what implications this effect will have on politics. However, there are counterarguments with supporting evidence that suggest the effects of echo chambers are significantly lowered than assumed and suggested.

Concept

The Internet has expanded the variety and amount of accessible political information. On the positive side, this may create a more pluralistic form of public debate; on the negative side, greater access to information may lead to selective exposure to ideologically supportive channels. In an extreme "echo chamber", one purveyor of information will make a claim, which many like-minded people then repeat, overhear, and repeat again until most people assume that some extreme variation of the story is true.
The echo chamber effect occurs online due to a harmonious group of people amalgamating and developing tunnel vision. Participants in online discussions may find their opinions constantly echoed back to them, which reinforces their individual belief systems due to the declining exposure to other's opinions. Their individual belief systems are what culminate into a confirmation bias regarding a variety of subjects. When an individual wants something to be true, they often will only gather information that supports their existing beliefs and disregard any statements they find that are contradictory or speak negatively upon their beliefs. Individuals who participate in echo chambers often do so because they feel more confident that their opinions will be more readily accepted by others in the echo chamber. This happens because the Internet has provided access to a wide range of readily available information. People are receiving their news online more rapidly through less traditional sources, such as Facebook, Google, and Twitter. These and many other social platforms and online media outlets have established personalized algorithms intended to cater specific information to individuals’ online feeds. This method of curating content has replaced the function of the traditional news editor. The mediated spread of information through online networks causes a risk of an algorithmic filter bubble, leading to concern regarding how the effects of echo chambers on the internet promote the division of online interaction.
It is important to note that members of an echo chamber are not fully responsible for their convictions. Once part of an echo chamber, an individual might adhere to seemingly acceptable epistemic practices and still be further misled. Many individuals may be stuck in echo chambers due to factors existing outside of their control, such as being raised in one.
Furthermore, the function of an echo chamber does not entail eroding a member's interest in truth; it focuses upon manipulating their credibility levels so that fundamentally different establishments and institutions will be considered proper sources of authority.

Echo Chambers vs Epistemic Bubbles

In recent years, closed epistemic networks have increasingly been held responsible for the era of post-truth and fake news. However, the media frequently conflates two distinct concepts of social epistemology: echo chambers and epistemic bubbles.
An epistemic bubble is an informational network in which important sources have been excluded by omission, perhaps unintentionally. It is an impaired epistemic framework which lacks strong connectivity. Members within epistemic bubbles are unaware of significant information and reasoning.
On the other hand, an echo chamber is an epistemic construct in which voices are actively excluded and discredited. It does not suffer from a lack in connectivity; rather it depends on a manipulation of trust by methodically discrediting all outside sources. According to research conducted by the University of Pennsylvania, members of echo chambers become dependent on the sources within the chamber and highly resistant to any external sources.
An important distinction exists in the strength of the respective epistemic structures. Epistemic bubbles are not particularly robust. Relevant information has merely been left out, not discredited. One can ‘pop’ an epistemic bubble by exposing a member to the information and sources that they have been missing.
Echo chambers, however, are incredibly strong. By creating pre-emptive distrust between members and non-members, insiders will be insulated from the validity of counter-evidence and will continue to reinforce the chamber in the form of a closed loop. Outside voices are heard, but dismissed.
As such, the two concepts are fundamentally distinct and cannot be utilized interchangeably. However, one must note that this distinction is conceptual in nature, and an epistemic community can exercise multiple methods of exclusion to varying extents.

Similar concepts

A filter bubble – a term coined by internet activist Eli Pariser – is a state of intellectual isolation that allegedly can result from personalized searches when a website algorithm selectively guesses what information a user would like to see based on information about the user, such as location, past click-behavior and search history. As a result, users become separated from information that disagrees with their viewpoints, effectively isolating them in their own cultural or ideological bubbles. The choices made by these algorithms are not transparent.
Homophily is the tendency of individuals to associate and bond with similar others, as in the proverb "birds of a feather flock together". The presence of homophily has been detected in a vast array of network studies.
Both echo chambers and filter bubbles relate to the ways individuals are exposed to content devoid of clashing opinions, and colloquially might be used interchangeably. However, echo chamber refers to the overall phenomenon by which individuals are exposed only to information from like-minded individuals, while filter bubbles are a result of algorithms that choose content based on previous online behavior, as with search histories or online shopping activity. It is equally important to understand that although similar, Homophily and echo chambers are not the same either.

Implications of Echo Chambers

Online Communities

Online social communities become fragmented by echo chambers when like-minded people group together and members hear arguments in one specific direction with no counter argument addressed. In certain online platforms, such as Twitter, echo chambers are more likely to be found when the topic is more political in nature compared to topics that are seen as more neutral. Social networking communities are communities that are considered to be some of the most powerful reinforcements of rumors due to the trust in the evidence supplied by their own social group and peers, over the information circulating the news. In addition to this, the reduction of fear that users can enjoy through projecting their views on the internet versus face-to-face allows for further engagement in agreement with their peers.
This can create significant barriers to critical discourse within an online medium. Social discussion and sharing can potentially suffer when people have a narrow information base and don't reach outside their network. Essentially, the filter bubble can distort our very own realities that we thought could not be altered by outside sources. The Farnam Street academic blog explains that the filter bubble can have a bigger impact on us than we think. It can create echo chambers that leads us to believe that what you are seeing through ads is the only opinion or perspective that is right.

Offline Communities

Many offline communities are also segregated by political beliefs and cultural views. The echo chamber effect may prevent individuals from noticing changes in language and culture involving groups other than their own. Online echo chambers can sometimes influence an individual's willingness to participate in similar discussions offline. A 2016 study found that “Twitter users who felt their audience on Twitter agreed with their opinion were more willing to speak out on that issue in the workplace”.

Examples

Ideological echo chambers have existed in many forms, for centuries. Some examples of this include:
However, since the creation of the internet, scholars have been curious to see the changes in political communication. Due to the new changes in information technology and how it's managed, understanding opposing perspectives and reaching a common ground in a democracy has been up for debate. The effects seen from the echo chamber effect has largely been cited to occur in politics as the effects of this fragmenting media exposure method as described in these examples:

From Media Companies

Some companies have also made efforts in combating the effects of an echo chamber on an algorithmic approach. A high-profile example of this is the changes Facebook made to its “Trending” page, which is an on-site news source for its users. Facebook modified their “Trending” page by transitioning from displaying a single news source to multiple news sources for a topic or event. The intended purpose of this was to expand the breadth of news sources for any given headline, and therefore expose readers to a variety of viewpoints. There are startups building apps with the mission of encouraging users to open their echo chambers. UnFound.news offers an AI curated news app to readers presenting them news from diverse and distinct perspectives, helping them form rational and informed opinion rather than succumbing to their own biases. It also nudges the readers to read different perspectives if their reading pattern is biased towards one side/ideology. Another example is a beta feature on BuzzFeed News, called “Outside Your Bubble". This experiment adds a module at the bottom of BuzzFeed News articles to show reactions from various platforms, like Twitter, Facebook, and Reddit. This concept aims to bring transparency and prevent biased conversations diversifying the viewpoints their readers are exposed to.