By John Cook, George Mason University and Sander van der Linden, University ofย Cambridge
In a world where โpost-truthโ was 2016โs word of the year, many people are starting to doubt the efficacy of facts. Can science make sense of anti-science and post-truthism? More generally, how can we understand what drives peopleโs beliefs, decisions andย behaviors?
Scientists have developed many theories to describe how people process and think about information. Unfortunately, thereโs an increasing tendency to see people as creatures whose reasoning mechanisms are largely dependent on a narrow set of processes. For example, one popular theory suggests that if we just communicate more accurate information to people, their behavior will change accordingly. Another suggests that people will reject evidence if it threatens their deeply held cultural worldviews and associatedย feelings.
Itโs more important than ever that our approach to communication is evidence-based and built on a strong, theoretical foundation. Many of these models contribute valuable insights and can help us design better communication, but each on its own is incomplete. And science communicators have a tendency to oversimplify, focusing on a single model and disregarding otherย theories.
We suggest that this is a dangerous practice and less effective than a more nuanced and holistic view. The apparent choice between โfactโ and โfeeling,โ or between โcognitionโ and โculture,โ is a false dilemma. In reality, both are related and address different pieces of the decision-makingย puzzle.
Thinking Versusย Feeling
One well-known theory about how people absorb new facts is the โinformation deficit model.โ The main idea here is straightforward: If you throw more facts at people, theyโll eventually come around on anย issue.
Most behavioral science scholars agree that this model of human thinking and behavior is clearly incomplete โ people rely on a range of other cues besides facts in guiding their attitudes and behavior. For example, sometimes we simply act based on how we feel about an issue. Unfortunately, the facts donโt alwaysย convince.
But the term โinformation deficitโ is problematic, too. People tend to have limited information in most areas of life. For example, we often donโt know the thoughts and feelings of other people we trust and value. Similarly, we might have limited knowledge about appropriate cultural norms when traveling to a new country, and so on. Information deficit isnโt a very meaningful term to use to theorize about humanย thinking.
Another theory about human thinking is called โcultural cognition.โ In brief, it suggests that our cultural values and worldviews shape how we think about science and society.
Itโs easy to be duped into thinking of the human brain as a sponge that soaks up only the information it wants to believe. For example, the theory suggests that peopleโs position on divisive issues such as climate change is not informed by scientific evidence but rather by the strong commitment people have to their political groups and ideologies. The idea is that to protect our cultural worldviews, we actively reject evidence that threatens them โ think of someone who fears that government action on climate change undermines the freeย market.
In short, this narrative sounds appealing on the surface, as humans organize themselves in groups, and much psychological research has shown that we derive part of our social identities from the group affiliations weย maintain.
Yet, its focus is overly narrow, and thereโs a logical fallacy in this conception of human thinking. We belong to many groups at any given time and we juggle many different public and private identities. The real question is about nuance; when and under what conditions is someone motivated to reject scientific facts in favor of their culturalย worldview?
Either/or Misses theย Point
To throw all our fact-disseminating eggs into one or the other theoretical basket is oversimplistic and deprives us of importantย insights.
A more nuanced perspective recognizes that facts and information are embedded in social and cultural contexts. For example, peopleโs perception of expert consensus matters a great deal, especially on contested issues, and is often described as a gateway belief that influences a range of other attitudes about an issue. The near-unanimous consensus that vaccines do not cause autism or that climate change is human-caused are all scientific facts. At the same time, consensus information is also inherently social: It describes the extent of agreement within an influential group ofย experts.
People often want to be accurate in their views, and, in an uncertain world bounded by limited time and effort, we make strategic bets on what information to take into account. Consensus acts as a natural heuristic, or mental shortcut, for complicated scientific issues. Numerous studies have found that highlighting scientific agreement on human-caused global warming can help neutralize and reduce conflicting views about climateย change.
Similarly, while some studies have found a limited effect of knowledge on judgment, when you dig deeper into the data, a more nuanced and insightful picture emerges. For example, some studies claim that a deficit in scientific โknowledgeโ does not explain why people are divided on contested issues such as climate change. But whatโs being measured in these experiments matters. Indeed, indicators such as how well people understand numbers or their scientific literacy โ which is what some of these studies actually quantify โ are categorically different from measuring specific knowledge people have about a topic such as climate change. In fact, a survey across six countries found that when people understand the causes of climate change, their concern increases accordingly, irrespective of their values. Similarly, other studies show that explanations about the mechanisms of climate change can reduce biased evaluations of evidence as well as politicalย polarization.
In short, facts doย matter.
How People Think Is Complex andย Nuanced
Indeed, there is no need to throw out the baby with the bathwater. Instead, we need to dispel false dichotomies and folk psychology about human thinking that currently dominate the media. Repeating the story that people donโt care about facts runs the risk of becoming a self-fulfilling prophecy. A holistic view acknowledges that people rely on cognitive shortcuts and emotions, care about social norms and group identities and are sometimes motivated in their reasoning, but it also recognizes the research showing that most people want to fundamentally hold accurate perceptions about theย world.
This is particularly important as the public is currently hampered by misinformation and fake news. In two separate studies, we each found that misinformation about climate change has a disproportionate influence on public attitudes and opinion. However, we also found that inoculating people against the false arguments neutralized misinformationโs influence, across the political spectrum. In essence, teaching people what false arguments might be deployed helped them overcome their cultural biases. Other work similarly shows that the politicization of science can be counteracted withย inoculation.
People are complex, social and affected by a diverse range of influences depending on the situation. We want to hold accurate views, but emotion, group identities and conflicting goals can get in the way. Incorporating these different insights into human thinking enriches our understanding of how people form opinions and makeย decisions.
Effective science communication requires an inclusive, holistic approach that integrates different social science perspectives. To simplistically focus on a single perspective paints a limited and increasingly inaccurate view of how humans form judgments about social and scientificย issues.
John Cookย isย Research Assistant Professor at theย Center for Climate Change Communication, George Mason University and Sander van der Lindenย isย Director of the Cambridge Social Decision-Making Lab, University of Cambridge.
This article was originally published on The Conversation. Read the original article.
Main image: An Army Corps of Engineers scientist explains wetland science to students from Savannah State University. Credit: U.S. Army Corps of Engineers, publicย domain
Subscribe to our newsletter
Stay up to date with DeSmog news and alerts