Combating Misinformation Through Nudging

. Recent studies have shown that the spread and consumption of misin-formation online can be attributed to errors in human decision making, facilitated by cognitive biases. The field of Behavioral Economics has contributed a repertoire of such cognitive biases that can be leveraged for the design of technological interventions. In particular, the concept of nudging refers to subtle changes in the ‘choice architecture’ that can alter people's behaviors in predictable ways. In this paper we present our ongoing work on the design of nudging interventions in the context of misinformation, including a systematic review of the use of nudging in HCI that has led to a design framework consisting of 23 mechanisms of nudg-ing tapping to 15 different cognitive biases, the translation of this framework into a set of design cards, the Nudge Deck , and its use in a planned workshop that aims to explore the design space of misinformation in the context of nudging.


Introduction
In the era of "post-truth", the rise of falsehood content dissemination in the information environment has sparked significant public concern, since it has been posed as a threat to the very existence of democratic societies.Meinert et al. (2018) define fake news and false information, in general, as purposeful publications of false, discrediting and deceitful content, motivated by various interests (e.g.political, financial, etc.) so as to manipulate and exploit its recipients.
A number of researchers have called for the examination of the factors that result in the dissemination and consumption of falsehood knowledge.As examined in the following sections, various cognitive and emotional biases facilitate the spread and consumption of misinformative content.For instance, Vosoughi et al. (2018) found that increased speed and depth at which fake news diffuse can be attributed to mere novel effect, while Badke (2018) discussed the role of the confirmation bias on the consumption and spread of misinformation.
Based on the concept of nudging, we suggest that knowledge about the different cognitive biases that we, as humans, are susceptible to, can be leveraged for the design of technological interventions that minimize the spread and consumption of misinformation.This paper presents a preliminary account of our ongoing work towards this goal.

2
The role of human decision making on the spread and consumption of misinformation  2018) found that, contrary to conventional wisdom, the spread of false news could not be attributed to the structure of social media outlets, website platforms and internet bots, but rather to a mere novelty effect.Novelty, as the authors claimed, "attracts human attention, contributes to productive decision-making, and encourages information sharing because novelty updates our understanding of the world".False news was found to be more novel than true news, suggesting that people were more likely to share novel information.Badke (2018) supports that humans see only what they expect or want to see, without inspecting news thoroughly.This, they argue, is a product of the confirmation bias, the internal tendency of people to seek out information that confirms and verifies what they already believe, instead of examining critically all the pieces of information.According to the theory of cognitive dissonance, whenever a presented piece of news includes information which conflicts with the currently held mental models of people, it immediately induces cognitive dissonance [4].People are motivated to scale down this dissonance, thus they may avoid or even discount knowledge that contrasts their personal positions.Weeks (2015) argues that emotional experience moderates the influence of partisanship on individuals' responses to misinformation.Specifically, when individuals experience anger, the influence of partisanship is boosted, making individuals more likely to believe claims that are associated with their political affiliation.On the contrary, anxiety reduces the influence of partisanship and increases the chance of making other political affiliations believable.Schwarz et al. (2016) argue that whenever people come across a new piece of information, they tend to assess its truthfulness by focusing on five criteria.People usually ask themselves about the social consensus of the story, its supporting evidence, its consistency, coherence and credibility.However, instead of evaluating these questions analytically, individuals tend to use mental shortcuts in order to minimize the time and energy spent.This makes them susceptible to errors in decision making.

3
Nudging away from misinformation Thaler and Sunstein (2013) define a nudge as "any aspect of the choice architecture that has the capacity to change people's behavior in a predictable manner but without preventing any other alternatives or altering their economic incentives".Nudges leverage knowledge about systematic biases in decision making to support people in making optimal decisions.For instance, the status-quo bias reflects our tendency to resist change and to go along with the path of least resistance [8].As such, we often chose the default option rather than taking the time to consider the alternatives, even when this is against our best interests.For example, several countries in Europe have changed their laws to make organ donation the default option.In such so called opt-out contexts, over 90% of the citizens donate their organs; while in opt-in contexts the rate falls down to 15%.Similarly, replacing cake with fruit in the impulse basket next to the cash register, has been found to lead people in buying more fruit and less cake, when both choices are still available [9].As evident, nudges work on the assumption that people don't always make rational choices and they attempt to direct human behavior, yet without imposing a particular choice [9].Our ongoing work seeks to explore the power of nudging in reducing the likelihood of spreading as well as consuming misinformation.Through a systematic review of the use of nudging in HCI research, we have identified 23 distinct mechanisms of nudging developed within HCI, clustered in 6 overall categories, and tapping to 15 different cognitive biases and heuristics [8].One such mechanism is called "reminding the consequences".Tapping on the availability heuristic which reflects our tendency to judge the probability of occurrence of an event based on the ease at which it can be recalled, thus making us susceptible to underestimate the probability of events when these are not readily available to our cognitive processing, such nudges prompt individuals to reflect on the consequences of their actions.One example is provided by Harbach et al. (2014) who redesigned the permissions dialogue of the Google Play Store to incorporate personalized scenarios that disclosed potential risks from app permissions.If the app required access to one's storage, the system would randomly select images stored on the phone along with the message "this app can see and delete your photos".
We have translated this framework of the 23 mechanisms of nudging into a set of design cards, the Nudge Deck [11].Each mechanism is described in a card (see figure 1 for an example card), where the front side includes a definition of the nudging mechanism, an image of a possible implementation and an explanation.The back side illustrates suggestions, directions, instructions to provoke motivation and inspiration during the design conception.Moreover, to distinguish the situations in which each nudge should be used, the 23 mechanisms are mapped into the three trigger types, as suggested by Fogg's Behavior Model: sparks (i.e., ones to increase motivation), facilitators (i.e., ones that increase ability) or signals (i.e., ones to remind of the behavior).
Using the Nudge Deck as a design support tool, we currently plan to conduct a workshop with the goal of ideating on nudging interventions to combat the spread, or consumption of misinformative content online.To provide an example, suppose we want to minimize the spread of fake news on Twitter through preventing unaware users from retweeting those fabricated stories.One nudge mechanism that can be exploited is called "throttling mindless activity" [8].This mechanism taps into regret aversion bias, people's tendency to become more reflective and break their mindless activities whenever risk is identified.In this case, given that the user is about to post a tweet that contains bogus news, the tool could notify the user with a message like: "We estimate 90% chance of the article containing falsehood information.Are you sure you want to publish this tweet?".Through instilling doubt, this nudge encourages users to reconsider the tweet's content.A link to a fact-checking article may further provide an opportunity for learning.Our workshop will seek to explore how different nudging mechanisms can be applied in different platformsfrom social media, to e-mail, and question and answer (Q&A) websites like WikiAnswers.We will engage students of interaction design in small design teams and seek to provide a first inquiry into the design space of technology-mediated nudging in the context of misinformation-resilience tools, through an analysis of the design ideas that come out of the workshop and the emerging design qualities of those ideas.

Figure 1 :
Figure 1: The Nudge Decks consists of 23 mechanism cards (top), six category cards (middle), and three trigger cards (bottom