1,815
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Hiding opinions by minimizing disclosed information: an obfuscation-based opinion dynamics model

ORCID Icon, ORCID Icon & ORCID Icon
Pages 315-341 | Received 28 Sep 2020, Accepted 11 May 2021, Published online: 17 Aug 2021

ABSTRACT

In the field of opinion dynamics, the hiding of opinions is routinely modeled as staying silent. However, staying silent is not always feasible. In situations where opinions are indirectly expressed by one’s observable actions, people may however try to hide their opinions via a more complex and intelligent strategy called obfuscation, which minimizes the information disclosed to others. This study proposes a formal opinion dynamics model to study the hitherto unexplored effect of obfuscation on public opinion formation based on the recently developed Action-Opinion Inference Model. For illustration purposes, we use our model to simulate two cases with different levels of complexity, highlighting that the effect of obfuscation largely depends on the subtle relations between actions and opinions.

1. Introduction

For diverse reasons, people may be unwilling to disclose their opinions to the public, especially when the topic is controversial. Instead, they may try to hide their opinions by adopting various strategies. As a prevalent behavior, hiding opinions has been studied in a number of opinion dynamics models. The majority of them assume that individuals hide their opinions by simply keeping silent (e.g., Gawronski et al., Citation2014; Ross et al., Citation2019; Sohn, Citation2019; Sohn & Geidner, Citation2016; Takeuchi et al., Citation2015). Although silence may help hide opinions from hearers of our words, it may not hide them from observers of our actions. That is, we usually learn others’ opinions by inferring them from their actions based on some universal knowledge about how opinions and actions relate (Tang & Chorus, Citation2019). In other words, in daily life, actions are known to signal opinions, and as such, keeping completely silent is no longer feasible when observers learn opinions by observing actions.Footnote1 For example, suppose a group of friends containing both vegetarians and omnivores dine in a restaurant where only two dishes are available: beef steak and vegetable salad. If an omnivore wants to hide the opinion that it is OK to eat meat, choosing steak is certainly a bad idea, but keeping silent is also impractical. A better strategy is to choose salad, because both vegetarians and omnivores are more or less likely to eat salad, but only omnivores will order steak. Such a strategy, where actions are chosen that provide minimal information about underlying beliefs and preferences, is called obfuscation; it can be conceived as a manner to minimize the information disclosed to others by producing ambiguity and uncertainty (Chorus et al., Citation2020). Obfuscation has been a popular topic in software engineering (You & Yim, Citation2010) and more recently in privacy research (Brunton & Nissenbaum, Citation2015), but has not yet attracted attention in the community of opinion dynamics. When obfuscation behaviors are prevalent, a failure to capture them in models of opinion dynamics could lead to a biased understanding of how hiding opinions affects public opinion formation.

In this paper, we present an obfuscation-based opinion dynamics model to study the role of obfuscation in public opinion formation by embedding the obfuscation mechanism (Chorus et al., Citation2020) into the Action-Opinion Inference (AOI) modeling framework (Tang & Chorus, Citation2019), where people choose actions according to their opinions and learn others’ opinions by interpreting their actions. Within this AOI framework, an obfuscating individual would hide her opinion by choosing the action that (i) is permitted by or in line with her opinion, yet (ii) releases the least amount of information about the opinion to others. This model fills the gap between existing models where hiding opinions equates to keeping silent and the reality that the mechanism of hiding opinions can be more subtle and complex than simply staying silent. As a result, our model can offer a more realistic and reasonable explanation of various social phenomena related to public opinion formation, particularly on (morally) sensitive topics. For example, incorporating obfuscation in the model can lead people to overestimate the popularity of the opinion that obliges the observed action. In a relatively simple setting, this may result in a significantly larger population believing in this opinion, which would have otherwise been different if only “silent-keeping” was considered.

The remainder of the paper is organized as follows: In Section 2, we review existing opinion dynamics models of hiding opinions and explain the foundation of our model. Section 3 describes the model in detail. In Section 4, we provide two illustrative examples abstracted from daily life and tales, in order to illustrate how this model works. Section 5 provides a brief summary and outlooks for further research.

2. Theoretical background

2.1. Hiding opinions in opinion dynamics

Opinion dynamics is one of the most popular and well-established fields in sociophysics. By modeling how opinions spread between individuals at a micro level, opinion dynamics models aim to explain macro-level phenomena such as polarization and consensus in a group of interacting individuals. Most opinion dynamics models pay little attention to the notion that people might want to hide their opinions and routinely assume that opinions can be directly observed, and that individuals always express opinions honestly (Mitsutsuji & Yamakage, Citation2020; Tang & Chorus, Citation2019). This assumption is likely to be unrealistic in circumstances where opinions are not completely visible, or individuals want to hide their opinions to avoid shame or to protect their privacy more generally.

Recently, however, a number of modelsFootnote2 involving opinion-hiding have been proposed. The majority of them are based on the so-called spiral of silence theory, postulating that due to the fear of social isolation, people are more likely to keep silent if they think they are in the minority (Noelle-Neumann, Citation1974). In spiral of silence models, the choice between keeping silent and expressing one’s opinion is determined by individual’s perception of the opinions of others. For example, Gawronski et al. (Citation2014) assume that the probability of expressing opinions is a negative function of the absolute difference between individual’s own opinion and her perceived public opinion. Others prefer a threshold rule: in Sohn and Geidner’s model (Sohn & Geidner, Citation2016), as well as a more recent one (Sohn, Citation2019), an individual speaks out if the intensity of her opinion is larger than the expression threshold, which is a personal and constant attribute. Following this tradition, Ross et al. (Citation2019) introduce a similar attribute called willingness to self-censor. The condition of speaking out is that an individual’s confidence in her opinion is larger than her willingness to self-censor, and the level of confidence is positively related to the proportional difference between the number of neighbors who agree and disagree with the individual.

Other models of hiding opinions follow different theories. For example, Grandi et al. (Citation2017) consider hiding or disclosing opinions as a strategy to achieve a certain goal by influencing others’ opinions. Fan and Pedrycz (Citation2015, Citation2016) adopt the social judgment theory and postulate that people remain silent if the intensity of their preference for one of two alternatives is not strong enough. As a conclusion, most models involving the behavior of hiding opinions, regardless of their theoretical basis, take it for granted that hiding opinions equates to keeping silent.

2.2. Obfuscation and action-opinion inference

Opinions are not always expressed by words but can also be revealed by actions. As argued in Section 1, when observers learn someone’s opinion by observing her actions, keeping silent is not (always) possible; we claim that in such a case, obfuscation becomes the best strategy.

In the past few decades, most obfuscation studies were conducted in the computer science domain, especially software engineering, where code obfuscation is a very popular topic (You & Yim, Citation2010). More recently, philosophers and social scientists started to pay attention to obfuscation with a special interest in how obfuscation can be used to defend one’s privacy on the Internet (Brunton & Nissenbaum, Citation2015; Davis, Citation2019; Doyle, Citation2018), and how obfuscation mitigates unfavorable moral reactions to morally disreputable economic exchanges (Rossman, Citation2014; Schilke & Rossman, Citation2018; Wherry et al., Citation2019).

In the context of a coordination problem, Dewan and Myatt (Citation2008) consider obfuscation as a technique of a leader to compete for audience by deliberately reducing the clarity of her message. Technically, obfuscation is modeled by manipulating “the variance of the noise in her speech” (Dewan & Myatt, Citation2008). In game theory, obfuscation is most closely related to intentional vagueness, i.e., deliberately choosing vague messages even if more precise alternatives are available (Blume & Board, Citation2014). A number of studies present the “game-theoretic rationale for vagueness” by showing that vagueness can “mitigate conflict” and “enhance efficiency” in a sender-receiver game (Blume & Board, Citation2014; De Jaegher, Citation2003; Serra-Garcia et al., Citation2011). Both Dewan-Myatt’s obfuscation and intentional vagueness mainly deal with verbal communications, and their analyses are often based on calculations of utilities. In this paper, we embed obfuscation in non-verbal communications where opinions are signaled by actions, and we are more interested in the effect of obfuscation on opinion dynamics rather than people’s utility or the equilibrium of a particular game.

In recent years, the concept of obfuscation has been introduced and formalized as a communication strategy in choice modeling. Chorus et al. (Citation2020) combine the notions of Bayesian inference and Shannon entropy, integrating them into a formal model of obfuscation-based decision-making. The idea is that a subject knows that her actions signal her underlying preferences (opinions) and selects the action that is in line with her preferences while providing as little as possible information to observers. The model is designed to describe the behaviors of humans whose actions are observed by others as well as the behaviors of autonomous agents under the surveillance of a human supervisor. In the model, agents choose actions based on a particular rule (here: opinion) that is unknown to the supervisor. Based on the observation of the agent’s action, the supervisor infers the opinion that motivates the action according to the Bayes’ Theorem. An obfuscating human or autonomous agent, being aware that it might be “punished” if the observer or supervisor learns that it has an “unwanted” opinion, will choose actions by maximizing the Shannon entropy generated by its choice while staying as close as possible to its opinion.

To utilize this mechanism in the context of opinion dynamics, we first need to formalize how opinions are learned by observing actions. In fact, such a formalization exists in the form of a so-called Action-Opinion Inference (AOI) model (Tang & Chorus, Citation2019). In the AOI model, the relation between opinions and actions is described by deontic logic: an opinion can oblige, permit, or prohibit an action. Equipped with the action-opinion relation, individuals “infer the opinions of others by observing and interpreting their actions” (Tang & Chorus, Citation2019). Based on the inference, individuals update their own opinions “according to the relative probability of each opinion in the neighborhood, calculated from the inferences of different opinions” (Tang & Chorus, Citation2019). As the final step, individuals choose new actions according to the newly updated opinions. The AOI model is compatible with Chorus et al.’s obfuscation mechanism not only because it formalizes the notion of “learning opinions by observing actions”, but also because of the deontic logic underlying the action-opinion relation, where an action may be driven by different opinions, and an opinion may permit different actions, allowing agents to obfuscate by choosing certain actions. If each action is driven by only one opinion, observers can then directly and correctly read opinions from actions, and there will be no room for obfuscation.

At the end of this section, we would like to point out the connection between the AOI model and social learning models in economics. Both types of models study how people infer (learn) and aggregate opinions (information) from their social environment (Golub & Sadler, Citation2016). In the AOI model, agents update their opinions in a Bayesian manner, which is a common setting in social learning (the so-called “Bayesian social learning”, e.g., Acemoglu et al., Citation2011; Gale & Kariv, Citation2003). The AOI model is also closely related to “observational social learning” in which agents observe choices made by their predecessors (Çelen & Kariv, Citation2004). Despite these similarities, the AOI model highlights the multiplicity of action-opinion relations, while social learning models may pay more attention to convergence and efficiency (Golub & Jackson, Citation2010; Lobel et al., Citation2009; Mossel et al., Citation2016). In particular, social learning models have a constant interest in convergence to the true/accurate opinion (Golub & Jackson, Citation2010; Jadbabaie et al., Citation2012) or the right/best action (Acemoglu et al., Citation2011) via learning, but the AOI model (or opinion dynamics models in general) does not involve any judgment or evaluation. We therefore conclude that the AOI model is located at the boundary (which itself is blurred) between opinion dynamics and social learning, and hence our obfuscation model – whose basis is the AOI model – relates to both disciplines other than opinion dynamics alone.

3. The model

In this section, we develop an opinion dynamics model of obfuscation by embedding the obfuscation mechanism (Chorus et al., Citation2020) in the framework of the Action-Opinion Inference (AOI) model (Tang & Chorus, Citation2019).

The basic model setup resembles the AOI model. We consider a population of N agents located on an undirected network G that describes how agents are connected. Agents are neighbors if they are directly connected in the network. Each agent i i=1,2,,N holds an invisible opinion oi from the opinion set O=o1,,ok,,oK, based on which she chooses a visible action ai from the action set A=a1,,ag,,aG. The relation between ok and ag is denoted by skg±1,0, where skg=1 implies ag is obliged by ok, skg=0 implies ag is permitted by ok, and skg=1 implies ag is forbidden by ok. All skgk=1,,K;g=1,,G compose the so-called action-opinion matrix S=skg. Agents are assumed to have the same action set, opinion set, and action-opinion matrix. This assumption will be relaxed in Section 4.2, where people may have difference perceptions of the relation between actions and opinions.

Assume that there is a fixed number of No obfuscators in the population who want to hide their opinions, and NNo non-obfuscators who do not care if their opinions are disclosed or not. Initially (i.e., stage 0), each agent (both obfuscators and non-obfuscators) is randomly assigned an opinion from the opinion set O, based on which she chooses an action from the action set A according to the rule of updating actions (the rule will be given in Section 3.1 and 3.2).

In each time step, an agent, whether an obfuscator or not, is randomly chosen to go through the following successive stages: (1) observing actions and inferring opinions, (2) updating opinions, and (3) updating actions. For the sake of clarity, we will demonstrate the behaviors of obfuscators and non-obfuscators separately.

3.1. Behavior of non-obfuscators

(0) choosing actions based on the initial opinions

Before any agent is chosen to go through the three main stages, each agent needs to choose an action based on her initial opinion. The rule of choosing actions of a non-obfuscator is as follows: if the opinion of a non-obfuscator i, oi=ok, obliges an action ag, she will certainly choose this action because it is the only option. Formally, the probability of choosing ag when holding ok, Pag|ok, equals 1 if skg=1. If oi=ok forbids ag, agent i will not choose ag. That is, Pag|ok=0 if skg=1. If oi=ok permits more than one action, agent i will choose one of these permitted actions with equal probability. Formally, Pag|ok=1W if skg=0, and W is the number of actions permitted by ok. To summarize:

(1) Pag|ok=101Wifskg=1ifskg=1ifskg=0(1)

(1) observing actions and inferring opinions

Once an agent is chosen, she first observes the actions chosen by her neighbors, based on which she infers neighbors’ opinions behind these actions. After observing neighbor j choosing action aj, agent i believes that the opinion of j is ok with probability Pioj=ok|aj, which takes the following form:

(2) Pioj=ok|aj=Paj|okz=1KPaj|oz(2)

where Paj|oz is the probability of choosing aj when holding opinion oz, and can be calculated by EquationEq. (1). We can derive EquationEq. (2) from the Bayes’ rule by assuming the prior probability Poz=1K for all z=1,2,,K. The rationale behind this assumption is that agents have no prior knowledge about which opinion is more likely to be adopted by their neighbors before observing their actions.

(2) updating opinions

After inferring the opinions of all neighbors, agent i evaluates the relative probability of each opinion in the neighborhood:

(3) Pˆiok=jMiPioj=ok|ajz=1KjMiPioj=oz|aj,k=1,2,,K(3)

where Mi is the set of all agent i’s neighbors. As a result of positive social influence (Flache et al., Citation2017), agent i will update her opinion to ok with probability Pˆiok. In case that other forms of social influence or mechanism are preferred, modelers can easily modify EquationEq. (3) accordingly.

(3) updating actions

In the last stage, the chosen agent updates her action based on her opinion that has just been updated in the previous stage. This stage follows the same rule as in stage 0 where non-obfuscators choose their actions based on their initial opinions. Then, the world goes to the next time unit.

To summarize, for a chosen agent, one time unit includes all the three stages: observing actions and inferring opinions, updating opinions, and updating actions. We define a time step as N successive time units, therefore on average in a time step everyone has one chance to update her opinion and action.

3.2. Behavior of obfuscators

The behavior of an obfuscator is the same as a non-obfuscator in stage 1 and 2. The only difference lies in the rule of choosing actions, which applies to both stage 0 and 3. First of all, an obfuscator is still governed by the action-opinion relation: she must choose the obliged action and cannot choose the forbidden action. As a result, an obfuscator can only play obfuscation when her opinion permits more than one action. Among all the actions that are permitted by her opinion, according to Chorus et al. (Citation2020), an obfuscator chooses the (permitted) action that reveals as little information as possible about the opinion by maximizing the uncertainty of her decision, measured by the Shannon entropy. For each action ag, the Shannon entropy is calculated by:

(4) Hag=k=1KPok|aglogPok|ag(4)

where Pok|ag is short for Pioj=ok|aj=ag, thus it can be calculated by EquationEq. (2). Larger entropy implies more uncertainty. If Hag=0 (i.e., the entropy is minimized), choosing ag reveals the full amount of information regarding the invisible opinion. To support this claim, we must show that Hag=0 only if there exists a k=k such that Pok|ag=1, and Pok|ag=0 for all kk. Fortunately, this has been proven by Shannon (Citation1948) as a basic property of the Shannon entropy. Meanwhile, the entropy is maximized, according to EquationEq. (4), when Pom|ag=Pon|ag for all m,n=1,,K, that is, the observer has no knowledge about which opinion is more likely to be the opinion of an agent choosing ag. In practice, this perfect maximization is not always achievable due to the restriction of the action-opinion matrix.

Formally, an obfuscator i will choose ai according to:

(5) ai=argmaxagAiHag(5)

where Ai is the set of actions available to i. In other words, Ai contains all the actions permitted or obliged by oi.

It is worth noting that both obfuscators and non-obfuscators know nothing about the identities (i.e., obfuscator or non-obfuscator) of their neighbors, nor do they know the number of obfuscators in the population. The assumption can be relaxed if modelers want to study more intelligent agents who are able to learn the identities of others.

gives a brief summary of the model. Firstly, each agent is randomly assigned an opinion. Then obfuscators and non-obfuscators choose actions based on different rules (stage 0). Afterward, a random agent is selected to update her opinion and action through a three-stage process: inferring opinions of others (stage 1), updating opinion based on the inference (stage 2), and updating action based on the updated opinion (stage 3).

Figure 1. Illustration of the model.

Figure 1. Illustration of the model.

4. Illustrative examples

As we will soon witness in this section, the effect of obfuscation on public opinion largely depends on the relations between actions and opinions. Understanding the effect of obfuscation in a particular case requires running simulations of the model under particular conditions. To illustrate how this works, we provide two examples. The first example that describes the dynamics of vegetarians and omnivores is extremely simple, aiming to provide a step-by-step demonstration. The second example, trying to explain the ironic situation in The Emperor’s New Clothes, is more subtle and complex, as people with different opinions have different perceptions of the relation between actions and opinions. It is important to note here, that the sole aim of these examples is to illustrate the workings of the obfuscation-based opinion dynamics model – as such we refrain from drawing any generic (i.e., not specific to the example) conclusions about the potential effect of obfuscation on opinion dynamics. For that, a larger number of more elaborate case studies are needed which are preferably grounded in real life opinion formation situations.

4.1. The battle between vegetarians and omnivores

We first look into a very simple case, the vegetarian-omnivore example mentioned in Section 1. Here we assume that there are N=10 friends going to the restaurant. Initially, NVeg=5 of them are vegetarians and Nomn=5 of them are omnivores. Given the relatively small population, it is reasonable to assume that everyone can observe the action of everyone else. We summarize this case by the following parameters and conditions: N=10, G is a complete graph (i.e., everyone is a neighbor of everyone else), A=a1=ChooseSteak,a2=ChooseSalad, O={o1=Veg,o2=Omn} (“Veg” is short for “Vegetarian”, and “Omn” is short for “Omnivore”), and the action-opinion matrix SVO (“VO” stands for “the battle between Vegetarians and Omnivores”):

SVO=SteakSaladVeg1+1Omn00

which means a vegetarian is prohibited from choosing steak and can only choose salad, while an omnivore can choose between steak and salad. Without any calculation, we can already see that an obfuscating omnivore will choose salad, and a non-obfuscating omnivore will choose randomly (i.e., flip a coin) between steak and salad. It is also worth noting that whether a person obfuscates does not depend on the chosen action. For example, an obfuscating vegetarian should make the same choice (i.e., salad) as a non-obfuscating vegetarian. However, the obfuscating vegetarian chooses salad because it gives the minimum information, while the non-obfuscating vegetarian makes the same choice because it is the only permitted option, regardless of how much information it releases. In practice they choose the same action, but their motivations for doing so are different.

The running time is set to be 500, which is sufficiently long to reach a stable outcome. shows how the number of obfuscators in the population affects public opinion, based on which we can conclude that obfuscation, in this particular case, suppresses the spread of omnivorism and promotes the popularity of vegetarianism. However, the effect is bounded: in ), even if everyone is an obfuscator, in equilibrium, there still exist a few omnivores (around 1 to 2), implying that obfuscation cannot completely eliminate the existence of omnivorism.

Figure 2. The battle between vegetarians and omnivores: population of believers in each opinion versus time step.

Figure 2. The battle between vegetarians and omnivores: population of believers in each opinion versus time step.

To further explore the relation between obfuscation and public opinion, we run the simulation 100 times for each No. In , fˉVeg (i.e., the y axis) is the fraction of vegetarians in the population averaged over the last 50 time steps of each simulation realization. Compared to , provides more details. We can see that although obfuscation (represented by No) has a significant impact on public opinion (represented by fˉVeg), it is not a fully determining factor, as there remains a remarkable degree of variation across realizations regardless of No. This statement comes from the observation that even if all conditions and parameters (including No) are the same, the public opinion in each realization can be very different. For example, when No=0, the lowest fˉVeg is close to 0.35, and the highest is about 0.6. However, there is a trend that this variation decreases as the number of obfuscators increases. As another interesting finding, we can conclude that obfuscation is likely to reduce the variation in public opinion for this particular case.

Figure 3. The battle between vegetarians and omnivores: fraction of vegetarians (fˉVeg) versus the number of obfuscators (No). For each No, we run 100 realizations of the simulation. fˉVeg is obtained by averaging the fraction of vegetarians in the last 50 time steps of a realization. Each data point represents one realization. The horizontal position of each data point is slightly adjusted in order to reduce overlap. The line across the figure is the smoothed conditional mean, and the shaded area indicates the 95% confidence interval.

Figure 3. The battle between vegetarians and omnivores: fraction of vegetarians (fˉVeg) versus the number of obfuscators (No). For each No, we run 100 realizations of the simulation. fˉVeg is obtained by averaging the fraction of vegetarians in the last 50 time steps of a realization. Each data point represents one realization. The horizontal position of each data point is slightly adjusted in order to reduce overlap. The line across the figure is the smoothed conditional mean, and the shaded area indicates the 95% confidence interval.

The rationale behind the simulation result lies in the discrepancy in observer’s inference and the reality: from EquationEq. (2) and SVO, we know that observers believe a salad-eating agent is an omnivore with a probability of 1/3. However, because an obfuscating omnivore always choose salads, this probability is in fact larger than 1/3 in a population containing obfuscators. In the extreme case where everyone is an obfuscator, the probability increases to 1/2, the same as the probability that a salad-eating agent is a vegetarian. Such a discrepancy leads to an underestimation of the population of omnivores (or, equivalently, overestimation of the population of vegetarians). Consequently, vegetarianism becomes more popular than omnivorism because of positive social influence. At the same time, omnivorism will not go extinct because observers believe that omnivores are always likely to exist (with a relatively small probability) even if everyone chooses salads.

The rationale described above is formally expressed in the Appendix, where analytical results of this example are derived. According to the derivation, the fraction of vegetarians, averaged over all realizations (trajectories) of the dynamics, should converge to 1+θ2+θ over time, where θ=NoN is the fraction of obfuscators in the population. This conclusion is validated by the simulation result in , where the average fˉVeg (averaged over 100 independent realizations) is well approximated by 1+θ2+θ. Furthermore, the derivation shows that the average fˉVeg only depends on the fraction of obfuscators (θ) and the action-opinion matrix (SVO), while other conditions such as the size of the population (N) and the initial distribution of each opinion (NVeg and NOmn) are irrelevant. Unexpectedly, the number of neighbors of each agent is also irrelevant, as long as everyone has the same number of neighbors.Footnote3

Finally, we show that the effect of obfuscation on public opinion largely depends on the relations between actions and opinions. If we replace the omnivores here with carnivores (abbreviated to “Car”) who only consume meats, the matrix is now SVC(“VC”stands for “The battle between Vegetarians and Carnivores”):

SVC=SteakSaladVeg1+1Car+11

It is obvious that obfuscation plays no role in this new case as vegetarians can only choose salad and carnivores can only choose steak. In other words, an obfuscator behaves the same as a non-obfuscator. The opinion dynamics described by SVC has been solved analytically in the studies of the voter model (Krapivsky et al., Citation2010), from which we learn that the population would eventually reach a consensus of either vegetarianism or carnivorism (Tang & Chorus, Citation2019). As a result, the conclusion drawn from SVO is invalid for the situation described by SVC.

4.2. The emperor’s new clothes

The Emperor’s New Clothes is a famous tale written by Hans Christian Andersen in 1837. The general plot is about how two swindlers pretending to be weavers, convince the Emperor that the suit of clothes they made is invisible to stupid people. Everyone in the country, after observing the naked Emperor, out of fear of being considered stupid, pretends that they could see the clothes, until a child spoke out the truth.

For sociologists, the tale, as a symbolic example of “support for a public lie” (Centola et al., Citation2005), is of particular interest because the ironic phenomenon that everyone pretends that they can see the clothes needs further explanation: besides the fear of being labeled stupid, is there any other mechanism underlying the phenomenon? One of the most popular explanations uses the concept of pluralistic ignorance (e.g., Bjerring et al., Citation2014; Centola et al., Citation2005; Hansen, Citation2012). Pluralistic ignorance describes a situation where most people privately reject or disapprove an opinion, but (incorrectly) believe that the opinion has been widely accepted by others (Miller & McFarland, Citation1987). To explain the tale by pluralistic ignorance, citizens in the tale are assumed to be “disbelievers” as they in fact think the Emperor is naked. Then the phenomenon is achieved when all disbelievers publicly praise the invisible suit based on the false belief that everyone else thinks the Emperor is not naked (Bjerring et al., Citation2014).

“Naked emperors are easy to find but hard to explain.” (Centola et al., Citation2005). Despite being a popular practice to explain the tale, pluralistic ignorance overlooks the dynamics of opinions in the population. In fact, keeping one’s opinion unchanged is a basic condition of pluralistic ignorance. For example, in Centola’s model (Centola et al., Citation2005), the population is divided into “true believers” who always admire the Emperor, and “disbelievers” who (privately) think the Emperor is naked. Both true believers and disbelievers are not allowed to change their private opinions, regardless of their compliance decision.Footnote4 This naturally raises the following questions: if individuals are allowed to change opinions, can this “public lie” become a “(false) public opinion” where everyone believes that the Emperor is dressed? In the other extreme, can this “public lie” become a “public truth” where everyone not only privately believes but also publicly claims that the Emperor is naked?

To answer these questions, we need to take an alternative approach. In the rest of the section, we will explain the tale from the perspective of opinion dynamics and obfuscation by investigating the role of obfuscation in the dynamics of opinions among citizens, including both “true believers” and “disbelievers”.

While some citizens believe that the Emperor is naked, others may believe that the Emperor is dressed, and that they cannot see the clothes because they are stupid. Naturally, the latter will have the false imagination that some other citizens can see the clothes. To summarize, there will be three opinions involved in the story:

  • o1: I can see the clothes because I am not stupid.

  • o2: I cannot see the clothes because I am stupid.

  • o3: I cannot see the clothes because the Emperor is naked.

It should be noted that o1 is imaginary, as in fact no one would hold this opinion.

Relevant actions, as one can imagine, include:

  • a1: publicly mock the Emperor/point out that the Emperor is naked.

  • a2: keep silent.

  • a3: publicly admire the Emperor’s clothes.

Citizens who believe in o2 perceive the following action-opinion relation:

SECo2=a1a2a3o111+1o2100o3001

where “EC” stands for “Emperor’s New Cloth”. Because citizens with o1only exist in the imagination of citizens with o2, the row that describes o1 is completely determined by how citizens with o2 think: without any social pressure, citizens with o1 are expected (by citizens with o2) to have no motivation to mock the Emperor (a1) or keep silent (a2). Meanwhile, citizens with o2, although they cannot see the clothes, will never mock the Emperor (a1) because they believe the clothes do exist. For citizens with o3 who disbelieve the lie, citizens with o2 assume that they would never admire the Emperor (a3).

Citizens with o2 are facing the (maybe imaginary) social pressure of being labeled as stupid people, and therefore have the incentive to hide their opinion by obfuscation. It seems that an obfuscator who believes in o2 should choose a2 over a3 due to the fact that the entropy of a2 is larger than that of a3 according to SECo2. However, in this particular case, citizens with o2 believe that the pressure only comes from those who believe in o1, because citizens with o3, by definition, do not accept the swindlers’ lie, hence they would not consider citizens with o2 to be stupid. As a result, citizens with o2 only care about the judgment from citizens with o1, and will choose actions based on the action-opinion relation perceived by citizens with o1 instead of their own perception SECo2. Because citizens with o1 only exist in the imagination of citizens with o2, the perception of the action-opinion relation by citizens with o1 is determined by citizens with o2, and is therefore denoted by SECo1|o2:

SECo1|o2=a1a2a3o111+1o2100

The absence of o3 is because we assume citizens with o2 believe that citizens with o1 would ignore the existence of o3. The rationale behind this assumption is that citizens with o1 might be so confident in their opinion that they do not expect others would think the Emperor is naked.Footnote5 It is clear from SECo1|o2 that an obfuscator with o2 would choose a3, because choosing a2 is a signal of being stupid in the eyes of citizens with o1, as citizens with o2 believe.

Observing more people choosing a3 makes citizens with o2 believe that there are more citizens with o1 (i.e., Pˆio1 increases, where i denotes citizens with o2). However, they cannot change their opinion from o2 to o1, therefore observing a3 only makes them more confident in their current opinion o2.

Now, we consider citizens with o3, the disbelievers. As they believe the Emperor is naked, to them, o1 does not exist. Therefore, their perception of the action-opinion relation isFootnote6

SECo3=a1a2a3o2100o3001

Citizens with o3 also have incentives to play obfuscation as they may not want to be considered stupid by citizens with o2. To hide their opinion, instead of referring to their own perception SECo3, they should utilize SECo2 because they think the pressure comes from citizens with o2. SECo2 implies that obfuscators with o3 should choose a2 to maximize the entropy.

The opinion dynamics of the citizens can be summarized as follows:

  • Citizens with o2: their perception of the action-opinion relation is encoded in SECo2. Non-obfuscators choose between a2 and a3 with equal probability according to SECo2; obfuscators choose a3 according to SEC(o1|o2). For both non-obfuscators and obfuscators, observing someone choosing a3 makes them more confident in their current opinion. The inferring process after observing other actions (i.e., stage 1) relies on SECo2 as described in Section 3.

  • Citizens with o3: their perception of the action-opinion relation is encoded in SECo3. Non-obfuscators with o3 will choose between a1 and a2 with equal probability according to SECo3; obfuscators with o3 will choose a2 according to SECo2. The observing actions and inferring opinions process (i.e., stage 1) relies on SECo3 as described in Section 3.

Readers must have realized that this case seems to be more complex than what we presented in Section 3. This is because the assumption that agents have the same action-opinion matrix has been relaxed. As we have discussed above, citizens with different opinions now have different perceptions of the action-opinion relation in this system. This is because citizens with o2 have an imaginary type of neighbors: citizens with o1. In addition, due to the different sources of social pressure (i.e., the motivation for hiding one’s opinion through obfuscation), citizens also rely on different action-opinion matrices to decide how to obfuscate. Namely, obfuscators with o2 believe that the pressure comes from citizens with o1, therefore they choose actions according to the perception of these imaginary neighbors SEC(o1|o2). Meanwhile, obfuscators with o3 believe that the pressure comes from citizens with o2, therefore they rely on SEC(o2) to hide opinions.

Indeed, the assumption that everyone knows and uses the same action-opinion matrix significantly simplifies the modeling process. Such a simplification is reasonable in many situations such as the vegetarian-omnivore case (Section 4.1), but here we show that it can be relaxed in order to capture the special mind-sets of different types of citizens.

Under a set of reasonable parameters and conditions, including (1) total population N=100, (2) initial population of believers in o2 and o3 are equal, and (3) everyone knows everyone else in the system (i.e., G is a complete graph), we obtain the simulation results shown in (dynamics of opinions) and (dynamics of actions). If none of the citizens obfuscates (i.e., No=0), o2 and o3 will dominate the population in turn, and the average population believing in each opinion over time is half of the whole population ()). Meanwhile, about half of the population will keep silent (a2), and the rest of the population is, on average, equally divided between citizens who mock the Emperor (a1) and admire the Emperor (a3) ()).

Figure 4. The emperor’s new clothes: opinion dynamics of the citizens. Population of believers in o1 is always zero and thus is not plotted.

Figure 4. The emperor’s new clothes: opinion dynamics of the citizens. Population of believers in o1 is always zero and thus is not plotted.

Figure 5. The emperor’s new clothes: population dynamics of actions chosen by the citizens versus time step.

Figure 5. The emperor’s new clothes: population dynamics of actions chosen by the citizens versus time step.

To conclude, if no one wants to obfuscate, there is still a considerable number of citizens mocking the Emperor even when the majority is silent. However, as the number of obfuscators increases, the popularity of o2 gradually grows (, ). When everyone becomes an obfuscator (No=100), it only takes a few time steps (note that each time step contains 100 individual updates) for the whole population to reach a consensus of o2 ()), that is, everyone becomes the “true believer” in Centola’s model, and the “public lie” becomes the “(false) public opinion” that the Emperor is dressed. In terms of actions, everyone will eventually admire the Emperor when everyone obfuscates ()).

Now let’s consider the other extreme: what if citizens, opposite to obfuscation, would like to be as transparent as possible to observers? In other words, what if citizens want their opinions to be correctly and clearly known by others? Transparent citizens with o2 will choose a2 according to SECo2: although a3 has a smaller entropy, it is misleading because it signals that the underlying opinion is more likely to be o1. They rely on their own perception SECo2 instead of SECo1|o2 (as the obfuscators do) because transparency is usually not directly related to the pressures from others. Meanwhile, transparent citizens with o3, according to SECo3, will choose a1 because it directly signals that the underlying opinion is o3. A population full of transparent citizens, with the same parameters and initial conditions as in and , would produce a completely different result (): in a few time steps, everyone will believe that the Emperor is naked (o3), and mock the Emperor (a1). In the context of Centola’s model, this means everyone is now a “disbeliever”, and the “public lie” is replaced by the “public truth”.

Figure 6. The emperor’s new clothes: population dynamics of opinions and actions chosen by the citizens. every citizen is transparent.

Figure 6. The emperor’s new clothes: population dynamics of opinions and actions chosen by the citizens. every citizen is transparent.

To conclude, by applying the obfuscation-based opinion dynamics model we have provided an alternative explanation for the collective behavior in the tale by modeling obfuscation in public opinion formation. The phenomenon that everyone sincerely admires the invisible clothes can emerge from a population full of obfuscators. The fundamental difference with pluralistic ignorance is that in our analyses, citizens not only publicly admire the invisible clothes but also privately believe the clothes exist. On the contrary, if there are fewer obfuscators, eventually more citizens will believe that the Emperor is naked and dare to speak out the truth. Furthermore, if everyone would like to openly disclose their opinions (i.e., being transparent), there will soon be no believers in the swindlers’ lie.

4.3. Qualitative conclusions from the examples

These two examples validate our early judgment that a universally correct answer to “how obfuscation affects public opinion” does not exist, but there are still some qualitative conclusions that worth mentioning. From the first example, we can arrive at a hypothesis that if an opinion only allows one action (vegetarianism in this example), it will be generally more popular than others in the presence of obfuscators. As argued in Section 4.1, this can be attributed to observer’s overestimation of the popularity of this opinion. Although the hypothesis is not applicable in the second example (because o1, the opinion that allows only one action, is imaginary), a similar logic can help us understand the simulation outcome. Obfuscators believing in o2 play the same role as believers in o1 because they always choose the same action a3, therefore we could conceptually divide o2 into two categories: o2 that is believed by obfuscators (denoted by “obfuscating o2”) and o2 that is believed by non-obfuscators (denoted by “non-obfuscating o2”). Obfuscating o2 can be viewed as an opinion that only allows one action, hence is expected to be more popular than other opinions (such as non-obfuscating o2) according to the hypothesis. As a result, given the total number of believers in o2 at any instance, the more people believe in obfuscating o2, the more popular o2 will be in the future. Meanwhile, the number of obfuscators (No) determines the initial number of believers in obfuscating o2, and therefore is positively related to the popularity of o2.

5. Conclusion and discussion

In the literature of opinion dynamics, we have witnessed two levels of details: most opinion dynamics models do not include the behavior of hiding opinions as they assume that opinions are always expressed publicly and truthfully; and studies into hiding opinions do not include the strategy of obfuscation as they assume that hiding opinions equates to keeping silent. These two omissions hamper our understanding of real-life opinion dynamics. This study contributes to the opinion dynamics literature by proposing an obfuscation-based opinion dynamics model that embodies a more complex and in some cases more realistic form of hiding opinions than keeping silent. The model embeds the obfuscation mechanism into the framework of the Action-Opinion Inference model, by formalizing a strategy of choosing the action that gives the least information about the underlying opinion.

For illustration purposes, we run the simulation of the model for two cases with different levels of complexity. The first vegetarian-omnivore case is relatively simple, providing a step-by-step demonstration. Simulation results indicate that in this particular case, obfuscation promotes the opinion (i.e., vegetarianism) that only allows one action while the more inclusive opinion (i.e., omnivorism) maintains a low popularity. The second case explains why the citizens in Han Christian Andersen’s tale admire the Emperor’s invisible clothes from the perspective of obfuscation. It is more complex because in this case obfuscators with different opinions have different perceptions of the relation between actions and opinions, and they rely on different perceptions to choose actions due to different motivations of obfuscation. The result suggests that obfuscation is able to facilitate the spread of the false opinion that the Emperor is dressed, while transparency can help popularize the true opinion that Emperor is naked.

Overall, the obfuscation-based opinion dynamics model expands the boundary of opinion dynamics studies by enabling agents to have a more intelligent strategy of hiding their opinions behind their actions. We hope that our study can initiate further discussions and developments about obfuscation and related notions. Directions of further research include (i) relaxing or modifying several assumptions such as undirected networks, positive influence, and sequential updating; (ii) calibrating the model to empirical data of public opinions to investigate obfuscation in real-world issues; and (iii) exploring concepts that are similar to (but subtly different from) obfuscation such as deception (Castelfranchi & Tan, Citation2001), strategic ambiguity (Eisenberg, Citation1984) and intentional vagueness (Blume & Board, Citation2014) as well as their roles in opinion dynamics.

Acknowledgments

This study was funded by the European Research Council as part of the Consolidator Grant BEHAVE (grant agreement No. 724431).

Notes

1 This conclusion is still valid even if we generalize “keeping silent” to “doing nothing”: although remaining quite in a debate is effortless, eating nothing in a dinner party seems less practical.

2 In some of these models (e.g., Gawronski et al., Citation2014 & Ross et al., Citation2019), opinions are fixed, and agents update their choices between expressing opinions and keeping silent. We regard them as an extended class of opinion dynamics models.

3 Readers should be aware that (1) the derivation benefits from the simplicity of SVO and the assumption that everyone has the same number of neighbors; and (2) the conclusions made here are completely based on the derivation. It is unclear if they are valid in other cases.

4 In the extension of Centola’s model within the same paper, disbelievers with false enforcement are allowed to convert to true believers, but true believers cannot convert to disbelievers by default.

5 Other forms of SEC(o1|o2) may also be feasible. We employ the current form because it helps illustrate the idea that different people have different perceptions of action-opinion relations, and they obfuscate based on different action-opinion matrices.

6 In this example, SEC(o1|o2) and SECo3 are composed of subsets of identical rows of SECo2. This does not mean the rows in different matrices for the same opinion are always the same. They only depend on agent’s perceptions. We thank the referee for pointing it out.

7 Most part of the derivation was modified from Tang and Chorus (Citation2019).

References

  • Acemoglu, D., Dahleh, M. A., Lobel, I., & Ozdaglar, A. (2011). Bayesian learning in social networks. The Review of Economic Studies, 78(4), 1201–1236. https://doi.org/10.1093/restud/rdr004
  • Bjerring, J. C., Hansen, J. U., & Pedersen, N. J. L. L. (2014). On the rationality of pluralistic ignorance. Synthese, 191(11), 2445–2470. https://doi.org/10.1007/s11229-014-0434-1
  • Blume, A., & Board, O. (2014). Intentional vagueness. Erkenntnis, 79(4), 855–899. https://doi.org/10.1007/s10670-013-9468-x
  • Brunton, F., & Nissenbaum, H. (2015). Obfuscation: A user’s guide for privacy and protest. The MIT Press.
  • Castelfranchi, C., & Tan, Y. H. (Eds.). (2001). Trust and deception in virtual societies. Kluwer.
  • Çelen, B., & Kariv, S. (2004). Observational learning under imperfect information. Games and Economic Behavior, 47(1), 72–86. https://doi.org/10.1016/S0899-8256(03)00179-9
  • Centola, D., Willer, R., & Macy, M. (2005). The emperor’s dilemma: A computational model of self-enforcing norms. American Journal of Sociology, 110(4), 1009–1040. https://doi.org/10.1086/427321
  • Chorus, C., Van Cranenburgh, S., Daniel, A. M., Sandorf, E. D., Sobhani, A., & Szép, T. (2020). Obfuscation maximization-based decision-making: Theory, methodology and first empirical evidence. Mathematical Social Sciences, 109, 28–44. https://doi.org/10.1016/j.mathsocsci.2020.10.002
  • Davis, R. C. (2019). Obfuscating authorship: Results of a user study on nondescript, a digital privacy tool [working paper]. CUNY Academic Works.
  • De Jaegher, K. (2003). A game-theoretic rationale for vagueness. Linguistics and Philosophy, 26(5), 637–659. https://doi.org/10.1023/A:1025853728992
  • Dewan, T., & Myatt, D. P. (2008). The qualities of leadership: Direction, communication, and obfuscation. In American political science review 102(3)(pp. 351–368). doi:10.1017/S0003055408080234
  • Doyle, T. (2018). Privacy, obfuscation, and propertization. IFLA Journal, 44(3), 229–239. https://doi.org/10.1177/0340035218778054
  • Eisenberg, E. M. (1984). Ambiguity as strategy in organizational communication. Communication Monographs, 51(3), 227–242. https://doi.org/10.1080/03637758409390197
  • Fan, K., & Pedrycz, W. (2015). Emergence and spread of extremist opinions. Physica A: Statistical Mechanics and Its Applications, 436, 87–97. https://doi.org/10.1016/j.physa.2015.05.056
  • Fan, K., & Pedrycz, W. (2016). Opinion evolution influenced by informed agents. Physica A: Statistical Mechanics and Its Applications, 462, 431–441. https://doi.org/10.1016/j.physa.2016.06.110
  • Flache, A., Mäs, M., Feliciani, T., Chattoe-Brown, E., Deffuant, G., Huet, S., & Lorenz, J. (2017). Models of social influence: Towards the next frontiers. Journal of Artificial Societies and Social Simulation, 20(4), 4. https://doi.org/10.18564/jasss.3521
  • Gale, D., & Kariv, S. (2003). Bayesian learning in social networks. Games and Economic Behavior, 45(2), 329–346. https://doi.org/10.1016/S0899-8256(03)00144-1
  • Gawronski, P., Nawojczyk, M., & Kulakowski, K. (2014). Opinion formation in an open system and the spiral of silence. arXiv Preprint arXiv:1407.2742.
  • Golub, B., & Sadler, E. (2016). Learning in social networks. In A. Galeotti & B. W. Roger (Eds.), The oxford handbook of the economics of networks (pp. 504–542). Oxford University Press.
  • Golub, B., & Jackson, M. O. (2010). Naive learning in social networks and the wisdom of crowds. American Economic Journal: Microeconomics, 2(1), 112–149. doi:10.1257/mic.2.1.112.
  • Grandi, U., Lorini, E., Novaro, A., & Perrussel, L. (2017). Strategic disclosure of opinions on a social network. In Proceedings of the 16th conference on autonomous agents and multi agent systems (pp. 1196–1204). Sao Paulo, Brazil.
  • Hansen, J. U. (2012). A logic-based approach to pluralistic ignorance. In J. D. Vuyst & L. Demey (Eds.), Future directions for logic—proceedings of PhDs in Logic III (pp. 67–80). College Publications.
  • Jadbabaie, A., Molavi, P., Sandroni, A., & Tahbaz-Salehi, A. (2012). Non-Bayesian social learning. Games and Economic Behavior, 76(1), 210–225. https://doi.org/10.1016/j.geb.2012.06.001
  • Krapivsky, P. L., Redner, S., & Ben-Naim, E. (2010). A kinetic view of statistical physics. Cambridge University Press.
  • Lobel, I., Acemoglu, D., Dahleh, M., & Ozdaglar, A. (2009). Rate of convergence of learning in social networks. In Proceedings of the American Control Conference (pp. 2825–2830), St. Louis, MO, USA.
  • Miller, D. T., & McFarland, C. (1987). Pluralistic ignorance: When similarity is interpreted as dissimilarity. Journal of Personality and Social Psychology, 53(2), 298. https://doi.org/10.1037/0022-3514.53.2.298
  • Mitsutsuji, K., & Yamakage, S. (2020). The dual attitudinal dynamics of public opinion: An agent-based reformulation of LF Richardson’s war-moods model. Quality & Quantity, 54(2), 439–461. https://doi.org/10.1007/s11135-019-00938-x
  • Mossel, E., Olsman, N., & Tamuz, O. (2016). Efficient bayesian learning in social networks with gaussian estimators. In 54th annual allerton conference on communication, control, and computing (allerton) (pp. 425–432). IEEE, Monticello, IL, USA.
  • Noelle-Neumann, E. (1974). The spiral of silence a theory of public opinion. Journal of Communication, 24(2), 43–51. https://doi.org/10.1111/j.1460-2466.1974.tb00367.x
  • Ross, B., Pilz, L., Cabrera, B., Brachten, F., Neubaum, G., & Stieglitz, S. (2019). Are social bots a real threat? An agent-based model of the spiral of silence to analyse the impact of manipulative actors in social networks. European Journal of Information Systems, 28(4), 394–412. https://doi.org/10.1080/0960085X.2018.1560920
  • Rossman, G. (2014). Obfuscatory relational work and disreputable exchange. Sociological Theory, 32(1), 43–63. https://doi.org/10.1177/0735275114523418
  • Schilke, O., & Rossman, G. (2018). It’s only wrong if it’s transactional: Moral perceptions of obfuscated exchange. American Sociological Review, 83(6), 1079–1107. https://doi.org/10.1177/0003122418806284
  • Serra-Garcia, M., Van Damme, E., & Potters, J. (2011). Hiding an inconvenient truth: Lies and vagueness. Games and Economic Behavior, 73(1), 244–261. https://doi.org/10.1016/j.geb.2011.01.007
  • Shannon, C. E. (1948). A mathematical theory of communication. The Bell System Technical Journal, 27(3), 379–423. https://doi.org/10.1002/j.1538-7305.1948.tb01338.x
  • Sohn, D. (2019). Spiral of silence in the social media era: A simulation approach to the interplay between social networks and mass media. In Communication Research . https://doi.org/10.1177/0093650219856510
  • Sohn, D., & Geidner, N. (2016). Collective dynamics of the spiral of silence: The role of ego-network size. International Journal of Public Opinion Research, 28(1), 25–45. https://doi.org/10.1093/ijpor/edv005
  • Takeuchi, D., Tanaka, G., Fujie, R., & Suzuki, H. (2015). Public opinion formation with the spiral of silence on complex social networks. Nonlinear Theory and Its Applications, IEICE, 6(1), 15–25. https://doi.org/10.1587/nolta.6.15.
  • Tang, T., & Chorus, C. G. (2019). Learning opinions by observing actions: Simulation of opinion dynamics using an action-opinion inference model. Journal of Artificial Societies and Social Simulation, 22(3), 3. https://doi.org/10.18564/jasss.4020
  • Wherry, F. F., Seefeldt, K. S., & Alvarez, A. S. (2019). To lend or not to lend to friends and kin: Awkwardness, obfuscation, and negative reciprocity. Social Forces, 98(2), 753–793. doi: 10.1093/sf/soy127
  • You, I., & Yim, K. (2010). Malware obfuscation techniques: A brief survey. In 2010 international conference on broadband, wireless computing, Communication and Applications, 297–300. Fukuoka, Japan

Appendix

In this Appendix, we derive an analytical solution of the dynamics described in Section 4.1, namely the Vegetarian-Omnivore example, whose simulation results have already been given in and . The derivation closely follows Tang and ChorusFootnote7 (Citation2019), which itself is an extension of a typical derivation of the voter model (Krapivsky et al., Citation2010). It should be noted that the derivation significantly benefits from the simplicity of SVO, and is therefore only applicable to this particular case.

Recall the action-opinion matrix used in the example:

SVO=SteakSaladVeg1+1Omn00

We start by rephrasing the notion of opinions in a binary fashion: denote the opinion of an agent i as a binary variable oi which can only take one of two values ±1. oi=1 means the agent is a vegetarian, and oi=1 means the agent is an omnivore. Additionally, we denote a1=Steak, and a2=Salad.

The probability that agent i changes her opinion (“flip rate”), based on Eq. (Equation2) and (Equation3), can be written as:

(A1) wi=121oizjMio±1Poj=o|ajo(A1)

where z is the number of neighbors of each agent (i.e., lattice coordination number) and is assumed to be constant. Mi is the set of all the neighbors of agent i. Poi=o|aj is equivalent to Pioi=o|aj as the inference is the same for everyone who observes aj.

Following Tang and Chorus (Citation2019), we focus on the average opinion of each agent Ri,t≡<oit>. Note that by “<>” we mean the average <FX>≡xPX=xFx, therefore Ri,toPoit=oo is the opinion of agent i averaged over all possible values of oi, which can be roughly interpreted as the opinion of agent i averaged over all (countless) realizations (or “trajectories” in the language of statistical physics) of the dynamics at time t. It is neither the average opinion of all agents nor agent i’s opinion averaged over time. To be precise, suppose there are Q systems (i.e., realizations) S1,,Sq,,SQ that all evolve independently from the same initial system S0, and the opinion of agent i in each system Sq at time t is oqit, then Ri,t=limQq=1QoqitQ. To keep things tidy, we omit t and write Ri.

The paper describes a discrete-time model where an agent is chosen to update her opinion in a time unit, and N successive time units define a time step. The discreteness helps implement simulation but not derivation. Here, we alternatively assume that time t in the dynamics is continuous to facilitate the derivation. The continuous alternative, as we will witness at the end of this Appendix, can produce good approximation of the discrete model given sufficiently long time.

In a continuous-time context, the dynamics of agent i’s opinion in a sufficiently short time interval Δt is:

(A2) oit+Δt=  oitoitwithprobability1wiΔtwithprobabilitywiΔt(A2)

According to Krapivsky et al. (Citation2010), the evolution of Ri is:

(A3) dRidt=d<oi>dt=2oiwi(A3)

The derivation of the last term is based on (A2). By substituting (A1) into (A3) and using the trick that oi2=1, we obtain:

(A4) dRidt=Ri+1zjMi<oPoj=o|ajo>(A4)

By denoting <oPoj=o|ajo> as Rj, (A4) can be expressed in a more elegant form:

(A5) dRidt=Ri+1zjMiRj(A5)

To describe the whole population, we define the “mean magnetization” (analogous to the same concept in spin dynamics) m1NiRi, which is the average opinion of the population averaged over all realizations. The mean magnetization m represents public opinion: if m=1, everyone is a vegetarian in all realizations without exception; if m=1, everyone is an omnivore in all realizations without exception; if m=0, the population as a whole does not have a preference. Note that dmdt=d1NiRidt=1NidRidt, then summing (A5) over all agents leads to:

(A6) Ndmdt=iRi+1zijMiRj(A6)

Note that Rioi=oPoi=oo and o can only take two values ±1, we have:

(A7) Ri=Poi=1P(oi=1)=2Poi=11(A7)

Similarly, we have Rj=Poj=1|ajPoj=1|aj, hence:

(A8) Rj=2Poj=1|aj1(A8)

According to the definition of “”:

(A9) Poj=1|aj=Poj=1|aj=a1Paj=a1+Poj=1|aj=a2Paj=a2(A9)

From SVO, we know that Poj=1|aj=a1=0, and Poj=1|aj=a2=23. Substituting them into (A9), we have:

(A10) Poj=1|aj=23Paj=a2(A10)

Substituting (A7), (A8), and (A10) into (A6):

(A11) Ndmdt=2iPoi=1+43zijMiPaj=a2(A11)

Note that:

(A12) ijMiPaj=a2=ziPai=a2(A12)

because everyone has been counted z times. Substituting (A12) into (A11):

(A13) Ndmdt=2i23Pai=a2Poi=1(A13)

From (A13) we know that z has been canceled out. This means the number of neighbors of each agent does not affect the dynamics of m, as long as everyone has the same number of neighbors.

Until (A13), what we have done is simply modifying the derivation of the AOI model by Tang and Chorus (Citation2019). From now on, we start to take into account obfuscation. Suppose the share of obfuscators in the population is θ 0θ1. In addition, we introduce another binary variable obi: obi=1 means agent i is an obfuscator, and obi=1 means agent i is not an obfuscator. Using this notion, we have:

(A14) Pai=a2|oi=1=1Pai=a2|oi=1,obi=1=1Pai=a2|oi=1,obi=1=0.5(A14)

and

(A15) Poi=1,obi=1=Poi=1Pobi=1=θPoi=1Poi=1,obi=1=Poi=1Pobi=1=1θPoi=1(A15)

The derivation of (A15) is based on the fact that being an obfuscator or not is independent of one’s opinion. Meanwhile, we can expand Pai=a2:

(A16) Pai=a2=Poi=1Pa2|oi=1+Poi=1,obi=1Pa2|oi=1,obi=1+Poi=1,obi=1Pa2|oi=1,obi=1(A16)

Substituting (A14) and (A15) into (A16):

(A17) Pai=a2=121θPoi=1+1+θ(A17)

By substituting (A17) into (A13), we obtain:

(A18) Ndmdt=2i131θPoi=1+1+θPoi=1(A18)

From (A7) we know Ri=2Poi=11, therefore:

(A19) m1NiRi=2NiPoi=11(A19)

According to (A19), (A18) can be rewritten as:

(A20) dmdt=θ+23m+θ3(A20)

The stable fixed point of (A20) is:

(A21) m=θ2+θ(A21)

which is equivalent to:

(A22) 1NiPoi=1=1+θ2+θ(A22)

(A22) tells us that at equilibrium, the share of vegetarians in the population is 1+θ2+θ. However, this does not mean for every realization, the system will converge to this equilibrium. Instead, the share of vegetarians averaged over all realizations of the dynamics will converge to 1+θ2+θ. This result is in line with , where fˉVeg averaged over all the realizations carried out in the simulation (which is not “all realizations” but an approximation of “all realizations”) is well approximated by 1+θ2+θ.