Hostname: page-component-8448b6f56d-tj2md Total loading time: 0 Render date: 2024-04-18T15:23:28.875Z Has data issue: false hasContentIssue false

Majority rules: how good are we at aggregating convergent opinions?

Published online by Cambridge University Press:  31 May 2019

Hugo Mercier*
Affiliation:
Institut Jean Nicod, PSL University, CNRS, ParisFrance
Olivier Morin
Affiliation:
Max Planck institute for the Science of Human History, Jena, Germany
*
*Corresponding author. Institut Jean Nicod, Département d'études cognitives, ENS, EHESS, PSL University, CNRS, ParisFrance. E-mail: hugo.mercier@gmail.com

Abstract

Mathematical models and simulations demonstrate the power of majority rules, i.e. following an opinion shared by a majority of group members. Majority opinion should be followed more when (a) the relative and absolute size of the majority grow, the members of the majority are (b) competent, and (c) benevolent, (d) the majority opinion conflicts less with our prior beliefs and (e) the members of the majority formed their opinions independently. We review the experimental literature bearing on these points. The few experiments bearing on (b) and (c) suggest that both factors are adequately taken into account. Many experiments show that (d) is also followed, with participants usually putting too much weight on their own opinion relative to that of the majority. Regarding factors (a) and (e), in contrast, the evidence is mixed: participants sometimes take into account optimally the absolute and relative size of the majority, as well as the presence of informational dependencies. In other circumstances, these factors are ignored. We suggest that an evolutionary framework can help make sense of these conflicting results by distinguishing between evolutionarily valid cues – that are readily taken into account – and non-evolutionarily valid cues – that are ignored by default.

Type
Review
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s) 2019

Social Media Summary: People are good at taking majority opinions into account, when the information is presented naturally.

Humans rely enormously on others to obtain information. In many cases, we rely not on a single individual, but on several informants – we seek advice from our colleagues, read different newspapers, ask for a second medical opinion. Sometimes these informants provide divergent advice, sometimes they concur. When several people agree, how much weight should we put on their opinion? We have known at least since Condorcet (1785) that following majority rules – adopting the opinion of the majority of the relevant group – can be a powerful heuristic. However, this heuristic should be used with caution, as it only applies under specific circumstances. The present article reviews the literature on how, and how well, humans use majority rules. Our review does not bear on how groups make collective decisions, but on how individuals take into account majority opinions. Although these two domains are linked, they are distinct: groups can make collective decisions without using majority rules, and individuals can use majority rules without any group decision having to be made, or without even belonging to any group. How individuals take into account the decisions made by others will often influence the group's collective capacity to arrive at the right decision. The rules an individual uses to aggregate the decisions of other individuals may thus be seen as an element of collective decision-making, itself an aspect of group cognition (Hutchins Reference Hutchins1996). Here, however, the important phenomenon of collective reasoning only interests us insofar as individuals avail themselves of the information generated by a collective.

First, we briefly look at mathematical and modelling work to establish the factors people should take into account when deciding how much weight to put on the majority opinion. Second, we review the literature to see if humans are able to take these factors into account: social psychology, judgment and decision-making, cultural evolution research, developmental psychology, experimental economics, science communication, and political science. Two fields will not be discussed here: research on non-human animals (for reviews, see, e.g. Claidière and Whiten Reference Claidière and Whiten2012; Haun et al. Reference Haun, Van Leeuwen and Edelson2013; Whiten Reference Whiten2019), and non-experimental research (for reasons explained below). Finally, we offer a theoretical framework that attempts to make sense of the divergent findings emerging from this review.

Because this article draws on several different studies, it might be confusing to adopt the terminology used in each to refer to the same concepts. Instead, we will rely on the following terms: the recipient is the individual who receives the information (typically the participant in experimental studies), the informants are those individuals who provide information to the participants. The information provided by informants is referred to as the informants’ opinions, and the grounding of these opinions – the pieces of information on which they are based – is referred to as their source(s).

As a final terminological note, the influence of majority rules has chiefly been studied by social psychologists under the label of conformity, distinguishing between normative conformity – aiming at ingratiation with the group one conforms to – and informational conformity – aiming at accuracy (Deutsch and Gerard Reference Deutsch and Gerard1955; for an attempt to experimentally disentangle the effects of both types of conformity, see Sowden et al. Reference Sowden, Koletsi, Lymberopoulos, Militaru, Catmur and Bird2018). It is often difficult to tell exactly how much these two types of conformity are engaged in a given task – a difficulty linked to the fact that the theoretical distinction is itself not entirely clear-cut (see Cialdini and Goldstein Reference Cialdini and Goldstein2004; Coultas and van Leeuwen Reference Coultas and van Leeuwen2015; Deutsch and Gerard Reference Deutsch and Gerard1955; Hodges Reference Hodges2015). The difficulty in distinguishing informational and normative conformity is even more acute in the case of non-experimental studies, which is the main reason why we have chosen to exclude them from the present review.

In spite of these difficulties, we should note that normative and informational conformity serve different functions: to be accepted by a group for the former, to form accurate beliefs for the latter. We can thus reasonably expect that distinct cognitive mechanisms would have evolved to fulfil each function, taking into account different cues, performing different computations, and yielding different outcomes. As a result, throughout the review, we have attempted to focus on studies that are more likely to elicit informational rather than normative conformity, as we are primarily interested in how people use majority rules to gain information. This means for instance that we have focused on experiments in which the answers are given in private (for evidence that such settings are less likely to elicit normative conformity, see Asch Reference Asch1956; Bond Reference Bond2005; Corriveau and Harris Reference Corriveau and Harris2010; Haun and Tomasello Reference Haun and Tomasello2011), the answers can be more or less accurate (in contrast with arbitrary conventions), and in some cases at least, there are incentives for accurate answers.

1. When should majority rules be followed?

The basic result supporting the rationality of following majority rules is that if each informant is more likely to be right than wrong then, as the number of informants that concur increases, their chances of agreeing on the correct answer increase.

The extension of this basic result that has played the most important role historically is the Condorcet Jury Theorem (Condorcet 1785). The Condorcet Jury Theorem was developed in the context of voting, but what matters is not the formal process of voting, but simply that several informants state their opinions, and that these opinions are dichotomous. The Condorcet Jury Theorem demonstrates that, providing some conditions are met, the chances that the majority supports the correct choice increase with the total number of informants, and converge towards 1 as the number of informants increases. As a result, the first two factors that should be taken into account when deciding whether to use majority rules are the relative and absolute size of the majority (i.e. how consensual is the majority opinion within a given group, and how large is the group in which this opinion is accepted by the majority).

Analytically, the extension of the Condorcet Jury Theorem to cases in which there are more than two opinions has raised some well-known issues. In particular voting paradoxes, in which no opinion is supported above the others, become a possibility. Still, simulations strongly suggest that, when more than two options are available, plurality rules – in which one follows the opinion supported by a plurality of informants – perform very well for a wide range of parameters (Hastie and Kameda Reference Hastie and Kameda2005).

Majority rules, understood as following the most common opinion in a group, thus have strong rational grounding. Another potential advantage of following majority rules is their simplicity (Boyd and Richerson Reference Boyd and Richerson1985). Compared with other means of information aggregation, majority rules can take a form that is computationally trivial, only requiring a counting of the opinions without any further weighing of the individual opinions. Moreover, majority rules sometimes outperform more sophisticated mechanisms of information aggregation (Hastie and Kameda Reference Hastie and Kameda2005), making majority rules one of those ‘simple heuristics that make us smart’ (Gigerenzer et al. Reference Gigerenzer and Todd1999).

For all their power, majority rules should not be blindly followed. There are well-known factors that limit their applicability. The most obvious constraint on majority rules is that the informants must, on average, be right rather than wrong. This is often framed in terms of competence. However, informants can also provide wrong information intentionally or through negligence (see, e.g. Austen-Smith and Banks Reference Austen-Smith and Banks1996). To be able to rely on majority rules, one should thus first ensure that the informants are competent and benevolent.

Another constraint on majority rules is that in many, probably most, cases, recipients of social information have a preconceived view regarding the opinion at hand. These priors should be taken into account, so that the output of majority rules should be less potent when it runs against stronger priors. This simple consequence of the fact that social influence is not the sole source of our beliefs is accepted by all theories of social learning (e.g. Asch Reference Asch1955; Festinger Reference Festinger1954). Accordingly, cultural evolutionary theorists show that purely social learning as distinct from purely individual learning is less likely to evolve when the latter is costly and uncertain owing to a fast-changing environment (Boyd and Richerson Reference Boyd, Richerson, Zentall and Galef1988). For individuals, ‘copy when uncertain’ is a sensible heuristic (Laland Reference Laland2004).

These factors – that the informants should be competent and benevolent, that the opinion should be weighted against the recipient's priors – are not specific to majority rules. The only factor that is specific to majority rules is that the opinions of the members of the majority should have been acquired independently – i.e. that there should be no informational dependencies between the opinions. Consider the case of two friends, Lana and Patricia, whom you both trust, telling you that Justin, an acquaintance of yours, is rude. How much weight you should give to your friends’ opinions depends on how independent they are. If the only reason Patricia believes Justin to be rude is because Lana told her so, then you should largely discount Patricia's opinion (not entirely, because Patricia's endorsement of Lana's opinion suggests that she has independent reasons to trust Lana in such matters; see Estlund Reference Estlund1994). Similarly, if both Patricia and Lana ground their opinions on witnessing the exact same behaviour (say, one time when Justin was very rude to a waiter), then their combined opinion does not weigh much more than their individual opinions. In contrast, if Patricia and Lana ground their opinion of Justin on non-overlapping observations, then the convergence in their opinions provides more grounds for accepting that Justin is rude (for a formal analysis of the role of causal dependencies in the Condorcet Jury Theorem, see Dietrich and Spiekermann Reference Dietrich and Spiekermann2013).

More recent modelling work has confirmed the robustness of Condorcet's original insight, for instance by looking at the trade-offs between personal and social decisions (King and Cowlishaw Reference King and Cowlishaw2007). Other work has extended the original model, for example by modelling the influence of the threshold (in terms of relative majority) on whether agents follow majority decisions (MacCoun Reference MacCoun2012). Finally, a few other studies appear to contradict the Condorcet Jury Theorem. For example, Kao and Couzin (Reference Kao and Couzin2014) have shown that, when the environment is particularly complex, small groups might be more likely to make correct decisions than larger groups – so that the absolute size of the majority would not be a sound cue anymore. However, it is likely this apparent reversal is caused by issues of informational dependency, which would make it consistent with the classical Condorcetian framework. Thus, even though we have not presented an exhaustive review of mathematical and modelling work on majority rules, the literature appears to support the Condorcetian framework, with some potential minor deviations.

To summarize, in order to aptly use majority rules, recipients should pay attention to the following factors: (a) relative size of the majority; (b) absolute size of the majority; (c) competence of the informants; (d) benevolence of the informants; and (e) degree of informational dependency between the opinion of the informants.

Before reviewing the experimental literature looking at whether people take these cues into account, we must specify that majority rules are only one of the factors deciding how much weight people give to others’ opinions: people also look at the content of the opinion and of the arguments supporting it (for reviews, see Mercier Reference Mercier2017; Sperber et al. Reference Sperber, Clément, Heintz, Mascaro, Mercier, Origgi and Wilson2010). In some cases, these other factors trump majority rules, as when the correct answer to a logical problem is defended by only one individual against many individuals agreeing on the incorrect answer. In such a situation, if the individual defending the correct answer is able to produce good arguments in its support, she will nearly always prevail against the majority (Laughlin Reference Laughlin2011; Trouche et al. Reference Trouche, Sander and Mercier2014; for slightly different problems, see Laughlin et al. Reference Laughlin, Hatch, Silver and Boh2006). However, here we focus on factors that affect how people use majority rules, not how majority rules are weighted against other, independent factors.

2. Review of the experimental literature

2.1. Absolute and relative size of the majority

Starting with the effect of absolute group size, several studies using different paradigms and participants of different ages have found that the absolute number of informants supporting an opinion can influence how much participants take this opinion into account. Even though some studies compare the influence of one vs more than one informant, we will use ‘group size’ to refer to the number of informants, even when there is only one informant. The most robust result is probably that obtained through a meta-analysis of the experiments that have collected private answers in the Crutchfield paradigm (Bond Reference Bond2005). In this paradigm, participants usually complete a non-trivial perceptual task and their answers are private, in contrast with the standard Asch paradigm (Crutchfield Reference Crutchfield1955). Participants are provided with a signal that the experimenter presents as the answers of one or more other participants (informants). In this context, participants are more likely to adopt an answer defended by more informants (Bond Reference Bond2005). This effect of absolute group size is found not only through the meta-analytical comparison of several studies with different group sizes, but within individual studies that have manipulated group size (Gerard et al. Reference Gerard, Wilhelmy and Conolley1968; McElreath et al. Reference McElreath, Lubell, Richerson, Waring, Baum, Edsten and Paciotti2005; Morgan et al. Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012; Toyokawa et al. Reference Toyokawa, Whalen and Laland2018; Wolf et al. Reference Wolf, Kurvers, Ward, Krause and Krause2013). Tanford and Penrod's (Reference Tanford and Penrod1984) meta-analysis, although it includes mostly Asch-like studies where participant choices are public, does find the expected effect of absolute majority sizes, even after controlling for the publicity of responses.

An effect of absolute group size has also been observed in young children. When presented with perceptually ambiguous items, 3- and 4-year-olds are more likely to change their mind regarding the identity of the items when confronted by three unanimous informants than by a single informant (Bernard et al. Reference Bernard, Harris, Terrier and Clément2015).

Other experiments, however, have failed to observe an effect of absolute group size on informational conformity. These failures probably stem from different factors. One factor might be that the experiment mostly engaged normative conformity. For instance, one experiment observed whether participants would engage in an unusual behaviour as they witnessed others engaging in this behaviour (Coultas Reference Coultas2004). Such an experiment is likely to tap into normative conformity: the behaviour is public, and there is no question of accuracy. This experiment found effects of relative group size, but not of absolute group size. These results are consistent with normative conformity, for which what mostly matters is whether the opinion is unanimous or not, rather than the absolute group size (see, e.g. Asch Reference Asch1955). Other factors explaining the lack of effect of absolute group size are floor and ceiling effects. As an example of ceiling effect, in an experiment the perceptual task was made very difficult by lowering the amount of attention people could give the stimuli (Campbell and Fairey Reference Campbell and Fairey1989). When participants were confronted with the opinion of a single informant, a large majority of the participants adopted it. It is not surprising then that an also high number of participants adopted an opinion defended by three informants. Conversely, Efferson et al. (Reference Efferson, Richerson, McElreath, Lubell, Edsten, Waring and Baum2007; see also Efferson et al. Reference Efferson, Lalive, Richerson, McElreath and Lubell2008, Reference Efferson, Lalive, Cacault and Kistler2016) found that participants, in a task aimed at simulating foraging decisions, did not take any heed of the answers given either by several informants or by a single informant – an example of a floor effect (possibly because their priors were too strong).

Turning to the effects of relative group size, a series of experiments, which involved different tasks with clear accuracy goals, systematically varied the proportion of informants agreeing on a given answer, from 50 to 100% (Morgan et al. Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012). The odds that the participants would adopt the answer supported by the majority increased with the relative size of the majority. Similar results have been observed in other experiments using a variety of materials – although some of these experiments might also have tapped into normative conformity (Claidière et al. Reference Claidière, Bowler and Whiten2012; Coultas Reference Coultas2004; Eriksson and Coultas Reference Eriksson and Coultas2009; Weizsäcker Reference Weizsäcker2010). Using a perceptual task (dot-discrimination), the same result was clearly observed among 6- and 7-year-olds, but not among younger children (Morgan et al. Reference Morgan, Laland and Harris2015, p. 201).

The exact shape of the relationship between the size of a majority and the propensity to follow it is a matter of debate. On the one hand, social psychologists from the Social Impact Theory school (e.g. Eriksson and Coultas Reference Eriksson and Coultas2009; Latané Reference Latané1981; Latané and Wolf Reference Latané and Wolf1981) have argued that the minority has a stronger impact, in proportion to its relative size, than the majority. This (in Eriksson and Coultas’ formulation) implies that the function relating the probability of following a given choice (assuming a binary choice between two options) to its frequency in a population of informants assumes the shape of an inverted ‘s’. In other words, majorities tend not to be followed in proportion to their size, but a little less. The opposite view has also been defended, under the name ‘conformist transmission’, also sometimes called ‘majority bias’ (Boyd and Richerson Reference Boyd and Richerson1996; Henrich and Boyd Reference Henrich and Boyd1998). In that view, individuals follow the majority choice in a biased and disproportionate way, that is to say, out of proportion with the majority's (relative) size. This time, the relation between a majority's relative size and the probability that it will be followed forms an s-shaped curve. There is currently no decisive empirical support in favour of either of these two conflicting views, as distinct from the more straightforward hypothesis of a linear or logarithmic relation (neither s-shaped nor inverted-s-shaped). Experimental results on human adults have yielded discrepant and inconclusive findings (Coultas Reference Coultas2004; Efferson et al. Reference Efferson, Lalive, Richerson, McElreath and Lubell2008; Eriksson and Coultas Reference Eriksson and Coultas2009; McElreath et al. Reference McElreath, Bell, Efferson, Lubell, Richerson and Waring2008; Morgan et al. Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012; Muthukrishna et al. Reference Muthukrishna, Morgan and Henrich2016). Studies of jury decisions in US trials (although not comparable with controlled experiments on many grounds) appear to support a model of influence where relative majority size impacts social influence in a non-linear fashion, minorities having an influence greater than their numerical weight, with no threshold effect at 50%, in violation of conformist transmission (Tanford and Penrod Reference Tanford and Penrod1984).

In the advice-taking literature, participants are typically asked to take into account numerical estimates – for instance how many calories there are in an apple – which means that few informants would agree on any single estimate. As a result, researchers rely on a different measure of convergence: the degree of agreement among informants, rather than how many informants agree on the exact same estimate. It has been repeatedly observed that high agreement among several informants makes participants more likely to move their own estimates towards that of the informants (Budescu et al. Reference Budescu, Rantilla, Yu and Karelitz2003; Budescu and Yu Reference Budescu and Yu2007; Harries et al. Reference Harries, Yaniv and Harvey2004; Yaniv Reference Yaniv1997; Mannes Reference Mannes2009; see also, in the context of herding in finance, Barron and Stuerke Reference Barron and Stuerke1998).

Instead of directly providing information from several informants, other experiments have provided numerical data regarding the majority size – say, 75% of citizens are in favour of this policy and 25% disapprove. Most notably, research conducted in political science has studied the effects of presenting polling data on political attitudes. These experiments do not reveal a consistent effect of majority information. Participants sometimes veer towards the majority (bandwagon effect), sometimes towards the minority (underdog effect) (for review, see Mutz Reference Mutz1998; for similar research in social psychology, see, e.g. Petty and Cacioppo Reference Petty and Cacioppo1986). Many experiments have provided majority information along with arguments supporting either the majority opinion or the minority opinion, or both. These experiments have also yielded conflicting results, revealing several moderating variables such as argument quality and the participants’ prior beliefs (for review, see Martin and Hewstone Reference Martin and Hewstone2008). In some cases, argument quality completely trumps majority information, so that a strong argument provided by a minority is very likely to be accepted, whereas a weak argument provided by a majority is nearly never accepted (Claidière, Trouche, and Mercier Reference Claidière, Trouche and Mercier2017; Laughlin Reference Laughlin2011; Trouche et al. Reference Trouche, Sander and Mercier2014).

A recent study has looked at whether participants have some understanding of the advantages of voting demonstrated by the Condorcet Jury Theorem (Mercier, Dockendorff, and Schwartzberg, Reference Mercier, Dockendorff and Schwartzbergsubmitted). Participants are asked to estimate the chances that, if a majority of members of an assembly voted for one option (out of two), the option selected was the best. Several relevant factors (size of the majority, competence of the members, etc.) are manipulated, and the results show that, although the participants are able to take some relevant factors into account, they consistently underestimate the odds that the majority selects the best option.

Finally, several studies in science communication have tested whether providing participants with information about the degree of scientific consensus affects their opinions on topics that have proved controversial among the general public (climate change, vaccination and GMOs). In these studies, consensus information is presented in different ways: the percentage of scientists who agree is provided (Deryugina and Shurchkov Reference Deryugina and Shurchkov2016; Dixon Reference Dixon2016; Dixon et al. Reference Dixon, Hmielowski and Ma2017; Kerr and Wilson Reference Kerr and Wilson2018; Landrum et al. Reference Landrum, Hallman and Jamieson2018; Lewandowsky et al. Reference Lewandowsky, Gignac and Vaughan2013), but it can be accompanied by other modes of presentation, such as pie charts, pictures of groups of scientists or metaphors (van der Linden et al. Reference van der Linden, Leiserowitz, Feinberg and Maibach2014, Reference van der Linden, Clarke and Maibach2015a, Reference van der Linden, Leiserowitz, Feinberg and Maibachb). Presenting consensus information is sometimes effective in moving participants’ opinions towards the consensus (Deryugina and Shurchkov Reference Deryugina and Shurchkov2016; Dixon Reference Dixon2016; Kerr and Wilson Reference Kerr and Wilson2018; Lewandowsky et al. Reference Lewandowsky, Gignac and Vaughan2013; van der Linden et al. Reference van der Linden, Leiserowitz, Feinberg and Maibach2014, Reference van der Linden, Clarke and Maibach2015a, Reference van der Linden, Leiserowitz, Feinberg and Maibachb), but not always (Dixon et al. Reference Dixon, Hmielowski and Ma2017; Landrum et al. Reference Landrum, Hallman and Jamieson2018).

2.2. Competence and benevolence of the informants, prior beliefs of the recipient

Here we focus on factors that are not specific to informational conformity, since they consist of individual characteristics, whether they are informant characteristics (competence and benevolence) or recipient characteristics (prior beliefs). These factors have been shown to play an important role in how people evaluate opinions coming from a single informant (Bonaccio and Dalal Reference Bonaccio and Dalal2006; Petty and Wegener Reference Petty, Wegener, Gilbert, Fiske and Lindzey1998; Yaniv Reference Yaniv2004; for review, see Mercier Reference Mercier2017). In contrast, few experiments have directly tested the effects of informant characteristics on informational conformity. Andersson et al. (Reference Andersson, Hedesström and Gärling2014) found that asking participants to focus on the accuracy of several informants who happened to have been inaccurate led participants to appropriately discount the opinion of these informants. Bernard, Proust, and Clément (Reference Bernard, Harris, Terrier and Clément2015) found that 6-year-olds – but not 4- or 5-year-olds – adequately discounted convergent opinions when the informants had been inaccurate in the past. Efferson et al. (Reference Efferson, Lalive, Cacault and Kistler2016) found that participants were more likely to copy convergent informants that were faced with the same task and the same payoffs as themselves. Still, in spite of these limited data, we should expect that informant characteristics which are known to be taken into account when processing information from single informants should keep playing their role when processing information from multiple informants.

Several experiments have shown that the prior beliefs of a recipient modulate the way she uses majority rules. It has been consistently found that, the less participants are sure of their own opinions, the more likely they are to adopt the majority opinion. This is true of comparisons across different paradigms: for instance, when answers are private, the group opinion is nearly entirely neglected in the Asch paradigm, while the group opinion has a strong influence in more difficult perceptual tasks (Bond Reference Bond2005). This result has also been obtained by experimental manipulations of task difficulty, for a variety of tasks (Baron et al. Reference Baron, Vandello and Brunsman1996; Campbell and Fairey Reference Campbell and Fairey1989; Morgan et al. Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012; Toyokawa et al. Reference Toyokawa, Whalen and Laland2018). Similarly, participants who form their opinions on the basis of more variable information, and thus form less reliable opinions, are more likely to adopt the opinion defended by a majority of informants (Andersson et al. Reference Andersson, Hedesström and Gärling2014; McElreath et al. Reference McElreath, Lubell, Richerson, Waring, Baum, Edsten and Paciotti2005; see also Antenore et al. Reference Antenore, Leone, Panconesi and Terolli2018). Children as young as 3 years of age have been found to be influenced by convergent informants to a greater extent when they face a difficult perceptual task, relative to an easier version of the task (Bernard et al. Reference Bernard, Harris, Terrier and Clément2015; see also Morgan et al. Reference Morgan, Laland and Harris2015).

The influence of recipients’ prior beliefs is, on the whole, positive: it makes for a finer-grained discrimination and allows for improved performance (see, e.g. Morgan et al. Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012, Reference Morgan, Laland and Harris2015). However, participants often put too much weight on their prior beliefs, displaying egocentric discounting. Studies of social influence often find that participants holding the right answer to a question rationally exhibit higher confidence and lower susceptibility to social influence (see, e.g. Jayles et al. Reference Jayles, Kim, Escobedo, Cezera, Blanchet, Kameda and Theraulaz2017). Egocentric discounting is different: it is the tendency of recipients to put more weight on their own opinion, relative to the opinion of another informant, even when they have no reason to believe they are more competent than the informant (Trouche et al. Reference Trouche, Johansson, Hall and Mercier2018; Yaniv and Kleinberger Reference Yaniv and Kleinberger2000). Egocentric discounting is apparent in various economic games involving convergent social information, where participants spurn the information manifested in other people's choices, even though it would be beneficial for them to take it into account (Efferson et al. Reference Efferson, Richerson, McElreath, Lubell, Edsten, Waring and Baum2007, Reference Efferson, Lalive, Richerson, McElreath and Lubell2008; Mannes Reference Mannes2009).

The clearest example is perhaps provided by the experimental literature on information cascades (Anderson and Holt Reference Anderson and Holt1997). In information cascade experiments, participants have to decide which colour is more common in the balls drawn from an urn. Participants base their opinion on a single draw that only they have access to, as well as on the actions exhibited by the sequence of all previous participants. The standard models for this experiment assumes agents to be competent and not intent on misleading others, an assumption that holds in laboratory settings (Bikhchandani et al. Reference Bikhchandani, Hirshleifer and Welch1998; Goeree et al. Reference Goeree, Palfrey, Rogers and McKelvey2007).

In the Bayes–Nash model of information cascades, the way that participants should weigh the information given by other participants’ behaviour is straightforward. If the choices of the last two players who made a move before me do not coincide, what is optimal is to follow my private information. If the choices of the last two participants coincide, on the other hand, the optimal choice is to copy them, regardless of my private information (the colour of the ball I see). In the latter case, it is possible for the wrong opinion to take hold in a group, if the first few participants’ draws are misleading (i.e. they drew the ball from the rare colour). Other participants should then bet on the rare colour, even if they themselves have drawn a ball from the common colour, potentially creating a negative information cascade. However, the opposite (a positive information cascade where players choose the common colour) is much more likely. Whether the cascade is positive or negative, however, a rational participant, not knowing whether past participants were right or wrong, should follow it (in this model).

One important finding from the experimental literature is that both negative and positive cascades are much less frequent than this Bayes–Nash model implies (Goeree et al. Reference Goeree, Palfrey, Rogers and McKelvey2007). This deviation from the normative model occurs for two reasons. First, there is a substantial element of randomness in the participants’ choices. Second, participants overweigh their private information, instead of following the informants, in the situation where the private signal they receive contradicts the choices of the previous participants (Goeree et al. Reference Goeree, Palfrey, Rogers and McKelvey2007; Weizsäcker Reference Weizsäcker2010; Ziegelmeyer et al. Reference Ziegelmeyer, March and Krügel2013). Egocentric discounting is not detrimental to participants when it breaks a negative cascade, but it is when it makes them depart from a positive one (which is to say, most of the time).

Information cascades, in conjunction with the other experimental studies reviewed above, demonstrate that recipients are able to weigh informational conformity as a function of various informant characteristics and of their prior beliefs. This weighing likely improves performance. The main bias in this weighing is a tendency for recipients to put too much weight on their prior beliefs, relative to the convergent opinion of a small number of informants.

2.3. Informational dependencies

Informational dependencies occur when the opinions of several informants are confounded. This can occur because the informants’ opinions have partially or entirely identical sources (common sources), because some informants had contact with one of the other informants’ opinions before forming their own (contact), or because they shared an extrinsic motivation to express the same opinion (external pressure). An extreme case of dependency would occur when several different opinions find their origins in one single individual (see Morgan et al. Reference Morgan, Rendell, Ehn, Hoppitt and Laland2012). Informational dependencies can be explored experimentally in two ways. The first is by providing participants with general cues of dependency: correlations between the informants that may or may not indicate common sources, contact or external pressure, without giving any direct evidence of either. The second is by providing specific cues of correlation, contact or external pressure. The results differ depending on the cues’ specificity.

General cues of dependency

A number of studies have examined how participants weigh the opinions offered by several informants as a function of how correlated these informants’ opinions are or have been in the past. Information about correlation has been presented in different ways: (a) as an explicit correlation (e.g. some ‘subjects were told that the correlation among forecast errors was approximately +0.8 for all forecaster pairs’; Maines Reference Maines1990, p. 36; see also, Kroll et al. Reference Kroll, Levy and Rapoport1988); (b) implicitly – the opinions happened to be based on more or less overlapping information (a fact that was unknown to the participants and could only be inferred through the distribution of the opinions) (Budescu and Yu Reference Budescu and Yu2007); (c) as a list of previous opinions from the same informants, opinions which were manipulated to be either highly correlated or largely uncorrelated (Andersson et al. Reference Andersson, Hedesström and Gärling2014; Maines Reference Maines1996; Votruba and Kwan Reference Votruba and Kwan2015); and (d) as information about how the informants were selected, in particular telling participants that some informants were selected because they were likely to agree with one another (Yaniv et al. Reference Yaniv, Choshen-Hillel and Milyavsky2009). These methodologies yield convergent results: participants robustly fail to take correlations between the informants’ opinions into account (for similar failures regarding non-social sources of information, see Slovic Reference Slovic1966; Soll Reference Soll1999).

Specific cues of common source

Whalen et al. (Reference Whalen, Griffiths and Buchsbaum2018) show that participants are more likely to trust the opinion of three informants (over their own private information) when the three informants have access to one independent piece of evidence each, rather than to one piece of evidence common to all three. Note, though, that participants’ endorsement of the informants’ opinion, in the latter case, is higher than it should normatively be, consistent with the ‘multiple informants’ effect (Harkins and Petty Reference Harkins and Petty1981). Relatedly, participants discount information from a single informant when this informant bases her opinion on the same information as the participants, according to Gonzalez (Reference Gonzalez1994).

Mercier and Miton (Reference Mercier and Miton2019) found similar results whether the common source was hearsay (i.e. the opinions were based on hearsay from the same source) or perception (i.e. the opinions were based on having perceived the same event). One study, in contrast, exposed participants to direct cues of common sources, with no effect. The participants were provided with several newspapers articles, issued by distinct newspapers, that either all relied on the same expert or on different experts (Yousif et al. Reference Yousif, Aboody and Keil2018). Participants did not put more weight on the opinion defended by different experts (possibly because they interpreted the endorsement of different newspapers as a sign of greater expertise; see Estlund Reference Estlund1994).

Specific cues of external pressure

In a study by Conway and Schaller (Reference Conway III and Schaller2005), participants were made to play the part of an executive in a firm who had the final say in an important purchase decision. In the control condition, each participant witnessed all their colleagues choosing option A over option B independently, whereas in the experimental condition, the colleagues were all ordered to choose option A by their common boss. Participants were more likely to choose option B in the latter case, suggesting that they discounted the opinions that reflected external pressure. Similarly, Mercier and Miton (Reference Mercier, Ballantyne and Dunning2019) found that participants put less weight on the convergent opinions of informants when the informants shared the same motivation in expressing the opinion (i.e. to say something positive about a restaurant owned by a friend).

Specific cues of contact

Most experimental manipulations of informational dependencies focus on contact: the informants’ exposure to the other informants’ opinions. These experiments can be sorted according to the strength of the evidence, which can be weak if it leaves open the possibility that the informants formed their opinions on the basis of independent evidence, or strong if most informants clearly derived their opinions from hearsay from another informant only.

Weak cues of contact

In one experiment, participants were provided with the opinions of several informants. In one condition, participants were told that each of the informants had seen the other informants’ respective opinions before forming their own final opinion. In the other condition, participants were told that the informants had had no access to the other informants’ opinions before forming their own. Participants put more weight on the opinions of the informants in the latter condition (Bloomfield and Hales Reference Bloomfield and Hales2009). In another experiment, participants were provided with the opinions of two informants, and they were told either that the opinions had been formed independently or that one informant had seen the other's opinion before forming hers. Participants were less influenced by the second informant's opinion when this informant had seen the opinion of the first informant (Keshk Reference Keshk2012). However, this discounting was only observed when participants were motivated to disregard the informants’ opinion. Relatedly, Wilder (Reference Wilder1977) showed that four multiple informants were less persuasive when they were described as one group of four, rather than two groups of two, and even more persuasive when they were described as four individuals.

Harkins and Petty (Reference Harkins and Petty1987) presented students with arguments about the organization of an academic exam. In one condition, the informants providing the arguments were presented as three students reflecting independently. In another condition, the arguments were described as the work of a committee formed by three students. Participants were less persuaded by the committee's arguments than by the three individual students. More precisely, the committee persuaded them no more than a single student did. Lopes et al. (Reference Lopes, Vala and Garcia-Marques2007) likewise show that a group's opinion is considered more valid when its members are described as having heterogeneous information. Hess and Hagen (Reference Hess and Hagen2006) found that a piece of gossip was judged more believable when it came from an independent informant, as opposed to someone who merely reported and endorsed the gossip, or to an independent witness with friendship ties with a concerned party. Recently, Einav (Reference Einav2017) asked children (5- to 9-year-olds) and adults to decide between information offered by two consensual groups. The members of one of the two groups had had access to one another's answers, so that their answers probably suffered from informational dependencies. There was a developmental shift in how the opinions of these two groups were weighted, with younger children favouring the group whose members had seen the other members’ answers, while older children and adults favoured the group whose members had not had such access. This result thus replicates the findings above, and shows their developmental limits.

Strong cues of contact

One experiment (with young children) makes it clear to the participants that most informants in a group formed their opinion by contact with other opinions alone. Participants were provided with the contradictory opinions of two groups of informants. Informants from the first group formed their opinions independently – they all had perceptual access to the relevant information. In the second group, one informant had perceptual access to the relevant information, and communicated her opinion to the second informant, who communicated it to the third, etc. When the two groups were of the same size, both adults and 4-year-olds were more likely to endorse the opinion of the first group (Hu et al. Reference Hu, Whalen, Buchsbaum, Griffiths and Xu2015). However, when the first group comprised four members and the second group six, only adults kept being more likely to endorse the opinion of the first group.

Evidence from information cascades

The experimental literature on information cascades (e.g. Anderson and Holt Reference Anderson and Holt1997) could be taken as showing that participants do not take into account informational dependencies in a felicitous way. To understand why, we should keep in mind a strange property of the Bayes–Nash model of cascade formation (Bikhchandani et al. 1992; Goeree et al. Reference Goeree, Palfrey, Rogers and McKelvey2007). As soon as two consecutive players make an identical move, all of the participants following them should copy that move, irrespective of their private information. Whether the cascade has been followed for two steps or 150 should not matter to participants. Once a chain is started, the copiers that make up the chain should take no account of their own private information, and for this reason their actions are strictly uninformative – they merely reflect the information of the first two players who started the cascade. Participants do not behave as expected by this model, however: they are much more likely to follow a sequence of identical moves when that sequence has been going on for a long time (Kübler and Weizsäcker Reference Kübler and Weizsäcker2005; Weizsäcker Reference Weizsäcker2010).

From the standard model's point of view, this demonstrates a lack of understanding of informational dependencies. This, however, is not a valid argument, since the Bayes–Nash model is inaccurate on this point. Participants show egocentric discounting (they sometimes fail to follow the cascade when it contradicts their private information). Because of this, a cascade that has lasted for a great number of steps does reflect substantial information concerning the state of the world, and should be given greater weight than a shorter one. Given this, the way participants choose to follow others is close to optimal, if we disregard their bias for egocentric discounting. If participants were not properly sensitive to informational dependencies, they would be prone to following the crowd against their own private information, mistakenly thinking the crowd to be better informed than them in cases where it is not. Participants overwhelmingly avoid this bias. According to Weizsäcker's (Reference Weizsäcker2010) meta-analysis, participants behave in a near-optimal fashion in such cases, reaping 75% of the possible rewards (while a perfectly optimal player would reap 76%). This performance stands in sharp contrast with the performance exhibited when it is optimal for participants to contradict their own information, in which case egocentric discounting leads to a loss of rewards.

Overall, specific cues indicating common source, contact or external pressure, are taken into account to a much greater extent than general cues that indicate possible dependencies through correlations, without pointing at a specific cause. Specific cues of common source, contact, or external pressure, seem more efficient when they are strong enough to exclude other causes (for instance, when it is obvious that the informants could only have formed their opinions through contact with other informants).

3. Conclusion: making sense of divergent findings

At first sight, this review might appear to yield confusing results. In some cases, participants appropriately use a given factor to decide how much they should rely on majority rules. In other cases, they completely overlook a factor. Moreover, the same factor – e.g. whether there are informational dependencies or not – is sometimes taken into account, sometimes ignored.

We suggest that the framework of the evolution of communication (Maynard Smith and Harper Reference Maynard Smith and Harper2003) can help make sense of these findings. Communication is known to be a risky endeavour, as recipients can be lied to or otherwise manipulated. Communication is particularly risky for humans, given its importance in their ecology. As a result, it has been suggested that humans have been endowed with cognitive mechanisms whose function is to evaluate communicated information (Mercier Reference Mercier2017; Sperber et al. Reference Sperber, Clément, Heintz, Mascaro, Mercier, Origgi and Wilson2010). These mechanisms of epistemic vigilance would take into account a variety of cues and influence which messages we accept and which we reject. Some of these cues pertain to the competence and benevolence of informants, others to the content of the message, in particular in relation with our prior beliefs. A wealth of evidence shows that humans, from a very young age (Clément Reference Clément2010; Harris Reference Harris2012; Harris and Lane Reference Harris and Lane2014), are able to appropriately take many such cues into account, so that plausible messages coming from competent and benevolent informants are more likely to be accepted (for review, see Mercier Reference Mercier2017).

The studies reviewed above suggest that these epistemic vigilance mechanisms are also applied to majority rules, so that the opinion of the majority is granted more weight when the informants are competent and benevolent, and when the majority opinion conflicts less with the recipient's prior beliefs. Moreover, an evolutionary framework can account for the main deviation from apparently optimal behaviour observed in this area: the tendency to put too much weight on the recipient's priors relative to communicated information (egocentric discounting). If we consider the uncertainty about informant benevolence, and the relative costs associated with trusting too much vs not trusting enough, a degree of egocentric discounting appears sensible enough (see, Sperber et al. Reference Sperber, Clément, Heintz, Mascaro, Mercier, Origgi and Wilson2010; Trouche et al. Reference Trouche, Johansson, Hall and Mercier2018).

At least as much as communication, the use of majority rules is probably evolutionarily ancient: many animals rely on such rules, for instance when making foraging decisions (Conradt and List Reference Conradt and List2009; Conradt and Roper Reference Conradt and Roper2003). Given the importance of group decision-making (Boehm et al. Reference Boehm, Antweiler, Eibl-Eibesfeldt, Kent, Knauft, Mithen and Wilson1996) and, more generally, of cooperation and communication in humans, we should expect them to be equipped with cognitive mechanisms dedicated to majority rules. These mechanisms would also pay attention to a variety of cues in order to determine how much weight to grant majority opinions.

However, not all cues are created equal. In particular, we can only expect cognitive mechanisms to have evolved in response to cues that were accessible and reliable in the environment that shaped these cognitive mechanisms – evolutionarily valid cues (e.g. Barkow et al. Reference Barkow, Cosmides and Tooby1992; Cosmides and Tooby Reference Cosmides, Tooby, Hirschfeld and Gelman1994; Sperber Reference Sperber and Dupoux2001). The concept of evolutionarily valid cues shares some traits with that of ecological validity (see, e.g., Gigerenzer Reference Gigerenzer, Koehler and Harvey2007; Gigerenzer et al. Reference Gigerenzer and Todd1999), but it stresses the importance of past, in addition to current, environments.

To the extent that the cognitive mechanisms we rely on to deal with majority rules are somewhat well adapted, we should expect them to primarily rely on evolutionary valid cues. This does not mean that we should be unable to learn to rely on other cues: after all, humans are typically endowed with learning mechanisms rather than purely innate mechanisms. Still, efficient learning mechanisms rely on strong built-in biases (e.g. Carruthers Reference Carruthers2006; Gallistel and Gibbon Reference Gallistel and Gibbon2000). Thus, even though humans can surely learn how to use new cues when dealing with majority rules, we should expect some cues to be more readily taken into account.

Even if we can only offer approximations for the environment that shaped the cognitive mechanisms dealing with majority rules, some plausible assumptions help make sense of which cues are more easily taken into account. Regarding the size of the majority, an evolutionarily valid cue would have been witnessing individuals offering different opinions. In contrast, written numbers, and in particular written statistical indicators (e.g. ‘75% of subjects chose X’) are a recent cultural invention, making them a potentially non-evolutionarily valid cue (e.g. Chrisomalis Reference Chrisomalis2010; Dehaene Reference Dehaene1999). The distinction between evolutionarily valid and non-evolutionarily valid cue is reminiscent of the distinction between decisions from experience (here, seeing the individual opinions of different informants) and decisions from description (here, being provided with a numerical description of the majority) (Hertwig and Erev Reference Hertwig and Erev2009).

The distinction between evolutionarily valid cues and non-evolutionarily valid cues fits the results reviewed in Section 2.1. Participants, including young children, appropriately weigh majority opinion as a function of the absolute and relative size of the majority when they can see the opinions of individual informants. In contrast, participants do not systematically follow majority rules when the information is presented in written numerical format (e.g. ‘75% of the people say’). Moreover, participants have no abstract understanding of the benefits of majority rules – in the same way as people have a good grasp of intuitive physics, but not of theoretical physics (see Mercier Reference Mercier, Ballantyne and Dunning2019).

Turning to informational dependencies, some relevant cues are omnipresent in all known human societies, and these cues can be assumed to have been present in the relevant evolutionary timeframe. In particular, recipients would have had some information about the source of the informants’ opinions. Recipients are sometimes able to infer this information – e.g. having seen Patricia talk to Lana before she talks to you. More often, information about the source of their opinions is volunteered by informants. All languages contain tools with which informants can specify the source of their opinions – evidentiality markers (see Aikhenvald Reference Aikhenvald2004, p. 10ff). Some evidentiality markers are explicit (‘Lana told me’), others implicit. Many languages have evidentials – gramaticalized, mandatory evidentiality markers – so that every assertion has to specify the source of the opinion offered (Aikhenvald Reference Aikhenvald2004; Lazard Reference Lazard2001).

Because people would often have information about the actual or potential sources of informants’ opinions, specific cues of common source, contact or external pressure should be evolutionarily valid, and thus appropriately taken into account. As we saw above (Section 2.3), this seems to be the case. In contrast, general cues of dependency – e.g. a correlation coefficient – should not be evolutionarily valid, and thus they should be largely ignored. Again, this seems to be the case.

That some cues are presumed not to be evolutionarily valid does not mean that they cannot be learned. People can learn to use non-evolutionarily valid cues if they are sufficiently motivated. For instance, observational evidence suggests that financial experts are able to take informational dependencies into account, even when they are not presented in a evolutionarily valid manner (Cooper et al. Reference Cooper, Day and Lewis2001; Jegadeesh and Kim Reference Jegadeesh and Kim2010; Kim and Pantzalis Reference Kim and Pantzalis2003; for another kind of expertise, more informal, see Zhou and Guo Reference Zhou and Guo2016).

The attempted exhaustiveness of the present review was made possible by restrictions on the range of evidence we considered. First, we focused on situations likely to elicit informational, rather than normative conformity. Normative conformity might dictate very different behaviours than informational conformity (as in, for instance, the Asch conformity experiments). While people are influenced by normative concerns in the public expression of their opinion or behaviour, it is often difficult to tell whether they are also following majority rules properly in the formation of their private opinions. This difficulty is particularly acute in the case of non-experimental studies, which is the main reason why we had decided to exclude them. A potential unfortunate side-effect of this choice was the focus on standard experimental participants, when non-experimental studies might cover a broader diversity of people. Finally, we only looked at how individuals use majority rules to form opinions, not how groups use such rules to make decisions. Although, as we have argued at the outset, the two issues are largely distinct, there is still potential for considerable interplay, which should be the topic of future studies.

In spite of these limitations, the present article offers, to the best of our knowledge, the first integration of the various fields that have conducted experimental studies relevant to the use of majority rules by humans. We have attempted to make sense of these results, first, by organizing them as a function of the various factors that normative models say should be taken into account, and second, by using an evolutionary framework to explain why some cues to the validity of majority rules are sometimes properly processed, and sometimes ignored by participants.

Future work should offer more direct tests of these hypotheses, in particular by contrasting evolutionarily valid and non-evolutionarily valid cues. We predict that evolutionarily valid cues will be taken into account in a way that better fits mathematical models of when majority rules should be taken into account (for some preliminary evidence, see Mercier, Majima, Claidière, and Léone, Reference Mercier, Majima, Claidière and Léonesubmitted). This does not mean that non-evolutionarily valid cues will always be ignored, as they can be processed by other mechanisms, merely that they will be taken into account less appropriately. Another prediction of the present evolutionary framework is that there should be less cross-cultural and developmental variation in how evolutionarily valid cues are taken into account. Taking evolutionarily valid cues into account should not depend on schooling or explicit learning. In contrast, making sound use of non-evolutionarily valid cues should often require explicit teaching. Evolutionary approaches can drive more research into similarities and differences across cultures (Apicella and Barrett Reference Apicella and Barrett2016), a move that would greatly benefit research on the use of majority rules (for some examples of cross-cultural research, see Corriveau and Harris Reference Corriveau and Harris2010; Van Leeuwen et al. Reference Van Leeuwen, Cohen, Collier-Baker, Rapold, Schäfer, Schütte and Haun2018). Finally, our evolutionary approach could help better distinguish between normative and informational conformity. As mentioned earlier, both types of conformity serve different functions, which are plausibly fulfilled by different cognitive mechanisms, taking different cues into account. An evolutionary perspective should make predictions regarding which cues to normative conformity are evolutionarily valid. These predictions could then be evaluated against the existing literature and in novel experiments comparing for example the weight granted evolutionarily valid cues with informational vs. normative conformity in different contexts.

Author ORCIDs

Hugo Mercier, 0000-0002-0575-7913 and Olivier Morin, 0000-0002-6216-1307.

Author contributions

Both authors took part in the literature review and in writing the article.

Financial support

H.M. acknowledges the support from a grant from the Agence Nationale de la Rercherche EUR FrontCog ANR-17-EURE-0017.

Conflict of interest

The authors report no conflict of interest.

References

Aikhenvald, AY (2004) Evidentiality. Oxford: Oxford University Press.Google Scholar
Anderson, LR and Holt, CA (1997) Information cascades in the laboratory. The American Economic Review, 87(5), 847862.Google Scholar
Andersson, M, Hedesström, M and Gärling, T (2014) A social-psychological perspective on herding in stock markets. Journal of Behavioral Finance, 15(3), 226234.Google Scholar
Antenore, M, Leone, G, Panconesi, A and Terolli, E (2018) Together we buy, alone i quit: some experimental studies of online persuaders. Proceedings of the 1st International Conference on Digital Tools and Uses Congress, Vol. 2. New York: ACM.Google Scholar
Apicella, CL and Barrett, HC (2016) Cross-cultural evolutionary psychology. Current Opinion in Psychology 7, 9297.Google Scholar
Asch, SE (1955) Opinions and social pressure. Scientific American 193(5), 3135.Google Scholar
Asch, SE (1956) Studies of independence and conformity: a minority of one against a unanimous majority. Psychological Monographs 70(9), 170.Google Scholar
Austen-Smith, D and Banks, JS (1996) Information aggregation, rationality, and the Condorcet jury theorem. American Political Science Review 90(01), 3445.Google Scholar
Barkow, JH, Cosmides, L and Tooby, J (1992) The Adapted Mind. Oxford: Oxford University Press.Google Scholar
Baron, RS, Vandello, JA and Brunsman, B (1996) The forgotten variable in conformity research: impact of task importance on social influence. Journal of Personality and Social Psychology 71(5), 915.Google Scholar
Barron, OE and Stuerke, PS (1998) Dispersion in analysts’ earnings forecasts as a measure of uncertainty. Journal of Accounting, Auditing and Finance 13(3), 245270.Google Scholar
Bernard, S, Harris, P, Terrier, N and Clément, F (2015) Children weigh the number of informants and perceptual uncertainty when identifying objects. Journal of Experimental Child Psychology 136, 7081.Google Scholar
Bernard, S, Proust, J and Clément, F (2015) Four-to six-year-old children's sensitivity to reliability versus consensus in the endorsement of object labels. Child Development.Google Scholar
Bikhchandani, S, Hirshleifer, D and Welch, I (1998) Learning from the behavior of others: conformity, fads, and informational cascades. The Journal of Economic Perspectives 12(3), 151170.Google Scholar
Bloomfield, R and Hales, J (2009) An experimental investigation of the positive and negative effects of mutual observation. The Accounting Review 84(2), 331354.Google Scholar
Boehm, C, Antweiler, C, Eibl-Eibesfeldt, I, Kent, S, Knauft, BM, Mithen, S, … Wilson, DS (1996) Emergency decisions, cultural-selection mechanics, and group selection [and Comments and Reply]. Current Anthropology 37(5), 763793.Google Scholar
Bonaccio, S and Dalal, RS (2006) Advice taking and decision-making: an integrative literature review, and implications for the organizational sciences. Organizational Behavior and Human Decision Processes, 101(2), 127151.Google Scholar
Bond, R (2005) Group size and conformity. Group Processes and Intergroup Relations 8(4), 331354.Google Scholar
Boyd, R and Richerson, PJ (1985) Culture and the Evolutionary Process. Chicago, IL: Chicago University Press.Google Scholar
Boyd, R and Richerson, PJ (1988) An evolutionary model of social learning: the effects of spatial and temporal variation. In Zentall, TR and Galef, BG (eds), Social Learning: Psychological and Biological Perspectives (pp. 2948) Hillsdale, NJ: Lawrence Erlbaum.Google Scholar
Boyd, R and Richerson, PJ (1996) Why culture is common, but cultural evolution is rare. Proceedings of the British Academy 88, 7793.Google Scholar
Budescu, DV and Yu, H-T (2007) Aggregation of opinions based on correlated cues and advisors. Journal of Behavioral Decision Making 20(2), 153177.Google Scholar
Budescu, DV, Rantilla, AK, Yu, H-T, and Karelitz, TM (2003) The effects of asymmetry among advisors on the aggregation of their opinions. Organizational Behavior and Human Decision Processes 90(1), 178194.Google Scholar
Campbell, JD and Fairey, PJ (1989) Informational and normative routes to conformity: the effect of faction size as a function of norm extremity and attention to the stimulus. Journal of Personality and Social Psychology 57(3), 457.Google Scholar
Carruthers, P (2006) The Architecture of the Mind: Massive Modularity and the Flexibility of Thought. Oxford: Oxford University Press.Google Scholar
Chrisomalis, S (2010) Numerical Notation: A Comparative History. New York: Cambridge University Press.Google Scholar
Cialdini, RB and Goldstein, NJ (2004) Social influence: compliance and conformity. Annual Review of Psychology 55, 591621.Google Scholar
Claidière, N and Whiten, A (2012) Integrating the study of conformity and culture in humans and nonhuman animals. Psychological Bulletin 138(1), 126.Google Scholar
Claidière, N, Bowler, M and Whiten, A (2012) Evidence for weak or linear conformity but not for hyper-conformity in an everyday social learning context. PloS One 7(2), e30970.Google Scholar
Claidière, N, Trouche, E and Mercier, H (2017) Argumentation and the diffusion of counter-intuitive beliefs. Journal of Experimental Psychology: General 146(7), 10521066.Google Scholar
Clément, F (2010) To Trust or not to trust? Children's social epistemology. Review of Philosophy and Psychology, 1(4), 119.Google Scholar
Condorcet (1785) Essai sur l'application de l'analyse à la probabilité des décisions rendues à la pluralité des voix. Paris: L'imprimerie royale.Google Scholar
Conradt, L and List, C (2009) Group decisions in humans and animals: a survey. Philosophical Transactions of the Royal Society of London B: Biological Sciences 364(1518), 719742.Google Scholar
Conradt, L and Roper, TJ (2003) Group decision-making in animals. Nature 421(6919), 155158.Google Scholar
Conway III, LG and Schaller, M (2005) When authorities’ commands backfire: attributions about consensus and effects on deviant decision making. Journal of Personality and Social Psychology 89(3), 311.Google Scholar
Cooper, RA, Day, TE and Lewis, CM (2001) Following the leader: a study of individual analysts’ earnings forecasts. Journal of Financial Economics 61(3), 383416.Google Scholar
Corriveau, KH and Harris, PL (2010) Preschoolers (sometimes) defer to the majority in making simple perceptual judgments. Developmental Psychology 46(2), 437.Google Scholar
Cosmides, L and Tooby, J (1994) Origins of domain specificity: the evolution of functional organization. In Hirschfeld, LA and Gelman, SA (eds), Mapping the Mind: Domain Specificity in Cognition and Culture (pp. 85116). Cambridge: Cambridge University Press.Google Scholar
Coultas, JC (2004) When in Rome … An evolutionary perspective on conformity. Group Processes and Intergroup Relations 7(4), 317331.Google Scholar
Coultas, JC and van Leeuwen, EJ (2015) Conformity: Definitions, types, and evolutionary grounding. In V Zeigler-Hill, LLM Welling & TK Shackelford (eds.), Evolutionary Perspectives on Social Psychology (pp. 189–202). London: Springer. Retrieved from http://link.springer.com/chapter/10.1007/978-3-319-12697-5_15Google Scholar
Crutchfield, RS (1955) Conformity and character. American Psychologist 10(5), 191.Google Scholar
Dehaene, S (1999) The number sense: How the mind creates mathematics. Oxford: Oxford University Press.Google Scholar
Deryugina, T and Shurchkov, O (2016) The effect of information provision on public consensus about climate change. PloS One 11(4), e0151469.Google Scholar
Deutsch, M and Gerard, HB (1955) A study of normative and informational social influences upon individual judgment. The Journal of Abnormal and Social Psychology 51(3), 629.Google Scholar
Dietrich, F and Spiekermann, K (2013) Independent opinions? On the causal foundations of belief formation and Jury Theorems. Mind fzt074.Google Scholar
Dixon, G (2016) Applying the gateway belief model to genetically modified food perceptions: new insights and additional questions. Journal of Communication. Retrieved from http://onlinelibrary.wiley.com/doi/10.1111/jcom.12260/fullGoogle Scholar
Dixon, G, Hmielowski, J and Ma, Y (2017) Improving climate change acceptance among US conservatives through value-based message targeting. Science Communication 39(4), 520534.Google Scholar
Efferson, C, Richerson, PJ, McElreath, R, Lubell, M, Edsten, E, Waring, TMBaum, W (2007) Learning, productivity, and noise: an experimental study of cultural transmission on the Bolivian Altiplano. Evolution and Human Behavior 28(1), 1117.Google Scholar
Efferson, C, Lalive, R, Richerson, PJ, McElreath, R and Lubell, M (2008) Conformists and mavericks: the empirics of frequency-dependent cultural transmission. Evolution and Human Behavior 29(1), 5664.Google Scholar
Efferson, C, Lalive, R, Cacault, MP and Kistler, D (2016) The evolution of facultative conformity based on similarity. PloS One, 11(12), e0168551. https://doi.org/10.1371/journal.pone.0168551Google Scholar
Einav, S (2017) Thinking for themselves? The effect of informant independence on children's endorsement of testimony from a consensus. Social Development, 27(1), 7386.Google Scholar
Eriksson, K and Coultas, JC (2009) Are people really conformist-biased? An empirical test and a new mathematical model. Journal of Evolutionary Psychology 7(1), 521.Google Scholar
Estlund, D (1994) Opinion leaders, independence, and Condorcet's jury theorem. Theory and Decision 36(2), 131162.Google Scholar
Festinger, L (1954) A theory of social comparison processes. Human Relations 7(2), 117140.Google Scholar
Gallistel, CR and Gibbon, J (2000) Time, rate, and conditioning. Psychological Review 107(2), 289344.Google Scholar
Gerard, HB, Wilhelmy, RA and Conolley, ES (1968) Conformity and group size. Journal of Personality and Social Psychology 8(1), 79.Google Scholar
Gigerenzer, G (2007) Fast and frugal heuristics: The tools of bounded rationality. In Koehler, D and Harvey, N (eds), Handbook of Judgment and Decision Making. Oxford: Blackwell.Google Scholar
Gigerenzer, G, Todd, PM and ABC Research Group (1999) Simple Heuristics That Make Us Smart. Oxford: Oxford University Press.Google Scholar
Goeree, JK, Palfrey, TR, Rogers, BW and McKelvey, RD (2007) Self-correcting information cascades. The Review of Economic Studies 74(3), 733762.Google Scholar
Gonzalez, R (1994) When words speak louder than actions: another's evaluations can appear more diagnostic than their decisions. Organizational Behavior and Human Decision Processes 58(2), 214245.Google Scholar
Harkins, SG and Petty, RE (1981) Effects of source magnification of cognitive effort on attitudes: an information-processing view. Journal of Personality and Social Psychology 40(3), 401.Google Scholar
Harkins, SG and Petty, RE (1987) Information utility and the multiple source effect. Journal of Personality and Social Psychology 52(2), 260.Google Scholar
Harries, C, Yaniv, I and Harvey, N (2004) Combining advice: the weight of a dissenting opinion in the consensus. Journal of Behavioral Decision Making 17(5), 333348.Google Scholar
Harris, PL (2012) Trusting What you're Told: How Children Learn from Others. Cambridge, MA: Belknap Press/Harvard University Press.Google Scholar
Harris, PL and Lane, JD (2014) Infants understand how testimony works. Topoi, 33(2), 443458.Google Scholar
Hastie, R and Kameda, T (2005) The robust beauty of majority rules in group decisions. Psychological Review 112(2), 49450814.Google Scholar
Haun, DBM and Tomasello, M (2011) Conformity to peer pressure in preschool children. Child Development 82(6), 17591767.Google Scholar
Haun, DBM, Van Leeuwen, EJ and Edelson, MG (2013) Majority influence in children and other animals. Developmental Cognitive Neuroscience 3, 6171.Google Scholar
Henrich, J and Boyd, R (1998) The evolution of conformist transmission and the emergence of between-group differences. Evolution and Human Behavior 19(4), 215241.Google Scholar
Hertwig, R and Erev, I (2009) The description-experience gap in risky choice. Trends in Cognitive Sciences 13(12), 517523.Google Scholar
Hess, NH and Hagen, EH (2006) Psychological adaptations for assessing gossip veracity. Human Nature 17(3), 337354.Google Scholar
Hodges, BH (2015) Conformity and divergence in interactions, groups, and culture. In The Oxford Handbook of Social Influence. New York: Oxford University Press.Google Scholar
Hu, J, Whalen, A, Buchsbaum, D, Griffiths, T and Xu, F (2015) Can children balance the size of a majority with the quality of their information? Proceedings of the Cognitive Science Society Conference, Pasadena, California, July 22–25, 2015.Google Scholar
Hutchins, E (1996) Cognition in the Wild. Cambridge MA: MIT Press.Google Scholar
Jayles, B, Kim, H, Escobedo, R, Cezera, S, Blanchet, A, Kameda, T, … Theraulaz, G (2017) How social information can improve estimation accuracy in human groups. Proceedings of the National Academy of Sciences, 201703695.Google Scholar
Jegadeesh, N and Kim, W (2010) Do analysts herd? An analysis of recommendations and market reactions. Review of Financial Studies 23(2), 901937.Google Scholar
Kao, AB and Couzin, ID (2014) Decision accuracy in complex environments is often maximized by small group sizes. Proceedings of the Royal Society B 281(1784), 20133305.Google Scholar
Kerr, JR and Wilson, MS (2018) Changes in perceived scientific consensus shift beliefs about climate change and GM food safety. PloS One 13(7), e0200295.Google Scholar
Keshk, W (2012) When are investors’ uncertainty judgments influenced by their perceptions about analyst herding? (University of Illinois at Urbana-Champaign) Retrieved from https://www.ideals.illinois.edu/handle/2142/42123.Google Scholar
Kim, C and Pantzalis, C (2003) Global/industrial diversification and analyst herding. Financial Analysts Journal 59(2), 6979.Google Scholar
King, AJ and Cowlishaw, G (2007) When to use social information: the advantage of large group size in individual decision making. Biology Letters 3(2), 137139.Google Scholar
Kroll, Y, Levy, H and Rapoport, A (1988) Experimental tests of the separation theorem and the capital asset pricing model. The American Economic Review, 78(3), 500519.Google Scholar
Kübler, D and Weizsäcker, G (2005) Are longer cascades more stable? Journal of the European Economic Association 3(2–3), 330339.Google Scholar
Laland, KN (2004) Social learning strategies. Animal Learning and Behavior 32(1), 414.Google Scholar
Landrum, AR, Hallman, WK and Jamieson, KH (2018) Examining the impact of expert voices: communicating the scientific consensus on genetically-modified organisms. Environmental Communication, 13(1), 120.Google Scholar
Latané, B (1981) The psychology of social impact. American Psychologist 36(4), 343356.Google Scholar
Latané, B and Wolf, S (1981) The social impact of majorities and minorities. Psychological Review 88(5), 438.Google Scholar
Laughlin, PR (2011) Group Problem Solving. Princeton, NJ: Princeton University Press.Google Scholar
Laughlin, PR, Hatch, EC, Silver, JS and Boh, L (2006) Groups perform better than the best individuals on letters-to-numbers problems: effects of group size. Journal of Personality and Social Psychology 90, 644651.Google Scholar
Lazard, G (2001) On the grammaticalization of evidentiality. Journal of Pragmatics 33(3), 359367.Google Scholar
Lewandowsky, S, Gignac, GE and Vaughan, S (2013) The pivotal role of perceived scientific consensus in acceptance of science. Nature Climate Change 3(4), 399404.Google Scholar
Lopes, D, Vala, J and Garcia-Marques, L (2007) Social validation of everyday knowledge: Heterogeneity and consensus functionality. Group Dynamics: Theory, Research, and Practice 11(3), 223.Google Scholar
MacCoun, RJ (2012) The burden of social proof: shared thresholds and social influence. Psychological Review 119(2), 345.Google Scholar
Maines, LA (1990) The effect of forecast redundancy on judgments of a consensus forecast's expected accuracy. Journal of Accounting Research, 28, 2947.Google Scholar
Maines, LA (1996) An experimental examination of subjective forecast combination. International Journal of Forecasting 12(2), 223233.Google Scholar
Mannes, AE (2009) Are we wise about the wisdom of crowds? The use of group judgments in belief revision. Management Science 55(8), 12671279.Google Scholar
Martin, R and Hewstone, M (2008) Majority versus minority influence, message processing and attitude change: the source-context-elaboration model. Advances in Experimental Social Psychology 40, 237326.Google Scholar
Maynard Smith, J and Harper, D (2003) Animal Signals. Oxford: Oxford University Press.Google Scholar
McElreath, R, Lubell, M, Richerson, PJ, Waring, TM, Baum, W, Edsten, EPaciotti, B (2005) Applying evolutionary models to the laboratory study of social learning. Evolution and Human Behavior 26(6), 483508.Google Scholar
McElreath, R, Bell, AV, Efferson, C, Lubell, M, Richerson, PJ and Waring, T (2008) Beyond existence and aiming outside the laboratory: estimating frequency-dependent and pay-off-biased social learning strategies. Philosophical Transactions of the Royal Society of London B: Biological Sciences 363(1509), 35153528.Google Scholar
Mercier, H (2017) How gullible are we? A review of the evidence from psychology and social science. Review of General Psychology 21(2), 103.Google Scholar
Mercier, H (2019) A paradox of information aggregation: We do it well but think about it poorly, and why this is a problem for institutions. In Ballantyne, N and Dunning, D (eds) Reason, Bias, and Inquiry: New Perspectives from the Crossroads of Epistemology and Psychology. New York: Oxford University Press, in press.Google Scholar
Mercier, H and Miton, H (2019) Utilizing simple cues to informational dependency. Evolution and Human Behavior, in press.Google Scholar
Mercier, H, Dockendorff, M and Schwartzberg, M (submitted) Democratic legitimacy and attitudes about information-aggregation procedures.Google Scholar
Mercier, H, Majima, Y, Claidière, N and Léone, J (submitted) Obstacles to the spread of unintuitive beliefs.Google Scholar
Morgan, TJH, Rendell, LE, Ehn, M, Hoppitt, W and Laland, KN (2012) The evolutionary basis of human social learning. Proceedings of the Royal Society of London B: Biological Sciences 279(1729), 653662.Google Scholar
Morgan, TJH, Laland, KN and Harris, PL (2015) The development of adaptive conformity in young children: effects of uncertainty and consensus. Developmental Science 18(4), 511524.Google Scholar
Muthukrishna, M, Morgan, TJ and Henrich, J (2016) The when and who of social learning and conformist transmission. Evolution and Human Behavior 37(1), 1020.Google Scholar
Mutz, DC (1998) Impersonal Influence: How Perceptions of Mass Collectives Affect Political Attitudes. Cambridge: Cambridge University Press.Google Scholar
Petty, RE and Cacioppo, JT (1986) Communication and Persuasion: Central and Peripheral Routes to Persuasion. New York: Springer.Google Scholar
Petty, RE and Wegener, DT (1998) Attitude change: multiple roles for persuasion variables. In Gilbert, DT, Fiske, S and Lindzey, G (eds), The Handbook of Social Psychology (pp. 323390). Boston, MA: McGraw-Hill.Google Scholar
Slovic, P (1966) Cue-consistency and cue-utilization in judgment. The American Journal of Psychology 79(3), 427434.Google Scholar
Soll, JB (1999) Intuitive theories of information: Beliefs about the value of redundancy. Cognitive Psychology 38, 317346.Google Scholar
Sowden, S, Koletsi, S, Lymberopoulos, E, Militaru, E, Catmur, C and Bird, G (2018) Quantifying compliance and acceptance through public and private social conformity. Consciousness and Cognition 65, 359367.Google Scholar
Sperber, D (2001) In defense of massive modularity. In Dupoux, E (ed.) Language, Brain and Cognitive Development: Essays in Honor of Jacques Mehler (pp. 4757) Cambridge, MA: MIT Press.Google Scholar
Sperber, D, Clément, F, Heintz, C, Mascaro, O, Mercier, H, Origgi, G and Wilson, D (2010) Epistemic vigilance. Mind and Language 25(4), 359393.Google Scholar
Tanford, S and Penrod, S (1984) Social influence model: A formal integration of research on majority and minority influence processes. Psychological Bulletin 95(2), 189.Google Scholar
Toyokawa, W, Whalen, A and Laland, KN (2018) Social learning strategies regulate the wisdom and madness of interactive crowds. BioRxiv 326637.Google Scholar
Trouche, E, Johansson, P, Hall, L and Mercier, H (2018) Vigilant conservatism in evaluating communicated information. PloS One 13(1), e0188825.Google Scholar
Trouche, E, Sander, E and Mercier, H (2014) Arguments, more than confidence, explain the good performance of reasoning groups. Journal of Experimental Psychology: General 143(5), 19581971.Google Scholar
van der Linden, SL, Leiserowitz, AA, Feinberg, GD and Maibach, EW (2014) How to communicate the scientific consensus on climate change: plain facts, pie charts or metaphors? Climatic Change 126(1–2), 255262.Google Scholar
van der Linden, SL, Clarke, CE and Maibach, EW (2015a) Highlighting consensus among medical scientists increases public support for vaccines: evidence from a randomized experiment. BMC Public Health 15(1), 1207.Google Scholar
van der Linden, SL, Leiserowitz, AA, Feinberg, GD and Maibach, EW (2015b) The scientific consensus on climate change as a gateway belief: Experimental evidence. PloS One 10(2), e0118489.Google Scholar
Van Leeuwen, EJ, Cohen, E, Collier-Baker, E, Rapold, CJ, Schäfer, M, Schütte, S and Haun, DB (2018) The development of human social learning across seven societies. Nature Communications 9(1), 2076.Google Scholar
Votruba, AM and Kwan, VS (2015) Disagreeing on whether agreement Is persuasive: perceptions of expert group decisions. PloS One 10(3), e0121426.Google Scholar
Weizsäcker, G (2010) Do we follow others when we should? A simple test of rational expectations. American Economic Review 100(5), 2340–60.Google Scholar
Whalen, A, Griffiths, TL and Buchsbaum, D (2018) Sensitivity to shared information in social learning. Cognitive Science 42(1), 168187.Google Scholar
Whiten, A (2019) Conformity and over-imitation: an integrative review of variant. Advances in the Study of Behavior 51, 31.Google Scholar
Wilder, DA (1977) Perception of groups, size of opposition, and social influence. Journal of Experimental Social Psychology 13(3), 253268.Google Scholar
Wolf, M, Kurvers, RH, Ward, AJ, Krause, S and Krause, J (2013) Accurate decisions in an uncertain world: collective cognition increases true positives while decreasing false positives. Proceedings of the Royal Society of London B: Biological Sciences 280(1756), 20122777.Google Scholar
Yaniv, I (1997) Weighting and trimming: heuristics for aggregating judgments under uncertainty. Organizational Behavior and Human Decision Processes 69(3), 237249.Google Scholar
Yaniv, I (2004) Receiving other people's advice: Influence and benefit. Organizational Behavior and Human Decision Processes 93, 113.Google Scholar
Yaniv, I, Choshen-Hillel, S and Milyavsky, M (2009) Spurious consensus and opinion revision: why might people be more confident in their less accurate judgments? Journal of Experimental Psychology: Learning, Memory, and Cognition 35(2), 558.Google Scholar
Yaniv, I and Kleinberger, E (2000) Advice taking in decision making: Egocentric discounting and reputation formation. Organizational Behavior and Human Decision Processes 83, 260281.Google Scholar
Yousif, SR, Aboody, R and Keil, FC (2018) Any consensus will do: the failure to distinguish between ‘true'and ‘false’ consensus. Proceedings of the Cognitive Science Society Conference, Madison, Wisconsin, 25–28 July 2018.Google Scholar
Zhou, S and Guo, B (2016) The order effect on online review helpfulness: a social influence perspective. Decision Support Systems 93, 7787.Google Scholar
Ziegelmeyer, A, March, C and Krügel, S (2013) Do we follow others when we should? A simple test of rational expectations: comment. The American Economic Review 103(6), 26332642.Google Scholar