Seeking advice: A sampling approach to advice taking

Judgment and Decision Making, Vol. 11, No. 4, July 2016, pp. 401-415

Seeking advice: A sampling approach to advice taking

Mandy Hütter*   Fabian Ache#

The present research addresses advice taking from a holistic perspective covering both advice seeking and weighting. We build on previous theorizing that assumes that underweighting of advice results from biased samples of information. That is, decision makers have more knowledge supporting their own judgment than that of another person and thus weight the former stronger than the latter. In the present approach, we assume that participants reduce this informational asymmetry by the sampling of advice and that sampling frequency depends on the information ecology. Advice that is distant from the decision maker’s initial estimate should lead to a higher frequency of advice sampling than close advice. Moreover, we assume that advice distant from the decision maker’s initial estimate and advice that is supported by larger samples of advisory estimates are weighted more strongly in the final judgment. We expand the classical research paradigm with a sampling phase that allows participants to sample any number of advisory estimates before revising their judgments. Three experiments strongly support these hypotheses, thereby advancing our understanding of advice taking as an adaptive process.


Keywords: advice taking, advice seeking, sampling, ecological approach

1  Introduction

Many decisions occur in an interactive, social context (Larrick, Mannes & Soll, 2012; Sniezek & Buckley, 1995). When we are uncertain about something, we often seek advice from others to enhance decision quality. Advice taking is considered an important, adaptive aspect of human decision making (Yaniv & Milyavsky, 2007). The increase in accuracy through combining independent judgments is a well-established finding (e.g., Harvey & Fischer, 1997; Yaniv, 2004a, 2004b). Aggregated judgments based on multiple opinions are usually more accurate than individual judgments because aggregation balances out random errors (Galton, 1907; Herzog & Hertwig, 2009).

Research has largely focused on how advice is used in judgment revision once it is received (e.g., Harvey & Fischer, 1997; Yaniv, 2004a; but see Gino, Brooks & Schweitzer, 2012, for an exception). The most consistent finding in this area is the underweighting of other people’s estimates, referred to as egocentric discounting (Yaniv & Kleinberger, 2000). Whereas averaging of estimates is usually the most successful strategy (e.g., Armstrong, 2001; Clemen, 1989; Larrick & Soll, 2006) people combine their own estimate with advice in only a subset of the cases (e.g., Harvey & Fischer, 1997; Soll & Larrick, 2009; Soll & Mannes, 2011; Yaniv & Choshen-Hillel, 2012) thereby failing to realize the full benefits of advice taking (Yaniv, 2004a). The informational asymmetry account (Yaniv, 2004b) asserts that egocentric discounting is due to differences in the accessibility of one’s own reasoning relative to that of others. More precisely, whereas one might have a number of reasons for one’s own judgment, one likely has less knowledge about another person’s judgment (Tversky & Koehler, 1994). Support for this assumption stems from studies reporting stronger reliance on advice with a decrease in self-reported knowledge (Yaniv & Choshen-Hillel, 2012) and an increase in task difficulty (Gino & Moore, 2007). Both factors likely lead to reduced support for one’s initial judgment, thereby reducing informational asymmetry. In contrast, informational asymmetry should be more pronounced the more strongly advice differs from one’s own judgment.

1.1  Seeking advice

Some researchers have investigated the impact of multiple advisory judgments (Yaniv, Choshen-Hillel & Milyavsky, 2009) or advice generated by groups (Mannes, 2009) on judgment revision, thereby (virtually) increasing the amount of information participants receive. However, this work does not take into account the fact that people can interact with their social ecology (Bonaccio & Dalal, 2006; Gino et al., 2012). For instance, the frequency with which advice is consulted could vary across situations. The present research investigates advice seeking behavior. Initial research presented factors that affect people’s general willingness to consider a piece of advice (Gino et al., 2012; Gino & Moore, 2007). However, advice taking is often a sequential process that can be truncated at different points in time. Thus, depending on characteristics of the decision task, personal factors (e.g., knowledge), or the advice received, people could seek smaller or larger samples of advice.

Given that advice taking is an instance of interactive decision making, it appears highly artificial to assume that people’s only options to react to advice are integrating it with their own estimate or rejecting it. In contrast, we assume that receiving advice can also instigate additional advice seeking. Specifically, the informational asymmetry account (Yaniv, 2004b) suggests that advice is evaluated with regard to how much information supports it. The advisee’s search for information might not be limited to her own mind. Thus, within a plausible range, advice might motivate people to consider more advice. Moreover, given the importance of independence of information to achieving substantial improvements by aggregation (Einhorn, Hogarth & Klempner, 1977; Herzog & Hertwig, 2009), the information samples people create by consulting different sources and quantities of advice constitute a necessary precondition for any improvement of decision quality through advice taking.

1.2  An ecological approach to advice taking

The present research expands the informational asymmetry account to investigate the process of advice taking — both seeking and integrating advice. We base this research on the premise that decision makers often have opportunities to consult multiple people and take an active role in the search for information. Such interactions between cognition and the information ecology have long been recognized (Brunswik, 1955; Simon, 1956) and have been studied recently as well (e.g., Hertwig & Erev, 2009; Hertwig, 2015). Sampling as a theoretical approach assumes that biases in the information sample on which judgments are based might be a more parsimonious explanation for biases often assumed to reside within the individual’s mind (Fiedler, 2000). The informational asymmetry account constitutes an instance of biased samples, on the basis of which the decision maker rationally chooses to integrate (or reject) advice (Ravazzolo & Røisland, 2011). However, the classical research paradigm never allowed decision makers to obtain additional information, leaving open the question of whether individuals spontaneously reduce the asymmetry by sampling. Related paradigms reveal that people often show profound biases in their tendency to search for different types of information (e.g., creating valence asymmetries in attitude acquisition; Denrell, 2005; Fazio, Eiser & Shook, 2004).

In the present research, we establish an advice taking paradigm that allows participants to sample any number of other people’s judgments. This paradigm allows for the investigation of whether the sample of advisory estimates is sensitive to features of the decision situation. We derive and test two straightforward assumptions of our ecological account:

Assumption 1: The frequency of sampling is greater when advice is distant from rather than close to the decision maker’s initial estimate. This distance has been shown to influence judgment revision (Yaniv, 2004b; Yaniv et al., 2009), but to our knowledge, has not been applied to advice seeking. The distance of the advice from the decision maker’s initial estimate constitutes a cue to advice independence, a condition under which highest improvements in judgment accuracy can be expected (e.g., Larrick & Soll, 2006). Hence, distant advice should be preferred over close advice, if advice sampling is motivated by the prospect of acquiring new, independent information about the decision problem. Such a motive can be distinguished from a motive to be affirmed by advisors (Kahneman & Tversky, 1973). In this case, close advice should be preferred.

If sampling behavior is indeed a function of informational asymmetry and thus motivated by the informational value of advice, sampling should be more frequent when advice is distant rather than close. The present research manipulates advice distance in order to test this crucial assumption of the expanded informational asymmetry account. Specifically, we expect that distant as compared to close advice will lead to a higher frequency of additional sampling.

The next assumptions presume that the integration of advice depends on the samples drawn.

Assumption 2a: Advice is more likely to be integrated with a decision maker’s initial estimate if it is distant from the initial estimate rather than close. Distance of advice has repeatedly been shown to constitute an important moderator of egocentric discounting in advice weighting. Yaniv (2004a) postulated a monotone negative relationship between advice distance and weighting with decreasing integration of more distant advice. This pattern was reported in a number of studies (Minson, Liberman & Ross, 2011; Yaniv & Milyavsky, 2007; Yaniv, 2004b). Recent research, however, reported a curvilinear relationship with the strongest integration of advice on intermediate distance levels, which is regarded most informative (Moussaïd, Kämmer, Analytis & Neth, 2013; Schultze, Rakotoarisoa & Schulz-Hardt, 2015). Advice that is very distant from the decision maker’s perspective might not be integrated as it might become incredible to the advisee and is thus not regarded informative. Close advice also offers little new information to the decision maker. In contrast, it confirms the decision maker’s intial judgment. Close advice is thus not utilized for judgment updating, but rather for the updating of confidence (Schultze et al., 2015) serving a social validation function (Kahneman & Tversky, 1973; Schulz-Hardt, Frey, Lüthgens & Moscovici, 2000). In sum, these findings provide initial evidence for decision makers’ sensitivity to the informational value of advice. In the present research we compare relatively close and distant advice and predict that distant advice will be weighted more strongly than close advice.

Assumption 2b: Advice is more likely to be integrated with a decision maker’s initial estimate the more estimates sampled. Judgment updating should not be a function of only the distance of the advice from the decision maker’s initial estimate. The degree to which advice is used for judgment updating should depend on the degree of supportive information (i.e., its amount and consistency). That is, the more pieces of advice support an alternative position, the more strongly this position should be held. In fact, initial research indicates that advice weighting is sensitive to the number of advisors providing it (Mannes, 2009). We thus expect that an alternative position will be weighted more strongly, the more pieces of advice support it. Therefore, advice integration should be a direct, positive function of sampling frequency.

In the present work, we focus on the number of pieces of advice sampled, rather than the number of reasons underlying a given judgment, for three reasons. First, in everyday life the judgments of other people can be observed more frequently than the reasons underlying a judgment. Second, when judgments are based on intuition, concrete reasons are often not available or constructed after the fact (Nisbett & Wilson, 1977). Third, we assume that a judgment is evaluated in accordance with the amount and consistency of information, as in an assessment of reliability, a principle that should be applicable to the number and consistency of advisory judgments as well as the number and consistency of supporting reasons (Koriat, 2012).

1.3  Overview of experiments

We conducted three experiments that followed a general procedure. As in the classical approach to advice taking, participants were presented with an estimation task, in this case the caloric content of various foods per serving. After giving an initial estimate, participants received an initial piece of advice, supposedly sampled from a distribution of 100 participants in an earlier study. This piece of advice was either similar to or diverging from their initial estimate. All three experiments implemented this factor in a within-participants design. Experiment 1 resembled the classical research approach without a sampling phase and served as an initial test of the weighting functions of our operationalization of close and distant advice. In Experiments 2 and 3, participants could sample additional pieces of advice at will before giving their final estimate. While Experiment 2 implemented advice that was consistent among advisors, Experiment 3 utilized ecologically valid and thereby considerably less consistent advice.

1.4  Data analysis

As dependent measures, we analyzed the size of the self-determined samples as well as the integration of advice when forming a final judgment. As a measure of advice integration, we calculated the weight of advice (WOA; Harvey & Fischer, 1997; Yaniv, 2004a), defined as

WOAij = (Fij − Iij) / (Aij − Iij)     (1)

where I, F, and A indicate the initial estimate, final estimate, and advice, respectively, on a given trial j in a given participant i. We calculated the mean of all pieces of advice received to arrive at an estimate for A. The WOA thus reflects the degree to which participants move towards the advice (i.e., a WOA of 1 indicates adoption of the advice; a WOA of 0 indicates adherence to the initial estimate; a WOA of 0.5 indicates an equal-weighting strategy). The WOA thus allows for an assessment of (increasing or decreasing) receptivity towards advice as a function of amount sampled or other variables.

The WOA is highly sensitive to outliers. Outliers primarily result from trials where the distance between initial estimate and advice is very small, so that even small alterations of the judgment may lead to values outside the range of 0 and 1. Many researchers thus truncate the WOA, setting values smaller than 0 to 0 and values greater than 1 to 1 (e.g., Gino, 2008; Schultze et al., 2015; Soll & Larrick, 2009), arguing that this practice is unproblematic if it concerns less than 5% of all trials. As participants in our expanded paradigm (Experiment 2 and 3) received multiple pieces of advice that scattered, there were more trials with values of WOA outside the range of 0 and 1 than in Experiment 1. For the analyses reported in the main text, we opted to apply statistical criteria to identify and remove outliers on a trial-by-trial basis (Tukey, 1977) rather than altering values of WOA. Additional analyses applying the truncation practice yielded similar results, but led to a higher number of alterations in Experiment 2 and 3. Note that, for the influence of sampling on advice weighting, the qualitative pattern of results is unaltered even when no trials are excluded. However, the effect of advice distance on WOA reverses (Experiment 1 and 3) in the absence of any exclusion.

For all experiments, we calculated multilevel model analyses for all dependent measures to assess relationships on a trial-by-trial basis (Judd, Westfall & Kenny, 2012). Analyses are based on 20 data points per participant. Following the recommendations by Judd and colleagues (2012), all models contained random intercepts for participants as well as items, which were fully crossed by design. The effects of our hypothesized predictors were always fixed. We determined the best-fitting model by sequentially including the hypothesized parameters and their interactions, one at a time and inspecting increase in model fit using the Kenward-Roger approximation (Kenward & Roger, 1997) for degrees of freedom. The resulting p-values as well as log-likelihood values and approximated Bayes factors (Masson, 2011) for each model are given in the Appendix. For the sake of brevity, we will report only the parameter estimates of the best-fitting model.

To summarize, we predict that sampling behavior will increase with the distance of advice from the decision maker’s initial estimate (Assumption 1). We test this assumption in Experiments 2 and 3, by looking for a main effect of advice distance on sample size for the multilevel model. We predict a larger sample size in distant as compared to close advice. We also expect advice integration to be sensitive to participants’ information samples. Specifically, we predict two main effects of advice distance (Assumption 2a; Experiments 1–3) and amount of information (Assumption 2b; Experiments 2 and 3) on the WOA in a multilevel model, such that the WOA will be larger with distant (rather than close) advice and larger (rather than smaller) information samples.

1.5  A-priori power analyses

We conducted a-priori power analyses to determine required sample sizes in all three experiments (Faul, Erdfelder, Lang & Buchner, 2007). As there is no clear guideline for power analyses regarding multilevel modelling utilized in the present research, we based our calculation on repeated measures ANOVA designs. Detecting a medium-sized effect (f = .25) with sufficient power (β = .95) in Experiments 1 and 2 required collecting data of at least 20 participants. We assumed a smaller effect (f = .15) in Experiment 3 due to the increased variance in advice, requiring at least 54 participants. As the experiments were part of multi-experiment sessions, we increased sample size in accordance with other studies that required a larger sample size.

2  Experiment 1

The first experiment served as a conceptual replication of advice taking procedures that do not include a sampling phase using the same materials and basic manipulation as in the next two experiments. Specifically, we compared two advice distance conditions. The close condition was intended to confirm participants’ initial judgments leading to little judgment revision. In the distant condition, advice was intended to increase participant’s tendency to revise their judgment, as in the intermediate distances in Moussaïd et al.’s (2013) and Schultze et al.’s (2015) work.

2.1  Method

Participants. The sample consisted of 35 University of Tübingen students of different subjects (10 males; Mage= 23.00 years, SDage= 6.18).1 Participants took part in a one-hour session comprised of several social cognitive experiments and were compensated with either course credit or 7€ and a chocolate bar. Additionally, participants received a small bonus payment depending on task performance. Participants were recruited via e-mail and online social networks.

Design. The experiment implemented a 2 (distance of advice: close vs. distant) × 2 (judgment phase: initial vs. final judgment) within-participants design.

Materials and procedure. After participants signed a consent form, we asked them to estimate the caloric content of various dishes per serving. Names, pictures and caloric content of dishes were retrieved from a web page of a German nutrition magazine (http://www.essen-und-trinken.de).2 We pretested 30 dishes using an online sample of 21 participants. Out of those 30 dishes, we selected the 20 dishes for which the average estimate most closely resembled the true caloric content of the dish (e.g., fish pasta). This procedure assured that the population had some knowledge about the caloric contents of those dishes (Gino, 2008).

The order of dishes was randomized for each participant, as was the order of the 10 trials per distance condition across the 20 trials.

For each trial, participants first saw a picture of the dish along with a descriptive label and a response box to give their initial estimate. If participants submitted an estimate lower than 50 or higher than 1250 calories per serving, the software prompted them to give a plausible estimate. Next, they were shown another estimate for that same dish. These estimates were allegedly drawn from a pool of estimates of 100 participants in a preceding experiment. Depending on distance condition, the advice was simulated as a pseudo-normal distribution centering on a mean relatively close to the participant’s initial estimate (on close trials) or centering on a mean relatively distant from the participant’s initial estimate (on distant trials). Specifically, the mean of the distribution was simulated as i ± .5 t (on distant trials) and i ± .05 t (on close trials) where i is the participant’s initial estimate and t is the true value of the dish. If the participant underestimated in comparison to the true value, the proportion was added to the initial estimate; if the participant overestimated in comparison to the true value the proportion was subtracted. Consequently, the advice would generally point in the direction of the true value. However, if the initial estimate was very close to the true value, the advice could also lead participants away from the true value, by causing them to adjust too much. The standard deviation of the distribution was equal for close and distant advice. Random noise was added to obscure the artificial nature of advice, drawn from a uniform distribution in the range of [–8, +8] calories.

After receiving advice, participants gave a final estimate for the given dish. For each of these final estimates that fell in a range extending 10% on either side of the true caloric content of the dish, participants would receive a bonus payment of 0.10€. During the experiment, participants did not receive feedback on their accuracy or their cumulative bonus. The bonus was revealed only at the end of the experiment. In total, participants could collect a bonus of 2€. Finally, participants were probed for suspicion, debriefed, paid, and thanked.

2.2  Results


Table 1: Sample size of self-determined information samples, weight of advice (WOA), deviation of WOA from equal weights averaging (Δ WOA), and confidence on close and distant trials for Experiments 1–3.
  Distance of advice 
DV
Experiment
Close
Distant
Cohen’s d
2*Sample size
2
7.29 (6.83)
8.82 (7.11)
.37
 
3
9.82 (5.97)
10.69 (5.97)
.21
3*WOA
1
.05 (.13)
.17 (.17)
.59
 
2
.14 (.35)
.40 (.33)
.57
 
3
.15 (.27)
.22 (.30)
.16
3*Δ WOA
1
–.45 (.13)
–.33 (.17)
.59
 
2
–.60 (.36)
–.39 (.31)
.47
 
3
–.70 (.29)
–.64 (.30)
.11
2*Confidence
2
0.29 (1.33)
–0.34 (1.21)
.58
 
3
–0.12 (1.37)
–0.43 (1.38)
.26
Note. Cells display mean values across participants and trials. Standard deviations are given in parentheses.

Distance of advice. As a manipulation check, we assessed the absolute difference between participants’ initial estimates and the initial piece of advice on the level of trials. Whereas participants’ initial estimate and the initial piece of advice differed by only M = 30.80 (SD = 21.19) calories on close trials, they differed by M = 241.85 (SD = 83.44) calories on distant trials.

We also calculated the normalized distance in line with Moussaïd et al. (2013):

Δ Eij = |Iij − Aij| / Iij,     (2)

where I and A denote the initial estimate of the participant and the advice, respectively, on a given trial j in participant i. Moussaïd et al. (2013) define values below 0.3 as “similar” advice, which they found to be related to an updating of confidence rather than an updating of one’s judgment. Values between 0.3 and 1.1 were defined as an “intermediate distance”, for which strong judgment updating was observed. Values above 1.1 were classified as “very distant” and were empirically related to diminishing influence on participants’ judgments. Employing the average advice on each trial as the standard of reference, the average Δ E amounted to M = .07 (SD = .06) in the close condition and to M = .53 (SD = .28) in the distant condition. Our operationalizations of the close and distant conditions thus resemble the “similar” and “intermediate distance” conditions of Moussaïd et al. (2013)

As a result of the distance manipulation, participants’ initial estimate and the advice enclosed the true value in 79% of the distant trials, but only 17% of the close trials.

Using advice. We excluded trials with a WOA < −0.42 and WOA > 0.67 (Tukey, 1977). In total, we excluded 49 of 700 trials (7.00%).

We tested the hypothesis that advice weighting is greater for distant as compared to close advice. We fitted the following model:

WOAij = α0 + b1Distancej + є,

where index i refers to participants and index j to questions. For this and all following models close advice was coded as 0 and distant advice as 1. The effect of distance thus indicates the consequences of increasing advice distance from close to distant advice. Replicating the results of other authors, advice distance significantly increased participants’ WOA, b1 = .12, se = .01, t = 10.65 (Table 1), although fit indices indicated worse fit for this model over the null model in this specific data set. In all subsequent experiments, including distance as a predictor of WOA improved model fit (Appendix).

2.3  Discussion

Participants weighted the advice stronger when it was distant from rather than close to their initial estimates. These results validate our manipulation of advice distance replicating the finding of greater weighting for advice of intermediate rather than small distance (Moussaïd et al., 2013; Schultze et al., 2015).

3  Experiment 2

The second experiment introduced a sampling phase that allowed participants to sample as many pieces of advice as desired. This experiment allowed us to assess whether distant advice indeed instigates a longer search for advice than close advice, and how several consistent pieces of advice are integrated during judgment revision. In order to investigate whether close advice in our paradigm actually led to higher levels of confidence, we introduced a confidence measure after each final estimate. We did that for only half of the participants to assess potential effects of reflecting on one’s confidence.

3.1  Method

Participants. The sample consisted of 44 University of Heidelberg students (14 males; Mage = 26.48 years, SDage = 8.81). The study was run as part of an experimental session of approximately 45 minutes. Participants received either course credit or a monetary compensation of 6€. Additionally, participants received a performance-contingent bonus as in Experiment 1.

Design. This experiment implemented a 2 (distance of advice: close vs. distant) × 2 (judgment phase: initial vs. final judgment) within-participants design. Additionally, half of the participants [N = 22] were asked to provide confidence ratings for their final estimates.

Materials and procedure. Up to eight participants took part at the same time, each seated in a separate cubicle. Materials and procedure were identical to Experiment 1 with two exceptions. First, after receiving a close or a distant piece of advice on the caloric content of a given dish, participants were allowed to sample as many other estimates as desired. The sampling phase was however capped at 20 pieces of advice. Advice was consistently drawn from either the close or distant distribution depending on distance condition. Second, participants in the confidence ratings condition also rated their confidence on a 7-point scale anchored at –3 = “very unconfident” and 3 = “very confident” after each trial, while participants in the no confidence ratings condition proceeded without any further questions. As the presence or absence of confidence ratings did not exert effects on any dependent measure, we will not discuss this factor any further.

3.2  Results

Distance of advice. As a manipulation check, we assessed the absolute difference between participants’ initial estimates and the average advice for each trial. Whereas participants’ initial estimate and the advice differed by only M = 26.34 (SD = 15.75) calories on close trials, they differed by M = 246.18 (SD = 77.57) calories on distant trials.

The Δ E to the average advice amounted to M = .06 (SD = .06) in the close condition and to M = .54 (SD = .26) in the distant condition. Hence, we realized a “similar” and “intermediate” distance condition in terms of Moussaïd et al. (2013). As a result, participants’ initial estimate was enclosed by the range of advice for 51% of close trials but for none of the distant trials. Initial estimate and range of advice enclosed the true value on 22% of the close and 84% of the distant trials.

Sampling of advice. We first inspected the average number of pieces of advice drawn as a necessary precondition for analyzing moderating variables of advice seeking behavior. Participants sampled more advice than the one piece classically offered. The average of M = 8.05 (SD = 7.00) advisory estimates sampled by participants across distance conditions was significantly larger than one, t(43) = 7.27, p < .001, d = 1.10.

Further, we expected greater informational asymmetry cued by distant as compared to close advice to increase sampling frequency. For advice sampling, the model with the best fit was the following:

Sample sizeij = α0 + b1Distancej + є .

Advice distance significantly increased sample size, b1 = 1.53, se = .20, t = 7.82. That is, receiving advice that was distant from their initial estimate led participants to consider about 1.5 pieces of advice more than when receiving advice that was close to their initial estimate (Table 1).

Using advice. We excluded trials with WOA < −0.90 and WOA > 1.40, amounting to 41 of 880 data points (4.66%; Tukey, 1977). Consistent with Experiment 1, we expected distant advice to lead to greater reliance on advice. Testing our assumption of sensitivity to the sampled information, we expected increased sampling to increase reliance on advice. The following model had the best fit:

WOAij = α0 + b1Distancej + b2Sample sizeij + є .

Replicating the results of Experiment 1, advice distance increased participants’ WOA, b1 = .23, se = .02, t = 11.21 (Table 1). Repeatedly encountering a consistent piece of advice increased the WOA as indicated by the effect of sample size, b2 = .015, se = .003, t = 5.66.3

The analysis of advice weighting strongly supports our assumption of increased reliance on an advisory estimate when it is supported by additionally sampled, consistent pieces of advice.

To complement the traditional WOA analysis, we also assessed how the weighting strategy employed by participants fit with the normative rule of equal weights averaging (Mannes, 2009). We therefore calculated the normative WOA participants should have assigned to the mean advice on a given trial as Nadvice / (Nadvice + 1), where Nadvice is the number of advisory estimates received on a given trial (Mannes, 2009), and subtracted it from the actual WOA participants assigned to receive a Δ WOA. For instance, if on a given trial a judge received nine pieces of advice, from an equal weighting perspective, the normative WOA should be 0.9. If the actual WOA on that trial is 0.6, the Δ WOA would amount to –0.3, indicating that the judge underweighted the advice. A positive score in this case implies overweighting of advice in comparison to the normative rule of equal weights averaging. The best fitting model for this deviation score was the following:

Δ WOAij = α0 + b1Distancej + b2Sample sizeij + b3Distancej × Sample sizeij + є.

Replicating numerous studies on failure to follow the normative rule of equal weights averaging (e.g., Mannes, 2009), participants fall short of this rule on both, close and distant trials (Table 1). However, assigned weights more closely matched the normative rule when advice was distant rather than close, b1 = .17, se = .03, t = 5.81.

Although sampling increased the weight participants placed on an advisory estimate, it did so suboptimally, as the deviation from the normative rule increased with increasing sample size, b2 = –.012, se = .003, t = –3.81. This effect was more pronounced for sampling on close as compared to distant trials, as indicated by the interaction, b3 = .008, se = .003, t = 2.78.

Confidence. In line with other authors (Moussaïd et al., 2013; Schultze et al., 2015), we assumed that receiving close as compared to distant pieces of advice would lead to higher levels of confidence. Additionally, we wanted to assess whether increased sampling of close versus distant advice has an effect on confidence. The following model had the best fit:

Confidenceij = α0 + b1Distancej + є.

Indeed, increasing distance of advice decreased confidence, b1 = –.63, se = .07, t = – 9.28 (Table 1). Additional sampling of consistent advice did not affect confidence, as indicated by the missing increase in model fit when including sample size as an additional predictor (Appendix).

3.3  Discussion

Allowing participants to sample additional pieces of advice, the results of Experiment 2 testify to participants’ willingness to obtain additional information. Further, the effect of advice distance on sample size demonstrates that receiving diverging as compared to similar pieces of advice increased participants’ sampling frequency. In line with our expanded informational asymmetry account, participants were sensitive to their information samples, shifting more strongly towards the advisory opinion the higher the number of consistent pieces of advice supporting this opinion they sampled.

Further validating our distance manipulation as providing information supporting or diverging from participants’ initial judgments, confidence was significantly lower on distant as compared to close trials.

4  Experiment 3

The third experiment was intended to increase the ecological validity of our approach, as the conclusions drawn from Experiment 2 are subject to a number of limitations. First, there was a cumulative effect of distance of advice. That is, if a participant, who received a distant (close) piece of advice, decided to consider an additional piece of advice, she received another distant (close) piece of advice. This cumulative effect could per se have an influence on participants’ sampling behavior, as it successively confirms the advisors’ (and in the case of close advice one’s own) position. Second, the advice always pointed in the direction of the true value, so that advice was generally of better quality than the own judgment. To eliminate these limitations, we replicated Experiment 2 with a more ecologically valid advice simulation method. Specifically, we implemented a procedure in which only the very first piece of advice differed between distance conditions. All subsequently sampled pieces of advice were drawn from the same distribution, without regard to condition, at random.

4.1  Method

Participants. The sample consisted of 58 University of Tübingen students (13 males; Mage= 21.10 years, SDage= 3.34)4, who took part in exchange for course credit or a monetary compensation of 7€ and a chocolate bar. The experiment was part of a session that lasted about 60 minutes. Additionally, participants received a performance-contingent bonus as in the previous experiments.

Design. This experiment implemented a 2 (distance of advice: close vs. distant) × 2 (judgment phase: initial vs. final judgment) within-participants design.

Materials and procedure. Materials and procedure were identical to Experiment 2, with two exceptions. First, the generation of advice differed in the following way. For each dish, the distribution of advice consisted of 100 values drawn at random from a normal distribution with the mean and standard deviation of the initial estimates from participants in Experiment 2. Depending on distance condition (close vs. distant advice) participants received an initial piece of advice that was either drawn from the same decile as their initial estimate (on close trials) or from the fifth next decile, above or below their initial estimate, depending on which decile their initial estimate fell into (on distant trials). For instance, if a participant’s initial estimate fell into the 6th decile, she would receive an initial piece of advice from the 6th decile (close trials) or from the 1th decile (distant trials). All additional pieces of advice received during the sampling phase were randomly drawn from the distribution of advice for the respective dish, independent of distance condition. The mean standard deviation of the distributions of advice experienced by participants was M = 163.61 with a range of [125.08; 214.29]. Thus, the advice received was more inconsistent than in Experiment 2, where the average standard deviation of the sampled advice was M = 24.93 with a range of [6.36; 46.67]. As a second deviation from Experiment 2, all participants provided confidence ratings for their final estimates.

4.2  Results

Distance of advice. As a manipulation check, we assessed the absolute difference between participants’ initial estimates and the initial piece of advice for each trial. Whereas participants’ initial estimate and the initial piece of advice differed by only M = 24.88 (SD = 29.28) calories on close trials, they differed by M = 274.05 (SD = 76.82) calories on distant trials. In contrast, the distance between participants’ initial estimate and the average advice received, if participants chose to sample, converged as we manipulated only the distance of the first piece of advice (close: M = 98.95, SD = 85.11; distant: M = 148.53, SD = 104.71).

Likewise the average Δ E between the decision maker’s initial estimate and the first piece of advice amounted to M = .06 (SD = .08) in the close condition and M = .60 (SD = .32) in the distant condition. Hence, we realized a “similar” and “intermediate” distance condition in terms of Moussaïd et al. (2013). Averaging across all pieces of advice, Δ E becomes more similar between conditions with M = .24 (SD = .32) in the close condition and M = .33 (SD = .32) in the distant condition.

As a result of the greater variance of advice compared to the previous experiments, participants’ initial estimate was included in the range of advice for 80% of close trials and for 77% of distant trials. Participants’ initial estimate and the sampled advice enclosed the true value on 83% of close and 95% of distant trials.

Sampling of advice. Participants again sampled significantly more advisory estimates than the one piece received in previous research, M = 10.26 (SD = 5.98), t(57) = 10.52, p < .001, d = 1.77. The following model had the best fit for sample size:

Sample sizeij = α0 + b1Distancej + є .

Although only the first piece of advice differed as a function of the distance manipulation, advice distance increased sample size, b1 = .87, se = .18, t = 4.83 (Table 1).

Using advice. In analyzing the WOA we excluded trials with WOA < –0.65 and WOA > 1.03, amounting to 119 of 1160 trials (10.26%; Tukey, 1977). The following model had the best fit for WOA:

WOAij = α0 + b1Distancej + b2Sample sizeij + є .

Increasing advice distance increased participants’ WOA, b1 = .06, se = .02, t = 3.64, (Table 1). Albeit being less consistent among advisors than in Experiment 2, sampling of advice again increased the WOA, b2 = .007, se = .002, t = 3.40.

We again complemented traditional WOA analysis by investigating participants’ behavior in comparison to the normative rule of equal weights averaging (see Experiment 2). The model with the best fit was:

Δ WOAij = α0 + b1Distancej + b2Sample sizeij + є.

As in Experiment 2, participants fell short of the normative rule on both close and distant trials (Table 1), but came closer when advice was distant, b1 = .05, se = .02, t = 3.23. Again, while sampling increased the WOA, it did so suboptimally, b2 = –.009, se = .002, t = –4.19.

Confidence. For confidence, the following model had the best fit:

Confidenceij = α0 + b1Distancej + b2Sample sizeij + є.

Replicating the findings of Experiment 2, advice distance decreased confidence, b1 = –.26, se = .05, t = –5.24 (Table 1). Interestingly, additional sampling of advice further reduced confidence, b2 = –.035, se = .008, t = –4.46.

4.3  Discussion

Experiment 3 replicated the findings of Experiment 2 under ecologically more valid conditions. Even though the distance of advice was manipulated via a single instance of advice only, initially receiving one diverging as compared to similar piece of advice led participants to sample more advice and to weight advice more strongly. Despite being less consistent among advisors than in Experiment 2, sampling of additional advice increased reliance on advice. We also replicated the effect of advice distance on confidence. Moreover, in contrast to the previous experiment, sampling of advice generally reduced confidence, as advice was less consistent.

5  General discussion

We analyzed advice taking from a broader perspective, covering both advice seeking and weighting. Specifically, we expanded the informational asymmetry account (Yaniv, 2004b), conceptualizing egocentric discounting as an instance of biased samples. Three experiments tested the main assumptions of this expanded informational asymmetry account using a research paradigm that allows for the sampling of advice. The results show that participants were generally disposed to sampling additional advice. Participants’ sampling behavior was sensitive to features of the information ecology, with increased sampling of more informative, diverging advice (Assumption 1). Moreover, advice integration was sensitive to the information samples created by sampling, with stronger reliance on distant advice (Assumption 2a) and a greater reliance on advice that is supported by additional pieces of advice (Assumption 2b). The current approach thus expands research on advice taking by revealing first relations between advice seeking and advice integration and advances current theoretical approaches.

5.1  A sampling approach to advice taking

By investigating the process of sequential advice seeking, the present research answers a repeated call from other scientists (Bonaccio & Dalal, 2006; Gino et al., 2012). The current results reveal advice seeking should no longer be neglected in research on advice taking. Previous research already pointed towards a generally large willingness to consider advice (Gino & Moore, 2007). However, by allowing participants to decide on the amount of advice, we created more variance and thereby the possibility to assess factors influencing advice seeking (which was prevented by ceiling effects in the above mentioned study by Gino and Moore [2007]) and to investigate how participants react to different types of advice. The current research is thus the first to assess the sequential process of seeking advice.

The present experiments not only demonstrate a general willingness to sample advice but also show that advice seeking is susceptible to features of the information ecology, such as the distance of advice from the decision maker’s initial estimate (Assumption 1). That is, distant as compared to close advice increased participants’ propensity to consider further advice. These findings challenge less optimistic conceptualizations of advice taking (Yaniv, 2004a). Specifically, in the classical empirical approach, participants show egocentric discounting that is attributed to unequal information samples (Yaniv, 2004b). The present results show that individuals do not stick to these information samples but expand them when given the opportunity. We found this to be the case particularly when (initial) advice deviated from their own views. That is, the more the advisory estimates differed from the own initial judgment, the more people searched for additional information. Interestingly, a post-hoc comparison of average sample sizes in Experiments 2 and 3 reveals that sampling frequency when advice was close was significantly higher in Experiment 3 (bexp = 2.54, se = 1.16, t = 2.18; bexp*distance = –0.67, se = 0.267, t = –2.49), hinting at overall variance or inconsistency as another moderator of sampling frequency.

Advice weighting was sensitive to the sampled information with distant advice (Assumption 2a) and frequency of sampling increasing the weight of advice (Assumption 2b). The latter finding is supported by a comparison of the WOA for Experiments 1 and 2. The opportunity to sample additional advice significantly reduced egocentric discounting (i.e., increased the WOA) on both close (bexp = .095, se = .036, t = 2.66) and distant trials, (bexp*distance = .132, se = .025, t = 5.22).

In sum, the present results demonstrate a high degree of adaptiveness of advice taking and add to the controversy about the roots of egocentric discounting (Soll & Mannes, 2011; Yaniv & Choshen-Hillel, 2012). Building on informational asymmetry as a core property of advice taking, the results show that egocentric tendencies can be reduced via the sampling of information.

The present research complements findings from a related research area that investigates sampling behavior prior to the choosing of options. In research on optional stopping, for instance, participants sample (costly) offers until they stop sampling and keep the latest offer (e.g., Bearden & Rapoport, 2005; Rapoport & Tversky, 1966; Seale & Rapoport, 1997). In research investigating decisions from experience, participants are usually presented with two risky lotteries and explore their outcomes before deciding which one of these lotteries to play for actual profit (Hertwig, Barron, Weber & Erev, 2004; Hertwig & Erev, 2009). Although there is some debate about the optimal sample size (e.g. Hertwig & Pleskac, 2010), this research generally demonstrates that participants sample substantially, investing money and/or time. Moreover, as in the present research the amount of sampling is adaptive, being sensitive to the variability of the sample (e.g., Lejarraga, Hertwig & Gonzalez, 2012; Mitra, Reiss & Capella, 1999; Pachur & Scheibehenne, 2012), the importance of the decision (Hau, Pleskac, Kiefer & Hertwig, 2008), and competition for the choice options (Phillips, Hertwig, Kareev & Avrahami, 2014). In these paradigms, participants try to maximize their profit. In the present research, the structure of motives is different in that the task invites both a motive for affirmation and an accuracy motive. The latter is only indirectly linked to financial outcomes. Moreover, participants are oriented towards the central tendency (or consistency) of the sampled information. Nevertheless, the two research areas converge on the adaptiveness of sampling.

At first glance, the present findings are at odds with an analysis presented by Soll and Mannes (2011). In an attempt to test whether differential access to reasons caused the differential weighting of own and others’ judgments, they created a paradigm in which participants predicted outcomes from experimentally controlled cues. Whether or not these cues were presented during judgment revision did not affect advice integration. From this null-effect, we could conclude that the differential availability of reasons does not explain egocentric discounting effects. Alternatively, we could conclude that the mere repetition of information underlying the initial judgment was not effective in increasing egocentric tendencies. In the present paradigm, we investigated the effects of seeking additional information on advice integration. In contrast to Soll and Mannes (2011), we show that an increase of the advisory information base by sampling reduces egocentric tendencies.

5.2  Limitations and future directions

The present results demonstrate the adaptiveness of advice taking in terms of its flexibility in light of certain affordances in the environment. Another connotation of the term “adaptiveness”, however, concerns the improvement that comes with this flexibility. The present research did not address the second type of adaptiveness for two reasons. First, the current research was focused on whether and under what conditions decision makers consult advisors. Second, the manipulated advice generally pointed towards the true value (except for the sampled advice in Experiment 3), so that we refrained from reporting effects on accuracy, as those are bound to this artificial setting. Future research should thus ask whether the sort of flexibility shown here improves accuracy under ecologically valid conditions.

The laboratory setting also limits the generalization of the present findings to real-life decision making. It is well established that people preferably turn to close and similar others for advice (McPherson, Smith-Lovin & Cook, 2001; Suls, Martin & Wheeler, 2002) and that they even assume these others to be representative of the actual population distribution (Galesic, Olsson & Rieskamp, 2012). One might thus argue that people would not even encounter diverging advice in their ecology. However, two findings render us optimistic. First, people exhibited substantial sampling even in light of similar advice. Thus, the chance of encountering even rare events during this sampling process is not negligible. Second, advice in Experiment 3 was simulated based on the actual population distribution parameters by a previous sample from the same student population, the same group of peers that participants would most likely turn to for advice in real life.

To assess preferences for close versus distant advice in an even stricter manner, both types of advice should be offered concurrently. As people preferably turn to close and similar others for advice (McPherson et al., 2001; Suls et al., 2002), participants might appear to be less well adapted when being offered both types of advice simultaneously, preferring advice consistent with their opinion. Moreover, the present paradigm can be extended to advisor characteristics that were shown to affect conformity in advice taking and other, social psychological paradigms such as persuasion. For instance, from the present perspective it can be expected that characteristics that communicate competence and trustworthiness such as expertise and reputation would strongly influence advisor choice. However, research has also shown that individuals sometimes utilize communicator characteristics that can be misleading (e.g., Chaiken, 1979). Consequently, it remains to be investigated whether the adaptiveness of advice seeking shown here can be extended to advisor choice.

The present research is the first to look at the sampling of advice. Based on the informational asymmetry hypothesis, we focused on the seeking of advice and its effect on judgment revision. We demonstrated that the breadth and consistency of the advisory estimates influences the degree to which they are integrated with the decision maker’s initial estimate. At the same time, however, the present research neglects the role of the breadth and consistency of the decision maker’s initial knowledge base. We assume that an assessment of the breadth and consistency of the decision maker’s knowledge base will result in a feeling of confidence (Koriat, 2012). Specifically, decision makers who have different pieces of information supporting a given judgment will feel more confident than decision makers with very little or inconsistent information. Presumably, the decision maker’s confidence will affect the amount of advice sampled, such that highly confident decision makers will sample less advice than less confident ones. Moreover, even if decision makers sampled the same amount of information, the degree of integration of the advice with their own estimate should depend on their confidence in their initial estimate. In sum, the present perspective should be complemented by an orientation towards the individual knowledge base in future research.

Although the present work focused on factors instigating sampling behavior, little is known about the factors determining the truncation of the sampling process in advice seeking. Recent work in the domain of decisions from experience assumes that judges set preference thresholds for the choice options (Markant, Pleskac, Diederich, Pachur & Hertwig, 2015). Sampling is terminated as soon as the preference state exceeds a given threshold. This model is in line with the influences of decision importance (i.e., a manipulation of reward magnitude) on sampling frequency in this paradigm (Hau et al., 2008). Such threshold assumptions could also be made for the sampling of advice. For instance, recent research has shown that people are more strongly inclined to receive advice when they are anxious (Gino et al., 2012) or when the task is difficult (Gino & Moore, 2007). Thus, it is conceivable that people compare their present level of knowledge to a threshold of desired knowledge, causing people to sample more information as long as this threshold is not reached and more information is still available. Hence, more frequent sampling could both be due to lower levels of knowledge as well as higher knowledge thresholds. Future research should thus investigate whether such thresholds are set prior to sampling and to what degree these thresholds are sensitive to the sampling process.

Despite the beneficial effects of sampling on advice weighting, averaging proper was not a dominant integration strategy in the present experiments. While the normatively desirable WOA is 0.5 (giving equal weight to one’s own position and the average advice), the largest average WOA achieved in the present experiments was 0.4. Presumably, final information samples were still biased towards the initial judgment. It is hardly conceivable that the sampled advice matched informational richness within individuals’ minds, so that information samples likely remain uneven, especially in terms of qualitative arguments underlying each judgment. Moreover, it is possible that motivational factors also influence advice sampling and weighting. One aspect to consider is that we asked participants to give an initial estimate and, thus, form an opinion. This procedural detail could also be responsible for lingering egocentric tendencies in the present research. Previous research (Koehler & Beauregard, 2006; Yaniv & Choshen-Hillel, 2012) has shown that judgments were less influenced by advice when participants gave an initial judgment. Without an initial judgment, participants were influenced by advice much more strongly (without being aware of that influence). Possibly, the sampling and weighting of advice would then depend mainly on the consensus among the advisors (Budescu & Yu, 2007), and equal weighting strategies would be more frequent.

5.3  Conclusion

The present research introduces a sampling approach to advice taking. Our findings corroborate the importance of the information ecology as a central influencing factor for people’s receptivity for advice (Gino et al., 2012; Gino & Moore, 2007; Sniezek & Buckley, 1995). Three experiments showed substantial sampling of advice that further increased when advice was different from rather than confirming participants’ own positions. Participants integrated diverging advice more strongly into their own judgment. Moreover, participants were sensitive to the information sampled, relying more strongly on advice when it was supported by additionally sampled advisory estimates. In summary, advice taking is sensitive to the information ecology and the human mind appears adaptive when allowed to be.

References

Armstrong, J. S. (2001). Combining judgments. In J. S. Armstrong (Ed.), Principles of Forecasting: A Handbook for Researchers and Practitioners. Norwell, MA: Kluwer Academic Publishers.

Bearden, J. N., & Rapoport, A. (2005). Operations research in experimental psychology. In INFORMS Tutorials in Operations Research (pp. 213–236). Hanover, MD: INFORMS. http://dx.doi.org/10.1287/educ.1053.0011.

Bonaccio, S., & Dalal, R. S. (2006). Advice taking and decision-making: An integrative literature review, and implications for the organizational sciences. Organizational Behavior and Human Decision Processes, 101(2), 127–151. http://dx.doi.org/10.1016/j.obhdp.2006.07.001.

Brunswik, E. (1955). Representative design and probabilistic theory in a functional psychology. Psychological Review, 62(3), 193–217. http://dx.doi.org/10.1037/h0047470.

Budescu, D. V., & Yu, H.-T. (2007). Aggregation of opinions based on correlated cues and advisors. Journal of Behavioral Decision Making, 20(2), 153–177. http://dx.doi.org/10.1002/bdm.

Chaiken, S. (1979). Communicator physical attractiveness and persuasion. Journal of Personality and Social Psychology, 37(8), 1387–1397. http://dx.doi.org/10.1037/0022-3514.37.8.1387.

Clemen, R. (1989). Combining forecasts: A review and annotated bibliography. International Journal of Forecasting, 5(4), 559–583. http://dx.doi.org/10.1016/0169-2070(89)90012-5.

Denrell, J. (2005). Why most people disapprove of me: Experience sampling in impression formation. Psychological Review, 112(4), 951–978. http://dx.doi.org/10.1037/0033-295X.112.4.951.

Einhorn, H., Hogarth, R., & Klempner, E. (1977). Quality of group judgment. Psychological Bulletin, 84(1), 158–172. http://dx.doi.org/10.1037/0033-2909.84.1.158.

Faul, F., Erdfelder, E., Lang, A.-G., & Buchner, A. (2007). G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39(2), 175–191. http://dx.doi.org/10.3758/BF03193146.

Fazio, R. H., Eiser, J. R., & Shook, N. J. (2004). Attitude formation through exploration: Valence asymmetries. Journal of Personality and Social Psychology, 87(3), 293–311. http://dx.doi.org/10.1037/0022-3514.87.3.293.

Fiedler, K. (2000). Beware of samples! A cognitive-ecological sampling approach to judgment biases. Psychological Review, 107(4), 659–676. http://dx.doi.org/10.1037/0033-295X.107.4.659.

Galesic, M., Olsson, H., & Rieskamp, J. (2012). Social sampling explains apparent biases in judgments of social environments. Psychological Science, 23(12), 1515–1523. http://dx.doi.org/10.1177/0956797612445313.

Galton, F. (1907). Vox populi. Nature, 75(7), 450–451. http://dx.doi.org/10.1038/075450a0.

Gino, F. (2008). Do we listen to advice just because we paid for it? The impact of advice cost on its use. Organizational Behavior and Human Decision Processes, 107(2), 234–245. http://dx.doi.org/10.1016/j.obhdp.2008.03.001.

Gino, F., Brooks, A. W., & Schweitzer, M. E. (2012). Anxiety, advice, and the ability to discern: Feeling anxious motivates individuals to seek and use advice. Journal of Personality and Social Psychology, 102(3), 497–512. http://dx.doi.org/10.1037/a0026413.

Gino, F., & Moore, D. (2007). Effects of task difficulty on use of advice. Journal of Behavioral Decision Making, 20(1), 21–35. http://dx.doi.org/10.1002/bdm.

Harvey, N., & Fischer, I. (1997). Taking advice: Accepting help, improving judgment, and sharing responsibility. Organizational Behavior and Human Decision Processes, 70(2), 117–133. http://dx.doi.org/10.1006/obhd.1997.2697.

Hau, R., Pleskac, T. J., Kiefer, J., & Hertwig, R. (2008). The description-experience gap in risky choice: The role of sample size and experienced probabilities. Journal of Behavioral Decision Making, 21(5), 493–518. http://dx.doi.org/10.1002/bdm.

Hertwig, R. (2015). Decisions from experience. In G. Keren & G. Wu (Eds.), Blackwell handbook of judgment and decision making (pp. 239–267). Oxford, UK: Blackwell.

Hertwig, R., Barron, G., Weber, E. U., & Erev, I. (2004). Decisions from experience and the effect of rare events in risky choice. Psychological Science, 15(8), 534–539. http://dx.doi.org/10.1111/j.0956-7976.2004.00715.x.

Hertwig, R., & Erev, I. (2009). The description-experience gap in risky choice. Trends in Cognitive Sciences, 13(12), 517–523. http://dx.doi.org/10.1016/j.tics.2009.09.004.

Hertwig, R., & Pleskac, T. J. (2010). Decisions from experience: Why small samples? Cognition, 115(2), 225–237. http://dx.doi.org/10.1016/j.cognition.2009.12.009.

Herzog, S., & Hertwig, R. (2009). The wisdom of many in one mind improving individual judgments with dialectical bootstrapping. Psychological Science, 20(2), 231–238. http://dx.doi.org/10.1111/j.1467-9280.2009.02271.x.

Judd, C. M., Westfall, J., & Kenny, D. A. (2012). Treating stimuli as a random factor in social psychology: a new and comprehensive solution to a pervasive but largely ignored problem. Journal of Personality and Social Psychology, 103(1), 54–69. http://dx.doi.org/10.1037/a0028347.

Kahneman, D., & Tversky, A. (1973). On the psychology of prediction. Psychological Review, 80(4), 237–251. http://dx.doi.org/10.1037/h0034747.

Kenward, M. G., & Roger, J. H. (1997). Small sample inference for fixed effects from restricted maximuml likelihood. Biometrics, 53(3), 983–997. http://dx.doi.org/10.2307/2533558.

Koehler, D. J., & Beauregard, T. A. (2006). Illusion of confirmation from exposure to another’s hypothesis. Journal of Behavioral Decision Making, 19(1), 61–78. http://dx.doi.org/10.1002/bdm.513.

Koriat, A. (2012). The self-consistency model of subjective confidence. Psychological Review, 119(1), 80–113. http://dx.doi.org/10.1037/a0025648.

Larrick, R. P., Mannes, A. E., & Soll, J. B. (2012). The social psychology of the wisdom of crowds. In J. I. Krueger (Ed.), Frontiers in social psychology: Social judgment and decision making (pp. 227–242). New York, NY: Psychology Press.

Larrick, R., & Soll, J. (2006). Intuitions about combining opinions: Misappreciation of the averaging principle. Management Science, 52(1), 111–127. http://dx.doi.org/10.1287/mnsc..

Lejarraga, T., Hertwig, R., & Gonzalez, C. (2012). How choice ecology influences search in decisions from experience. Cognition, 124(3), 334–342. http://dx.doi.org/10.1016/j.cognition.2012.06.002.

Mannes, A. E. (2009). Are we wise about the wisdom of crowds? The use of group judgments in belief revision. Management Science, 55(8), 1267–1279. http://dx.doi.org/10.1287/mnsc.1090.1031.

Markant, D., Pleskac, T. J., Diederich, A., Pachur, T., & Hertwig, R. (2015). Modeling choice and search in decisions from experience[202F?]: A sequential sampling approach. In R. Dale, C. Jennings, P. Maglio, T. Matlock, D. Noelle, A. Warlaumont, & J. Yoshimi (Eds.), Proceedings of the 37th Annual Conference of the Cognitive Science Society. Austin, TX: Cognitive Science Society (pp. 1512–1517). Austin, TX: Cognitive Science Society.

Masson, M. E. J. (2011). A tutorial on a practical Bayesian alternative to null-hypothesis significance testing. Behavior Research Methods, 43(3), 679–690. http://dx.doi.org/10.3758/s13428-010-0049-5.

McPherson, M., Smith-Lovin, L., & Cook, J. (2001). Birds of a feather: Homophily in social networks. Annual Review of Sociology, 27, 415–444. http://dx.doi.org/10.1146/annurev.soc.27.1.415.

Minson, J. A., Liberman, V., & Ross, L. (2011). Two to tango: effects of collaboration and disagreement on dyadic judgment. Personality & Social Psychology Bulletin, 37(10), 1325–1338. http://dx.doi.org/10.1177/0146167211410436.

Mitra, K., Reiss, M. C., & Capella, L. M. (1999). An examination of perceived risk, information search and behavioral intentions in search, experience and credence services. Journal of Services Marketing, 13(3), 208–228. http://dx.doi.org/10.1108/08876049910273763.

Moussaïd, M., Kämmer, J. E., Analytis, P. P., & Neth, H. (2013). Social influence and the collective dynamics of opinion formation. PloS One, 8(11), 1–8. http://dx.doi.org/10.1371/journal.pone.0078433.

Nisbett, R. E., & Wilson, T. D. (1977). Telling more than we can know: Verbal reports on mental processes. Psychological Review, 84(3), 231–259. http://dx.doi.org/10.1037/0033-295X.84.3.231.

Pachur, T., & Scheibehenne, B. (2012). Constructing preference from experience: The endowment effect reflected in external information search. Journal of Experimental Psychology. Learning, Memory, and Cognition, 38(4), 1108–1116. http://dx.doi.org/10.1037/a0027637.

Phillips, N. D., Hertwig, R., Kareev, Y., & Avrahami, J. (2014). Rivals in the dark: How competition influences search in decisions under uncertainty. Cognition, 133(1), 104–119. http://dx.doi.org/10.1016/j.cognition.2014.06.006.

Rapoport, A., & Tversky, A. (1966). Cost and accessibility of offers as determinants of optional stopping. Psychonomic Science, 4(1), 145–146. http://dx.doi.org/10.3758/BF03342220.

Ravazzolo, F., & Røisland, Ø. (2011). Why do people place lower weight on advice far from their own initial opinion? Economics Letters, 112(1), 63–66. http://dx.doi.org/10.1016/j.econlet.2011.03.032.

Schultze, T., Rakotoarisoa, A.-F., & Schulz-Hardt, S. (2015). Effects of distance between initial estimates and advice on advice utilization. Judgment and Decision Making, 10(2), 144–171. Retrieved from http://journal.sjdm.org/14/141112a/jdm141112a.html.

Schulz-Hardt, S., Frey, D., Lüthgens, C., & Moscovici, S. (2000). Biased information search in group decision making. Journal of Personality and Social Psychology, 78(4), 655–669. http://dx.doi.org/10.1037/0022-3514.78.4.655.

Seale, D. A., & Rapoport, A. (1997). Sequential decision making with relative ranks: An experimental investigation of the “secretary problem.” Organizational Behavior and Human Decision Processes, 69(3), 221–236. http://dx.doi.org/10.1006/obhd.1997.2683.

Simon, H. A. (1956). Rational choice and the structure of the environment. Psychological Review, 63(2), 129–138. http://dx.doi.org/10.1037/h0042769.

Sniezek, J., & Buckley, T. (1995). Cueing and cognitive conflict in judge-advisor decision making. Organizational Behavior and Human Decision Processes, 62(2), 159–174. http://dx.doi.org/10.1006/obhd.1995.1040.

Soll, J. B., & Larrick, R. P. (2009). Strategies for revising judgment: how (and how well) people use others’ opinions. Journal of Experimental Psychology. Learning, Memory, and Cognition, 35(3), 780–805. http://dx.doi.org/10.1037/a0015145.

Soll, J. B., & Mannes, A. E. (2011). Judgmental aggregation strategies depend on whether the self is involved. International Journal of Forecasting, 27(1), 81–102. http://dx.doi.org/10.1016/j.ijforecast.2010.05.003.

Suls, J., Martin, R., & Wheeler, L. (2002). Social comparison: Why, with whom, and with what effect? Current Directions in Psychological Science, 11(5), 159–163. http://dx.doi.org/10.1111/1467-8721.00191.

Tukey, J. W. (1977). Exploratory data analysis. Reading, MA: Addison-Wesley.

Tversky, A., & Koehler, D. (1994). Support theory: a nonextensional representation of subjective probability. Psychological Review, 101(4), 547–567. http://dx.doi.org/10.1037/0033-295X.101.4.547.

Yaniv, I. (2004a). Receiving other people’s advice: Influence and benefit. Organizational Behavior and Human Decision Processes, 93(1), 1–13. http://dx.doi.org/10.1016/j.obhdp.2003.08.002.

Yaniv, I. (2004b). The benefit of additional opinions. Current Directions in Psychological Science, 13(2), 75–79. http://dx.doi.org/10.1111/j.0963-7214.2004.00278.x.

Yaniv, I., & Choshen-Hillel, S. (2012). Exploiting the wisdom of others to make better decisions: Suspending judgment reduces egocentrism and increases accuracy. Journal of Behavioral Decision Making, 25(5), 427–434. http://dx.doi.org/10.1002/bdm.

Yaniv, I., Choshen-Hillel, S., & Milyavsky, M. (2009). Spurious consensus and opinion revision: why might people be more confident in their less accurate judgments? Journal of Experimental Psychology. Learning, Memory, and Cognition, 35(2), 558–63. http://dx.doi.org/10.1037/a0014589.

Yaniv, I., & Kleinberger, E. (2000). Advice taking in decision making: Egocentric discounting and reputation formation. Organizational Behavior and Human Decision Processes, 83(2), 260–281. http://dx.doi.org/10.1006/obhd.2000.2909.

Yaniv, I., & Milyavsky, M. (2007). Using advice from multiple sources to revise and improve judgments. Organizational Behavior and Human Decision Processes, 103(1), 104–120. http://dx.doi.org/10.1016/j.obhdp.2006.05.006.

Appendix: Sequential pair-wise model comparisons


Experiment 1
DV
Predictors
Log Likelihood
Df
(denominator)
p
approx. Bayes factor
WOA
Rnd. Intercepts
    
 
(Participant + Item)
248.57
   
 
+ Distance
300.8
616.45
< .001
1
Δ WOA
Rnd. Intercepts
    
 
(Participant + Item)
248.57
   
 
+ Distance
300.8
616.45
< .001
1






Experiment 2

DV

Predictors
Log Likelihood
Df
(denominator)
p
approx. Bayes factor
Sample size
Rnd. Intercepts
    
 
(Participant + Item)
2317.2
   
 
+ Distance
2287.7
832.74
< .001
1

WOA

Rnd. Intercepts
    
 
(Participant + Item)
-275.57
   
 
+ Distance
-203.42
789.55
< .001
1
 
+ Sample size
-188.28
197.92
< .001
.999
 
+ Distance * Sample size
-186.84
798.58
.089
.127

Δ WOA

Rnd. Intercepts
    
 
(Participant + Item)
–255.73
   
 
+ Distance
–205.96
789.88
< .001
1
 
+ Sample size
–202.31
194.12
.008
.570
 
+ Distance * Sample size
–198.48
798.21
.006
.613

Confidence

Rnd. Intercepts
    
 
(Participant + Item)
–552.38
   
 
+ Distance
–513.25
416.52
< .001
1
 
+ Sample size
–513.25
429.79
.999
.045
 
+ Distance * Sample size
–511.47
416.23
.061
.222


Experiment 3

DV

Predictors
Log Likelihood
Df
(denominator)
p
approx. Bayes factor

Sample size

Rnd. Intercepts
    
 
(Participant + Item)
–3067.8
   
 
+ Distance
–3056.2
1099.53
< .001
.999

WOA

Rnd. Intercepts
    
 
(Participant + Item)
–151.25
   
 
+ Distance
–143.36
981.51
< .001
.988
 
+ Sample size
–137.63
218.68
< .001
.905
 
+ Distance * Sample size
–137.31
992.40
.428
.040

Δ WOA

Rnd. Intercepts
    
 
(Participant + Item)
–166.29
   
 
+ Distance
–162.38
980.50
.005
.608
 
+ Sample size
–154.22
249.76
< .001
.990
 
+ Distance * Sample size
–154.21
990.55
.898
.030

Confidence

Rnd. Intercepts
    
 
(Participant + Item)
–1567.6
   
 
+ Distance
–1550.6
1091.48
< .001
.999
 
+ Sample size
–1540.7
1119.16
< .001
.998
 
+ Distance * Sample size
–1540.7
1087.72
.841
.029
Note. Degrees of freedom are based on the Kenward-Roger approximation (Kenward & Roger, 1997). For each dependent variable (DV), fit of each model was compared to the fit of the model in the row above. Approximate Bayes factors were calculated using the BIC (Masson, 2011). The approx. Bayes factor indicates the conditional probability of the model being true given the data obtained in comparison to the model in the row above.
































*
Fachbereich Psychologie, Eberhard Karls Universität Tübingen, Schleichstr. 4, 72076 Tübingen, Germany, mandy.huetter@uni-tuebingen.de.
#
Fachbereich Psychologie, Eberhard Karls Universität Tübingen, Tübingen, Germany.
Both authors contributed equally to this paper.

The present research was supported by a grant from the German Research Foundation awarded to Mandy Hütter and Klaus Fiedler (HU 1978/3–2). The authors thank Klaus Fiedler and Wolfgang Gaissmaier for valuable comments on a previous draft of this paper.

Copyright: © 2016. The authors license this article under the terms of the Creative Commons Attribution 3.0 License.

1
Due to a programming error, data of five additional participants were not recorded.
2
Materials are available from the authors.
3
For all models containing sample size as a predictor, we recoded sample size by subtracting 1 to account for the fact that each participant received at least one piece of advice on each trial.
4
Due to a programming error, data of two additional participants were not recorded.

This document was translated from LATEX by HEVEA.