Abstract

We find consistent evidence of negative autocorrelation in decision making that is unrelated to the merits of the cases considered in three separate high-stakes field settings: refugee asylum court decisions, loan application reviews, and Major League Baseball umpire pitch calls. The evidence is most consistent with the law of small numbers and the gambler’s fallacy—people underestimating the likelihood of sequential streaks occurring by chance—leading to negatively autocorrelated decisions that result in errors. The negative autocorrelation is stronger among more moderate and less experienced decision makers, following longer streaks of decisions in one direction, when the current and previous cases share similar characteristics or occur close in time, and when decision makers face weaker incentives for accuracy. Other explanations for negatively autocorrelated decisions such as quotas, learning, or preferences to treat all parties fairly are less consistent with the evidence, though we cannot completely rule out sequential contrast effects as an alternative explanation.

I. I ntroduction

Does the sequencing of decisions matter for decision making? Controlling for the quality and merits of a case, we find that the sequence of past decisions matters for the current decision, that is, decision makers exhibit negatively autocorrelated decision making. Using three independent and high-stakes field settings (refugee asylum court decisions in the United States, loan application reviews from a field experiment by Cole, Kanz, and Klapper (2015 ), and Major League Baseball home plate umpire calls on pitches) we show consistent evidence of negatively autocorrelated decision making, despite controlling for case quality, which leads to decision reversals and errors.

In each of the three high-stakes settings, we show that the ordering of case quality is likely to be conditionally random. However, a significant percentage of decisions, more than 5% in some samples, are reversed or erroneous due to negative autocorrelation induced by the behavior of decision makers. The three settings provide independent evidence of negatively autocorrelated decision making across a wide variety of contexts for decision makers in their primary occupations, and across a very large sample size of decisions in some cases. Each field setting offers unique advantages and limitations in terms of data analysis that taken together portray a compelling picture of negatively autocorrelated decision making arising from belief biases.

First, we test whether U.S. judges in refugee asylum cases are more likely to deny (grant) asylum after granting (denying) asylum to the previous applicant. The asylum courts setting offers administrative data on high-frequency judicial decisions with very high stakes for the asylum applicants—a judge’s decision determines whether refugees seeking asylum will be deported from the United States. The setting is also convenient because cases filed within each court (usually a city) are randomly assigned to judges within the court and judges decide on the queue of cases on a first-in-first-out basis. By controlling for the recent approval rates of other judges in the same court, we are able to control for time variation in court-level case quality to ensure that our findings are not generated spuriously by negative autocorrelation in underlying case quality. A limitation of the asylum court data, however, is that we cannot discern whether any individual decision is correct given the case’s merits. We estimate that judges are up to 3.3 percentage points more likely to reject the current case if they approved the previous case. This translates into 2% of decisions being reversed purely due to the sequencing of past decisions, all else being equal. This effect is also stronger following a longer sequence of decisions in the same direction, when judges have “moderate” grant rates close to 50% (calculated excluding the current decision), and when the current and previous cases share similar characteristics or occur close in time (which is suggestive of coarse thinking as in Mullainathan, Schwartzstein, and Shleifer 2008 ). We also find that a judge’s experience mitigates the negative autocorrelation in decision making.

Second, we test whether loan officers are more likely to deny a loan application after approving the previous application by using data from a loan officer field experiment conducted in India by Cole, Kanz, and Klapper (2015) . The field experiment offers controlled conditions in which the order of loan files, and hence their quality, within each session is randomized by the experimenter. In addition, loan officers are randomly assigned to one of three incentive schemes, allowing us to test whether strong pay-for-performance incentives reduce the bias in decision making. The setting is also convenient in that we can observe true loan quality, so we can discern loan officer mistakes. Another advantage of the field experiment setting is that payoffs only depend on accuracy. Loan officers in the experiment are told that their decisions do not affect actual loan origination and they do not face quotas. Therefore, any negative autocorrelation in decisions is unlikely to be driven by concerns about external perceptions, quotas, or by the desire to treat loan applicants in a certain fashion. We find that up to 9% of decisions are reversed due to negative autocorrelation in decision making under the flat incentive scheme among moderate decision makers. The effect is significantly weaker under the stronger incentive schemes and among less moderate decision makers. Across all incentive schemes, the negative autocorrelation is stronger following a streak of two decisions in the same direction. Education, age, experience, and a longer period of time spent reviewing the current loan application reduce the negative autocorrelation in decisions.

Third, we test whether Major League Baseball (MLB) umpires are more likely to call the current pitch a ball after calling the previous pitch a strike and vice versa. An advantage of the baseball umpire data is that it includes precise measures of the three-dimensional location of each pitch. Thus, while pitches may not be randomly ordered over time, we can control for each pitch’s true “quality” or location and measure whether mistakes in calls conditional on a pitch’s true location are negatively predicted by the previous call. We find that umpires are 1.5 percentage points less likely to call a pitch a strike if the previous pitch was called a strike, holding pitch location fixed. This effect more than doubles when the current pitch is close to the edge of the strike zone (meaning it is a less obvious call) and is also significantly larger following two previous calls in the same direction. Put differently, MLB umpires call the same pitches in the exact same location differently depending solely on the sequence of previous calls. We also show that any endogenous changes in pitch location over time are likely to be biases against our findings.

Altogether, we show that negatively autocorrelated decision making in three diverse settings is unrelated to the quality or merits of the cases considered and hence results in decision errors. We explore several potential explanations that could be consistent with negatively autocorrelated decision making, including belief biases such as the gambler’s fallacy and sequential contrast effects, and other explanations such as quotas, learning, and a desire to treat all parties fairly. We find that the evidence across all three settings is most consistent with the gambler’s fallacy and/or sequential contrast effects, and in several tests we are able to reject the other theories.

The “law of small numbers” and the “gambler’s fallacy” are both names for the well-documented tendency of people to overestimate the likelihood that a short sequence will resemble the general population ( Tversky and Kahneman 1971 , 1974 ; Rabin 2002 ; Rabin and Vayanos 2010 ) or underestimate the likelihood of streaks occurring by chance. For example, people often believe that a sequence of coin flips such as “HTHTH” is more likely to occur than “HHHHT” even though each sequence occurs with equal probability. Similarly, people may expect flips of a fair coin to generate high rates of alternation between heads and tails even though streaks of heads or tails often occur by chance. This misperception of random processes can lead to errors in predictions.

In our analysis of decision making under uncertainty, a decision maker who himself suffers from the gambler’s fallacy may similarly believe that streaks of good or bad quality cases are unlikely to occur by chance. Consequently, the decision maker may approach the next case with a prior belief that the case is likely to be positive if she deemed the previous case to be negative, and vice versa. Assuming that decisions made under uncertainty are at least partially influenced by the agent’s priors, these priors then lead to negatively autocorrelated decisions. Similarly, a decision maker who fully understands random processes may still engage in negatively autocorrelated decision making in an attempt to appear fair if she is being evaluated by others, for example promotion committees or voters who suffer from the gambler’s fallacy.

Our analysis differs from the existing literature on the gambler’s fallacy in several ways. First, most of the existing empirical literature examines behavior in gambling or laboratory settings (e.g., Bar-Hillel and Wagenaar 1991 ; Rapoport and Budescu 1992 ; Clotfelter and Cook 1993 ; Terrell 1994 ; Ayton and Fischer 2004 ; Croson and Sundali 2005 ; Asparouhova, Hertzel, and Lemmon 2009 ; Benjamin, Moore, and Rabin 2013 ; Suetens, Galbo-Jorgensen, and Tyran 2015 ) and does not test whether the gambler’s fallacy can bias high-stakes decision making in real-world or field settings such as those involving judges, loan officers, and professional baseball umpires. 1

Second, our analysis differs from the existing literature because we focus on decisions. We define a decision as the outcome of an inference problem using both a prediction and investigation of the current case’s merits. In contrast, the existing literature on the gambler’s fallacy typically focuses on predictions or bets made by agents who do not also assess case merits. Our focus on decisions highlights how greater effort on the part of the decision maker or better availability of information regarding the merits of the current case can reduce errors in decisions even if the decision maker continues to suffer from the gambler’s fallacy when forming predictions. Our findings support this view across all three of our empirical settings.

Finally, we study the behavior of experienced decision makers making decisions in their primary occupations. In some settings, we have variation in incentives to be accurate and show that stronger incentives can reduce the influence of decision biases on decisions. In addition, and in contrast to the laboratory setting as well as other empirical settings studied in the literature, our decision makers see a large sample of cases—many hundreds for an asylum judge and tens of thousands or more for an umpire—which affords us very large samples of decisions.

Other potential alternative and perhaps complementary explanations appear less consistent with the data, though in some cases we cannot completely rule them out. One potential alternative explanation is that decision makers face quotas for the maximum number of affirmative decisions, which could induce negative autocorrelation in decisions since a previous affirmative decision implies that fewer affirmative decisions can be made in the future. However, in all three of our empirical settings, agents do not face explicit quotas or targets. For example, loan officers in the field experiment are only paid based upon accuracy and their decisions do not affect loan origination. Asylum judges are not subject to any explicit quotas or targets and neither are baseball umpires. Nevertheless, one may be concerned about self-imposed quotas or targets. We show that such self-imposed quotas are unlikely to explain our results by contrasting the fraction of recent decisions in one direction with the sequence of such decisions. In a quotas model, the only thing that should matter is the fraction of affirmative decisions. We find, however, that agents negatively react to extreme recency, holding the fraction of recent affirmative decisions constant. That is, if one of the last N decisions was decided in the affirmative, it matters whether the affirmative decision occurred most recently or further back in time. This behavior is consistent with the sequencing of decisions mattering, and is largely inconsistent with self-imposed quotas unless the decision maker also has very limited memory and cannot remember beyond the most recent decision.

Another related potential explanation is a learning model where decision makers do not necessarily face quotas, but they believe that the correct fraction of affirmative decisions should be some level. The decision makers are unsure of where to set the quality bar to achieve that target rate and therefore learn over time, which could lead to negative autocorrelation in decisions. However, baseball umpires should not have a target rate and instead have a quality bar (the official strike zone) that is set for them. Further, decision makers in all of our settings are highly experienced and should therefore have a standard of quality calibrated from many years of experience. As a consequence, they are probably not learning much from their most recent decision or sequence of decisions. In addition, a learning model would not predict a strong negative reaction to the most recent decision either, especially when we also control for their own recent history of decisions, which should be a better proxy for learning.

Another potential interpretation specific to the baseball setting is that umpires may have a preference to be equally “fair” to both teams. Such a desire is unlikely to drive behavior in the asylum judge and loan officers settings because the decision makers review sequences of independent cases that are not part of “teams.” However, a preference to be equally nice to two opposing teams in baseball may lead to negative autocorrelation of umpire calls if, after calling a marginal or difficult-to-call pitch a strike, the umpire chooses to “make it up” to the team at bat by calling the next pitch a ball. We show that such preferences are unlikely to drive our estimates for baseball umpires. Indeed, we find that the negative autocorrelation remains equally strong or stronger when the previous call was obvious (i.e., far from the strike zone boundary) and correct. In these cases, the umpire is less likely to feel guilt about making a particular call because the umpire probably could not have called the pitch any other way (e.g., he and everyone else knew it was the right call to make). Nevertheless, we find strong negative autocorrelation following these obvious and correct calls, suggesting that a desire to undo marginal calls or mistakes is not the sole driver of our results.

Finally, we investigate several potential explanations closely related to the gambler’s fallacy. Since these are empirically indistinguishable, we present them as possible variants of the same theme, though we argue that they may be less plausible in some of our settings. The first is sequential contrast effects (SCE), in which the decision maker’s perception of the quality of the current case is negatively biased by the quality of the previous case ( Pepitone and DiNubile 1976 ; Simonsohn and Loewenstein 2006 ; Simonsohn 2006 ). For example, Bhargava and Fisman (2014) find that speed dating subjects are more likely to reject the next candidate if the previous candidate was very attractive, and Hartzmark and Shue (2016) find that investors perceive today’s earnings news as less impressive if unrelated firms released good earnings news in the previous day. Theoretically, the gambler’s fallacy and SCE can predict the same patterns in decision outcomes. The distinction is mainly with regard to when the subject makes a quality assessment. Under the gambler’s fallacy, a subject who sees a high quality case will predict that the next case is likely to be lower in quality in a probabilistic sense even before seeing the next case, whereas SCE predicts the subject will make a relative comparison after seeing both cases. While the laboratory or prediction markets may be able to separate these two biases, they will be observationally equivalent when looking at only decision outcomes since we cannot observe what is inside a decision maker’s head. Complicating matters further, it may also be the case that the gambler’s fallacy affects the decision maker’s perception of quality, thus leading to a contrast effect. For example, a subject may believe that the next case is likely to be lower in quality after seeing a high-quality case, and this makes him perceive the next case as indeed being less attractive.

We present suggestive evidence that our results are more consistent with a simple gambler’s fallacy model than the SCE model. The SCE may be less likely to occur in the context of baseball because there is a well-defined quality metric (the regulated strike zone), although SCE may still bias perceptions of quality on the margin. In both the asylum court and loan approval settings, we find that decisions are unrelated to continuous quality measures of the previous case after we condition on the previous binary decision. This is consistent with a simple gambler’s fallacy model in which agents expect binary reversals, and less supportive of a SCE model in which agents should react negatively to the continuous quality of the previous case. However, our tests cannot fully reject SCE because we may measure the true quality of the previous case with error.

Another possibility is that the decision maker is rational but cares about the opinions of others, for example promotion committees or voters, who are fooled by randomness. In other words, it is the outside monitors who have the gambler’s fallacy and decision makers merely cater to it. These rational decision makers will choose to make negatively autocorrelated decisions to avoid the appearance of being too lenient or too harsh. While concerns about external perceptions could be an important driver of decisions, they are unlikely to drive the results in the context of loan approval, which is an experimental setting where monetary payouts depend only on accuracy (and loan officers know this) and the ordering of decisions is never reported to an outside party.

Lastly, a related explanation is that agents may prefer to alternate being “mean” and “nice” over short time horizons. This preference could, again, originate from the gambler’s fallacy. A decision maker who desires to be fair may over-infer that she is becoming too negative from a short sequence of “mean” decisions. However, a preference to alternate mean and nice is unlikely to drive behavior in the loan approval setting where loan officers in the experiment know that they do not affect real loan origination (so there is no sense of being mean or nice to loan applicants).

Overall, we show that belief biases possibly stemming from misperceptions of what constitutes a fair process can lead to decision reversals and errors. While we cannot completely distinguish between variants of the gambler’s fallacy and SCE, our evidence is unique to the literature on decision making biases in its breadth, particularly in terms of studying large samples of important decisions made as part of the decision maker’s primary occupation. We also find heterogeneity in the field data that may have useful policy implications. For example, we find that negative autocorrelation in decisions declines if the current and previous case considered are separated by a greater time delay, consistent with experimental results in Gold and Hester (2008) , in which the gambler’s fallacy diminishes in coin flip predictions if the coin is allowed to “rest.” We further find that education, experience, and strong incentives for accuracy can reduce biases in decisions. Finally, our research also contributes to the sizable body of psychology literature using vignette studies of small samples of judges that suggests unconscious heuristics (e.g., anchoring, status quo bias, availability) play a role in judicial decision making (e.g., Guthrie et al. 2000 ). In addition, our results contribute to the theoretical literature on decision making, for example, Bordalo, Gennaioli, and Shleifer (2014) , which models how judges can be biased by legally irrelevant information.

The rest of the article is organized as follows. Section II outlines our empirical framework and discusses how it relates to theory. Section III presents the results for asylum judges, Section IV presents results for the loan officer experiment, and Section V presents the MLB umpire results. Section VI discusses our findings in relation to various theories, including the gambler’s fallacy, while Section VII concludes.

II. E mpirical F ramework and T heory

We describe our empirical framework for testing autocorrelation in sequential decision making across the three empirical contexts and relate it to various theories of decision making.

II.A. Baseline Specification

Our baseline specification simply tests whether the current decision is correlated with the lagged decision, conditional on a set of control variables:
Yit=β0+β1Yi,t1+Controls+ϵit.
Here, Y it represents binary decisions by decision maker i ordered by time t , and β1 measures the change in the probability of making an affirmative decision if the previous decision was affirmative rather than negative. If the ordering of cases is conditionally random, then β1 should be 0 if the quality of the case is the only determinant of decisions. An autocorrelation coefficient, β1 , different from 0 indicates that decision makers are basing their decisions on something other than quality or satisfying an objective function that contains more than just accuracy. Further, β1 < 0 is evidence in favor of negatively autocorrelated decision making unrelated to quality, and β1 > 0 is evidence of positive autocorrelation unrelated to quality. For instance, β1 > 0 might imply some belief in the “hot hand,” that is, seeing a recent streak of positive (or negative) cases implies something about the conditional quality of subsequent cases being higher (lower), even though the conditional quality has not changed. 2β1 < 0 could be consistent with several theories, including the gambler’s fallacy, which we show through a simple extension of Rabin’s (2002) model of the law of small numbers in  Appendix B . The basic idea is that, if the ordering of cases is random and decisions are made only based on case merits, a decision maker’s decision on the previous case should not predict her decision on the next case after controlling for base rates of affirmative decisions. However, a decision maker who misperceives random processes may approach the next decision with a prior belief that the case is likely to be more negative if she deemed the previous case to be positive, and vice versa, leading to negatively autocorrelated decisions. Negative autocorrelation in decisions could also be consistent with sequential contrast effects (SCE), quotas, and learning.

In some empirical settings we can also determine whether any particular decision was a mistake. If we include a dummy for the correct decision as part of Controls , then any nonzero estimate of β1 is evidence of mistakes. In other settings when we cannot definitively determine a mistake, we use β1 to estimate the fraction of decisions that are reversed due to autocorrelated decision making. For example, in the case of negative autocorrelation bias (what we find in the data), the reversal rate is: 2β1a(1a) , where a represents the base rate of affirmative decisions in the data (see  Appendix A for details).

Even if the ordering of cases is random within each decision maker, we face the problem that our estimate of β1 may be biased upward when it is estimated using panel data with heterogeneity across decision makers. The tendency of each decision maker to be positive could be a fixed individual characteristic or slowly changing over time. If we do not control for heterogeneity in the tendency to be positive across decision makers (and possibly within decision makers over time), that would lead to an upward bias for β1 since the previous and current decision are both positively correlated with the decision maker’s unobserved tendency to be positive.

We control for decision maker heterogeneity in several ways. One simple method is to control for heterogeneity using decision maker fixed effects. However, decision maker fixed effects within a finite panel can lead to negative correlation between any two decisions by the same decision maker, which biases toward β1 < 0. To remove this bias, we alternatively control for a moving average of the previous n decisions made by each decision maker, not including the current decision. A benefit of this specification is that it also tests whether or not the decision maker reacts more to the most recent decision, controlling for the average affirmative rate among a set of recent decisions. The drawback of using a moving average is that it may imprecisely measure the tendency of each decision maker to be positive due to small samples and hence be an inadequate control for heterogeneity. Thus, we also control for the decision maker’s average decision in all other settings other than the current decision. 3 In our baseline results we report estimates that control for individual heterogeneity using recent moving averages and leave-out-means because these methods do not bias toward β1 < 0. In the Online Appendix , we show that the results are very similar with the inclusion of decision maker fixed effects, although point estimates tend to be more negative, as expected. Finally, we cluster standard errors by decision maker or decision maker×session as noted.

A second important reason we include control variables is that the sequence of cases considered is not necessarily randomly ordered within each decision maker. To attribute β1 < 0 to decision biases, the underlying quality of the sequence of cases considered, conditional on the set of controls, should not itself be negatively autocorrelated. We discuss for each empirical setting why the sequences of cases appear to be conditionally random. 4

Because many of our regressions include fixed effects (e.g., nationality of asylum applicant), we estimate all specifications using the linear probability model, allowing for clustered standard errors, as suggested by Angrist and Pischke (2008) . However, we recognize there is debate in the econometrics literature concerning the relative merits of various binary dependent variable models. In the Online Appendix , we reestimate all baseline tables using logit and probit models and estimate similar marginal effects.

II.B. Streaks

We also test whether agents are more likely to reverse decisions following a streak of two or more decisions in the same direction. Specifically, we estimate
Yit=β0+β1I(1,1)+β2I(0,1)+β3I(1,0)+Controls+ϵit.
All controls are as described in the baseline specification. Here, I(Yi,t2,Yi,t1) is an indicator representing the two previous decisions. All β ’s measure behavior relative to the omitted group I (0, 0), in which the decision maker has decided negatively two-in-a-row. Tests for streaks can help differentiate among various theories. For example, a basic gambler’s fallacy model predicts that β1 < β2 < β3 < 0. The intuition is that agents mistakenly believe that streaks are unlikely to occur by chance, and longer streaks are particularly unlikely to occur. Following a (1,1) another 1 would constitute a streak of length three, which agents may believe is very unlikely to occur. Similarly, following a (0,1), agents may believe that another 1 is less likely to occur than a 0 because the former would create a streak of length two.

The predictions under an SCE model are less obvious and depend on the specific assumptions of the model. For instance, if agents only contrast current case quality with the case that preceded it, then the decision in time t – 2 should not matter, so we would expect β1 = β2 < β3 = 0. However, if agents contrast the current case with the previous case and, to a lesser degree, the case before that, a SCE model could deliver similar predictions to those of the gamblers’ fallacy model, implying β1 < β2 < β3 < 0.

A quotas model, on the other hand, yields very different predictions. For quotas, β1 should be the most negative, since two affirmative decisions in the past puts a more binding constraint on the quota limit than following only one affirmative decision. However, when the decision maker decided in the affirmative for only one out of the two most recent cases, it should not matter whether the affirmative decision was most recent, hence β2 = β3 . The learning model also does not predict β2 < β3 unless it is a particular form of learning where more weight is given to the most recent decision. We test these various predictions across each of our three settings.

III. A sylum J udges

Our first empirical setting is U.S. asylum court decisions.

III.A. Asylum Judges: Data Description and Institutional Context

The United States offers asylum to foreign nationals who can (i) prove that they have a well-founded fear of persecution in their own countries, and (ii) that their race, religion, nationality, political opinions, or membership in a particular social group is one central reason for the threatened persecution. Decisions to grant or deny asylum have potentially very high stakes for the asylum applicants. An applicant for asylum may reasonably fear imprisonment, torture, or death if forced to return to her home country. For a more detailed description of the asylum adjudication process in the United States, we refer the interested reader to Ramji-Nogales et al. (2007) .

We use administrative data on U.S. refugee asylum cases considered in immigration courts from 1985 to 2013. Judges in immigration courts hear two types of cases: affirmative cases in which the applicant seeks asylum on her own initiative, and defensive cases in which the applicant applies for asylum after being apprehended by the Department of Homeland Security (DHS). Defensive cases are referred directly to the immigration courts while affirmative cases pass a first round of review by asylum officers in the lower level Asylum Offices. For these reasons, a judge may treat these cases differently or, at the very least, categorize them separately. Therefore, we also test whether the negative autocorrelation in decision making is stronger when consecutive cases have the same defensive status (both affirmative or both defensive). 5

The court proceeding at the immigration court level is adversarial and typically lasts several hours. Asylum seekers may be represented by an attorney at their own expense. A DHS attorney cross-examines the asylum applicant and argues before the judge that asylum is not warranted. Those who are denied asylum are ordered to be deported. Decisions to grant or deny asylum made by judges at the immigration court level are typically binding, although applicants may further appeal to the Board of Immigration Appeals.

Our baseline tests explore whether judges are less likely to grant asylum after granting asylum in the previous case. To attribute negative autocorrelation in decisions to a cognitive bias, we first need to show that the underlying quality of the sequence of cases considered by each judge is not itself negatively autocorrelated. Several unique features of the immigration court process help us address this concern. Each immigration court covers a geographic region. Cases considered within each court are randomly assigned to the judges associated with the court (on average, there are eight judges per court). The judges then review the queue of cases following a “first-in-first-out” rule. 6 In other words, judges do not reshuffle the ordering of cases to be considered.

Thus, any time variation in case quality (e.g., a surge in refugees from a hot conflict zone) should originate at the court level. This variation in case quality is likely to be positively autocorrelated on a case-by-case level and therefore a bias against our findings of negative autocorrelation in decisions. We also directly control for time variation in court-level case quality using the recent approval rates of other judges in the same court and test autocorrelation in observable proxies of case quality in the Online Appendix .

Judges have a high degree of discretion in deciding case outcomes. They face no explicit or formally recommended quotas with respect to the grant rate for asylum. They are subject to the supervision of the Attorney General, but otherwise exercise independent judgment and discretion in considering and determining the cases before them. The lack of quotas and oversight is further evidenced by the wide disparities in grant rates among judges associated with the same immigration court Ramji-Nogales et al. (2007) . For example, within the same four-year time period in the court of New York, two judges granted asylum to fewer than 10% of the cases considered, while three other judges granted asylum to over 80% of cases considered. Because many judges display extreme decision rates (close to 0 or 1), we also present subsample analysis excluding extreme judges or limiting to moderate judges (grant rate close to 0.5). We exclude the current observation in the calculation of moderate status, so our results within the moderate subsample will not spuriously generate findings of negative autocorrelation in the absence of true bias.

Judges are appointed by the U.S. Attorney General. In our own data collection of immigration judge biographies, many judges previously worked as immigration lawyers or at the Immigration and Naturalization Service (INS) for some time before they were appointed. Judges typically serve until retirement. Their base salaries are set by a federal pay scale and locality pay is capped at Level III of the Executive Schedule. In 2014, that rate was $167,000. Based on conversations with the president of the National Association of Immigration Judges, no bonuses are granted. See  Appendix C for more background information.

Our data come from a Freedom of Information Act (FOIA) request filed through the Transactional Records Access Clearinghouse (TRAC). We exclude non-asylum-related immigration decisions and focus on applications for asylum, withholding of removal, or protection under the convention against torture (CAT). Applicants typically apply for all three types of asylum protection at the same time. As in Ramji-Nogales et al. (2007) , when an individual has multiple decisions on the same day on these three applications, we use the decision on the asylum application because a grant of asylum allows the applicant all the benefits of a grant of withholding of removal or protection under the withholding-CAT, while the reverse does not hold. In the Online Appendix we redefine a grant of asylum as affirmative if any of the three applications are granted and find qualitatively similar results. 7 We merge TRAC data with our own hand-collected data on judicial biographies. We exclude family members, except the lead family member, because in almost all cases all family members are either granted or denied asylum together.

We also restrict the sample to decisions with known time ordering within day or across days and whose immediate prior decision by the judge is on the same day or previous day, or over the weekend if it is a Monday. Finally, we restrict the sample to judges who review a minimum of 100 cases for a given court and courts with a minimum of 1,000 cases in the data. These exclusions restrict the sample to 150,357 decisions across 357 judges and 45 courthouses.

Table I summarizes our sample of asylum decisions. Judges have long tenures, with a median of eight years of experience. For data on tenure, we only have biographical data on 323 of the 357 judges, accounting for 142,699 decisions. The average case load of a judge is approximately two asylum cases per day. The average grant rate is 0.29; 94% of cases have a lawyer representing the applicant, and 44% are defensive cases initiated by the government. The average family size is 1.21. Time-wise, 47% of hearings occurred between 8 AM and 12 PM, 38% occurred between 12 PM and 2 PM, and 15% occurred between 2 PM and 8 PM. We mark the clock time according to the time that a hearing session opened.

Table I

Asylum Judges: Summary Statistics

MeanMedianStd. Dev.
Number of judges357
Number of courts45
Years since appointment8.4186.06
Daily caseload of judge1.8920.84
Family size1.2110.64
Grant indicator0.29
Nonextreme indicator0.54
Moderate indicator0.25
Lawyer indicator0.939
Defensive indicator0.437
Morning indicator0.47
Lunchtime indicator0.38
Afternoon indicator0.15
MeanMedianStd. Dev.
Number of judges357
Number of courts45
Years since appointment8.4186.06
Daily caseload of judge1.8920.84
Family size1.2110.64
Grant indicator0.29
Nonextreme indicator0.54
Moderate indicator0.25
Lawyer indicator0.939
Defensive indicator0.437
Morning indicator0.47
Lunchtime indicator0.38
Afternoon indicator0.15

Notes . This table presents summary statistics of the asylum judges data that we use in our decision making analysis.

Table I

Asylum Judges: Summary Statistics

MeanMedianStd. Dev.
Number of judges357
Number of courts45
Years since appointment8.4186.06
Daily caseload of judge1.8920.84
Family size1.2110.64
Grant indicator0.29
Nonextreme indicator0.54
Moderate indicator0.25
Lawyer indicator0.939
Defensive indicator0.437
Morning indicator0.47
Lunchtime indicator0.38
Afternoon indicator0.15
MeanMedianStd. Dev.
Number of judges357
Number of courts45
Years since appointment8.4186.06
Daily caseload of judge1.8920.84
Family size1.2110.64
Grant indicator0.29
Nonextreme indicator0.54
Moderate indicator0.25
Lawyer indicator0.939
Defensive indicator0.437
Morning indicator0.47
Lunchtime indicator0.38
Afternoon indicator0.15

Notes . This table presents summary statistics of the asylum judges data that we use in our decision making analysis.

The nonextreme indicator tags decisions for which the average grant rate for the judge for that nationality-defensive category, calculated excluding the current observation, is between 0.2 and 0.8. The moderate indicator tags decisions for which the average grant rate for the judge for that nationality-defensive category, excluding the current observation, is between 0.3 and 0.7. 8

III.B. Asylum Judges: Empirical Specification Details

Observations are at the judge × case order level; Y it is an indicator for whether asylum is granted. Cases are ordered within day and across days. Our regression sample includes observations in which the lagged case was viewed on the same day or the previous workday (e.g., we include the observation if the current case is viewed on Monday and the lagged case was viewed on Friday), and for which we know the ordering of cases considered within the same day. 9

Control variables in the regressions include, unless otherwise noted, a set of dummies for the number of affirmative decisions over the past five decisions (excluding the current decision) of the judge. This controls for recent trends in grants, case quality, or judge’s mood. We also include a set of dummies for the number of grant decisions over the past five decisions across other judges (excluding the current judge) in the same court. This controls for recent trends in grants, case quality, or mood at the court level. To control for longer-term trends in judge- and court-specific grant rates, we control for the judge’s leave-out-mean grant rate for the relevant nationality × defensive category, calculated excluding the current observation. We also control for the court’s average grant rate for the relevant nationality × defensive category, calculated by excluding the judge associated with the current observation. In our baseline results, we do not include judge fixed effects because they mechanically induce a small degree of negative correlation between Y it and Y i,t−1 . In the Online Appendix we report results using judge’s fixed effects and obtain similar results with slightly more negative coefficient estimates, as expected. Finally, we control for the characteristics of the current case: presence-of-lawyer indicator, family size, nationality × defensive status fixed effects, and time-of-day fixed effects (morning/lunchtime/afternoon). Including time-of-day fixed effects is designed to control for other factors such as hunger or fatigue, which may influence judicial decision making (as shown in the setting of parole judges by Danziger et al. 2011 ).

III.C. Asylum Judges: Results

In Table II , column (1), we present results for the full sample of case decisions and find that judges are 0.5 percentage points less likely to grant asylum to the current applicant if the previous decision was an approval rather than a denial, all else being equal. In the remaining columns we focus on cumulative subsamples in which the magnitude of the negative autocorrelation increases substantially. First, the asylum data cover a large number of judges who tend to grant or deny asylum to almost all applicants from certain nationalities. More extreme judges necessarily exhibit less negative autocorrelation in their decisions. In column (2) of Table II , we restrict the sample to nonextreme judge observations (where nonextreme is calculated excluding the current decision). The extent of negative autocorrelation doubles to 1.1 percentage points.

Table II

Asylum Judges: Baseline Results

Grant Asylum Dummy
(1)(2)(3)(4)(5)
Lag grant−0.00544* −0.0108 *** −0.0155 ** −0.0326 ***
(0.00308)(0.00413)(0.00631)(0.00773)
β1 : Lag grant-grant −0.0549 ***
(0.0148)
β2 : Lag deny-grant −0.0367 **
(0.0171)
β3 : Lag grant-deny −0.00804
(0.0157)
p -value: β1 = β2 = β3.0507
p -value: β1=β2.290
p -value: β1=β3.0214
p -value: β2=β3.0503
Exclude extreme judgesNoYesYesYesYes
Same-day casesNoNoYesYesYes
Same defensive casesNoNoNoYesYes
N150,35780,73336,38923,99010,652
R20.3740.2070.2230.2280.269
Grant Asylum Dummy
(1)(2)(3)(4)(5)
Lag grant−0.00544* −0.0108 *** −0.0155 ** −0.0326 ***
(0.00308)(0.00413)(0.00631)(0.00773)
β1 : Lag grant-grant −0.0549 ***
(0.0148)
β2 : Lag deny-grant −0.0367 **
(0.0171)
β3 : Lag grant-deny −0.00804
(0.0157)
p -value: β1 = β2 = β3.0507
p -value: β1=β2.290
p -value: β1=β3.0214
p -value: β2=β3.0503
Exclude extreme judgesNoYesYesYesYes
Same-day casesNoNoYesYesYes
Same defensive casesNoNoNoYesYes
N150,35780,73336,38923,99010,652
R20.3740.2070.2230.2280.269

Notes. This table tests whether the decision to grant asylum to the current applicant is related to the decision to grant asylum to the previous applicant. Observations are at the judge × case level. Observations are restricted to decisions that occurred within one day or weekend after the previous decision. Column (2) excludes extreme judge observations (the average grant rate for the judge for the nationality-defensive category of the current case, calculated excluding the current observation, is below 0.2 or above 0.8). Column (3) further restricts the sample to decisions that follow another decision on the same day. Column (4) further restricts the sample to decisions in which the current and previous case have the same defensive status (both defensive or both affirmative). Column (5) tests how judges react to streaks in past decisions. The sample is further restricted to observations in which the current, previous, and previous-previous cases share the same defensive status. To retain sample size, we keep the restriction that the current and previous case must occur on the same day, but allow the previous-previous case to occur on the previous day. Lag grant-grant is an indicator for whether the judge approved the two most recent asylum cases. Lag deny-grant is an indicator for whether the judge granted the most recent case and denied the case before that. Lag grant-deny is an indicator for whether the judge denied the most recent case and granted the case before that. The omitted category is Lag deny-deny . All specifications include the following controls: indicator variables for the number of grants out of the judge’s previous five decisions (excluding the current decision); indicator variables for the number of grants within the five most recent cases in the same court, excluding those of the judge corresponding to the current observation; the judge’s average grant rate for the relevant nationality × defensive category (excluding the current observation); the court’s average grant rate for the relevant nationality × defensive category (excluding the current judge); presence of lawyer representation indicator; family size; nationality × defensive fixed effects, and time of day fixed effects (morning/lunchtime/afternoon). Standard errors are clustered by judge. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table II

Asylum Judges: Baseline Results

Grant Asylum Dummy
(1)(2)(3)(4)(5)
Lag grant−0.00544* −0.0108 *** −0.0155 ** −0.0326 ***
(0.00308)(0.00413)(0.00631)(0.00773)
β1 : Lag grant-grant −0.0549 ***
(0.0148)
β2 : Lag deny-grant −0.0367 **
(0.0171)
β3 : Lag grant-deny −0.00804
(0.0157)
p -value: β1 = β2 = β3.0507
p -value: β1=β2.290
p -value: β1=β3.0214
p -value: β2=β3.0503
Exclude extreme judgesNoYesYesYesYes
Same-day casesNoNoYesYesYes
Same defensive casesNoNoNoYesYes
N150,35780,73336,38923,99010,652
R20.3740.2070.2230.2280.269
Grant Asylum Dummy
(1)(2)(3)(4)(5)
Lag grant−0.00544* −0.0108 *** −0.0155 ** −0.0326 ***
(0.00308)(0.00413)(0.00631)(0.00773)
β1 : Lag grant-grant −0.0549 ***
(0.0148)
β2 : Lag deny-grant −0.0367 **
(0.0171)
β3 : Lag grant-deny −0.00804
(0.0157)
p -value: β1 = β2 = β3.0507
p -value: β1=β2.290
p -value: β1=β3.0214
p -value: β2=β3.0503
Exclude extreme judgesNoYesYesYesYes
Same-day casesNoNoYesYesYes
Same defensive casesNoNoNoYesYes
N150,35780,73336,38923,99010,652
R20.3740.2070.2230.2280.269

Notes. This table tests whether the decision to grant asylum to the current applicant is related to the decision to grant asylum to the previous applicant. Observations are at the judge × case level. Observations are restricted to decisions that occurred within one day or weekend after the previous decision. Column (2) excludes extreme judge observations (the average grant rate for the judge for the nationality-defensive category of the current case, calculated excluding the current observation, is below 0.2 or above 0.8). Column (3) further restricts the sample to decisions that follow another decision on the same day. Column (4) further restricts the sample to decisions in which the current and previous case have the same defensive status (both defensive or both affirmative). Column (5) tests how judges react to streaks in past decisions. The sample is further restricted to observations in which the current, previous, and previous-previous cases share the same defensive status. To retain sample size, we keep the restriction that the current and previous case must occur on the same day, but allow the previous-previous case to occur on the previous day. Lag grant-grant is an indicator for whether the judge approved the two most recent asylum cases. Lag deny-grant is an indicator for whether the judge granted the most recent case and denied the case before that. Lag grant-deny is an indicator for whether the judge denied the most recent case and granted the case before that. The omitted category is Lag deny-deny . All specifications include the following controls: indicator variables for the number of grants out of the judge’s previous five decisions (excluding the current decision); indicator variables for the number of grants within the five most recent cases in the same court, excluding those of the judge corresponding to the current observation; the judge’s average grant rate for the relevant nationality × defensive category (excluding the current observation); the court’s average grant rate for the relevant nationality × defensive category (excluding the current judge); presence of lawyer representation indicator; family size; nationality × defensive fixed effects, and time of day fixed effects (morning/lunchtime/afternoon). Standard errors are clustered by judge. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

In column (3) of Table II , we further restrict the sample to cases that follow another case on the same day (rather than the previous day). We find stronger negative autocorrelation within same-day cases. The stronger negative autocorrelation when two consecutive cases occur more closely in time is broadly consistent with saliency and the gambler’s fallacy decision making model because more recent cases may be more salient and lead to stronger expectations of reversals. These results are also consistent with experimental results in Gold and Hester (2008) , which finds that laboratory subjects who are asked to predict coin flips exhibit less gambler’s fallacy after an interruption when the coin “rests.” The higher saliency of more recent cases could also be consistent with stronger SCE, but is less likely to be consistent with a quotas constraint unless judges self-impose daily but not overnight or multiday quotas.

Column (4) of Table II restricts the sample further to cases in which the current and previous case have the same defensive status. Individuals seeking asylum affirmatively, where the applicant initiates, can be very different from those seeking asylum defensively, where the government initiates. In affirmative cases, applicants typically enter the country legally and are applying to extend their stay. In defensive cases, applicants often have entered the country illegally and have been detained at the border or caught subsequently. Judges may view these scenarios to be qualitatively different; the negative autocorrelation increases to 3.3 percentage points.

Hence, from an unconditional 0.5 percentage points, the negative autocorrelation increases sixfold to 3.3 percentage points if we examine moderate judges on same-day cases with the same defensive status. Using the estimate in column (4) of Table II within the sample of nonextreme, same-day, same defensive cases, the coefficient implies that 1.6% of asylum decisions would have been reversed absent the negative autocorrelation in decision making. Table A.I in the Online Appendix reports the extent of the negative autocorrelation among each omitted sample and presents formal statistical tests for whether the estimates in each cumulative subsample significantly differ from one another. We find that the negative autocorrelation among extreme-judge and different-defensive-status subsamples are close to zero and significantly differ from the nonomitted samples. However, the negative autocorrelation is economically substantial even across consecutive days (with insignificant differences), although the effect size doubles when the judge considers cases within the same day.

Finally, column (5) of Table II tests whether decisions are more likely to be reversed following streaks of previous decisions. After a streak of two grants, judges are 5.5 percentage points less likely to grant asylum relative to decisions following a streak of two denials. Following a deny then grant decision, judges are 3.7 percentage points less likely to grant asylum relative to decisions following a streak of two denials, whereas behavior following a grant then deny decision is insignificantly different from behavior following a streak of two denials. In terms of our empirical framework introduced in Section II.B, we find that β1 < β2 < β3 < 0. A formal statistical test of the difference in the β ’s appears at the bottom of the table, where we reject that the betas are all equal and that β2 = β3 . (The only insignificant difference is between β1 and β2 , though β1 has, as predicted, a more negative point estimate.) These results are consistent with the gambler’s fallacy affecting decisions and inconsistent with a basic quotas model. Moreover, the magnitudes are economically significant. Using the largest point estimate following a streak of two grant decisions, a 5.5 percentage point decline in the approval rate represents a 19% reduction in the probability of approval relative to the base rate of approval of 29%.

We report robustness tests of our findings in the Online Appendix . Table A.II reports results using logit and probit models. The economic magnitudes are similar. Table A.III reports results using judge fixed effects. As expected, the coefficients are slightly more negative due to a mechanical negative autocorrelation between any two decisions by the same judge induced by the fixed effects. However, the bias appears to be small and the coefficient estimates are of similar magnitude to our baseline results that control for a moving average of each judge’s past five decisions, as well as her leave-out-mean grant rate. In addition, the precision of the estimates does not change much between the two specifications, suggesting that controlling for heterogeneity using the moving average of a judge’s decisions and her leave-out-mean instead of judge fixed effects yields similar identification despite the former containing more measurement error. Table A.IV presents results for an alternative definition of the granting of asylum, where instead of using the asylum grant decision, we code a decision as a grant if the judge granted any of the asylum, withholding of removal, or protection under the U.S. CAT applications. The results are very consistent with slightly smaller point estimates.

Finally, a potential concern with the sample split among moderate and extreme decision makers is that we may mechanically measure stronger negative autocorrelation among moderates. We emphasize that because we do not use the current observation in the calculation of whether a decision maker is moderate, restricting the sample to moderates does not mechanically generate β1 < 0 if the true autocorrelation is 0 (e.g., a judge who decides based on random coin flips would be classified as a moderate, but would display zero autocorrelation). However, another potential issue that could mechanically generate greater measured negative autocorrelation for moderate judges is our use of a binary statistical model. The autocorrelation of a binary variable is biased away from 1 and the size of the bias increases as the base rate of decisions gets closer to 0 or one. This may lead us to estimate a lower degree for negative autocorrelation for “extreme” decision makers. One method to address this issue is to use the tetrachoric correlation, which models binary variables as functions of continuous (bivariate normal) latent variables. The bivariate probit model extends the tetrachoric correlation to allow for additional control variables. Using the bivariate probit model, Table A.V shows that there is strong negative autocorrelation in decisions for moderate decision makers that is statistically significant and of similar economic magnitude as those from our baseline regressions. Conversely, for extreme decision makers, there is no evidence of any autocorrelation. These results match our estimates from the linear probability, logit, and probit regressions and indicate that the potential mechanical correlation coming from binary models is not driving our results.

Table III explores additional heterogeneity across judges and cases. In this and subsequent tables, we restrict our analysis to the sample defined in column (4) of Table II —observations for which the current and previous case were decided by nonextreme judges on the same day and with the same defensive status. Column (1) of Table III shows that the reduction in the probability of approval following a previous grant is 4.2 percentage points greater when the previous decision corresponds to an application with the same nationality as the current applicant. While there is significant negative autocorrelation when sequential cases correspond to different applicant nationalities, the negative autocorrelation is three times larger when the two cases correspond to the same nationality. This suggests that the negative autocorrelation in decisions may be tied to saliency and coarse thinking. Judges are more likely to engage in negatively autocorrelated decision making when the previous case considered is similar in terms of characteristics, in this case nationality. These results are consistent with the stronger autocorrelation also found when the previous case occurred close in time with the current case or shared the same defensive status (as shown in Table II ).

Table III

Asylum Judges: Heterogeneity

Grant Asylum Dummy
(1)(2)(3)(4)
Lag grant −0.0196 **0.00180 −0.0484 *** −0.0553 ***
(0.00801)(0.00900)(0.0115)(0.0115)
Same nationality 0.0336 ***
(0.0108)
Lag grant × same nationality −0.0421 ***
(0.0126)
Moderate judge 0.0326 ***
(0.0116)
Lag grant × moderate judge −0.0700 ***
(0.0136)
Experienced judge0.01380.0253*
(0.0106)(0.0140)
Lag grant × experienced judge 0.0327 ** 0.0456 ***
(0.0152)(0.0156)
Judge FENoNoNoYes
N23,99023,99022,96522,965
R20.2290.2290.2290.247
Grant Asylum Dummy
(1)(2)(3)(4)
Lag grant −0.0196 **0.00180 −0.0484 *** −0.0553 ***
(0.00801)(0.00900)(0.0115)(0.0115)
Same nationality 0.0336 ***
(0.0108)
Lag grant × same nationality −0.0421 ***
(0.0126)
Moderate judge 0.0326 ***
(0.0116)
Lag grant × moderate judge −0.0700 ***
(0.0136)
Experienced judge0.01380.0253*
(0.0106)(0.0140)
Lag grant × experienced judge 0.0327 ** 0.0456 ***
(0.0152)(0.0156)
Judge FENoNoNoYes
N23,99023,99022,96522,965
R20.2290.2290.2290.247

Notes . Column (1) tests whether the gambler’s fallacy is stronger when the previous decision concerned an applicant with the same nationality as the current applicant. Column (2) tests whether the gambler’s fallacy is stronger among moderate judge observations (the average grant rate for the judge for the nationality-defensive category of the current case, calculated excluding the current observation, is between 0.3 and 0.7). Columns (3) and (4) test whether the gambler’s fallacy declines with experience. Experienced in an indicator for whether the judge, at the time when the case was decided, had more than the median experience in the sample (8 years). Column (4) adds judge fixed effects, so the interaction term measures the within-judge effect of experience. All other variables and restrictions are as described in Table II , column (3). Standard errors are clustered by judge. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table III

Asylum Judges: Heterogeneity

Grant Asylum Dummy
(1)(2)(3)(4)
Lag grant −0.0196 **0.00180 −0.0484 *** −0.0553 ***
(0.00801)(0.00900)(0.0115)(0.0115)
Same nationality 0.0336 ***
(0.0108)
Lag grant × same nationality −0.0421 ***
(0.0126)
Moderate judge 0.0326 ***
(0.0116)
Lag grant × moderate judge −0.0700 ***
(0.0136)
Experienced judge0.01380.0253*
(0.0106)(0.0140)
Lag grant × experienced judge 0.0327 ** 0.0456 ***
(0.0152)(0.0156)
Judge FENoNoNoYes
N23,99023,99022,96522,965
R20.2290.2290.2290.247
Grant Asylum Dummy
(1)(2)(3)(4)
Lag grant −0.0196 **0.00180 −0.0484 *** −0.0553 ***
(0.00801)(0.00900)(0.0115)(0.0115)
Same nationality 0.0336 ***
(0.0108)
Lag grant × same nationality −0.0421 ***
(0.0126)
Moderate judge 0.0326 ***
(0.0116)
Lag grant × moderate judge −0.0700 ***
(0.0136)
Experienced judge0.01380.0253*
(0.0106)(0.0140)
Lag grant × experienced judge 0.0327 ** 0.0456 ***
(0.0152)(0.0156)
Judge FENoNoNoYes
N23,99023,99022,96522,965
R20.2290.2290.2290.247

Notes . Column (1) tests whether the gambler’s fallacy is stronger when the previous decision concerned an applicant with the same nationality as the current applicant. Column (2) tests whether the gambler’s fallacy is stronger among moderate judge observations (the average grant rate for the judge for the nationality-defensive category of the current case, calculated excluding the current observation, is between 0.3 and 0.7). Columns (3) and (4) test whether the gambler’s fallacy declines with experience. Experienced in an indicator for whether the judge, at the time when the case was decided, had more than the median experience in the sample (8 years). Column (4) adds judge fixed effects, so the interaction term measures the within-judge effect of experience. All other variables and restrictions are as described in Table II , column (3). Standard errors are clustered by judge. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Columns (2) and (3) of Table III show that moderate judges and judges with less experience display stronger negative autocorrelation in decisions. Judges who have less than the median experience in the sample (eight years) display stronger negative autocorrelation. The fourth column repeats the regression including judge fixed effects. We find that experience is also associated with significantly less negatively autocorrelated decisions for a given judge over time. 10

Because we measure decisions rather than predictions, reduced negative autocorrelation does not necessarily imply that experienced judges are more sophisticated in terms of understanding random processes. Both experienced and inexperienced judges could suffer equally from the gambler’s fallacy in terms of forming prior beliefs regarding the quality of the current case. However, experienced judges may draw, or believe they draw, more informative signals regarding the quality of the current case. If so, experienced judges will rely more on the current signal and less on their prior beliefs, leading to reduced negative autocorrelation in decisions.

Finally, we present evidence supporting the validity of our analysis. To attribute negative autocorrelation in decisions to cognitive biases and not case quality, we show that the underlying quality of the sequence of cases considered by each judge is not itself negatively autocorrelated. Within a court, the incoming queue of cases is randomly assigned to judges associated with that court, and the judges review the queue of cases following a first-in-first-out rule. Therefore, time variation in case quality (e.g., a surge in refugees from a hot conflict zone) should originate at the court level and is likely to be positively autocorrelated on a case-by-case level. We support this assumption in Table A.VI in the Online Appendix . We find that case quality does not appear to be negatively autocorrelated in terms of observable proxies for quality. However, our identifying assumption requires that autocorrelation in unobserved aspects of case quality is also not negative.

IV. L oan O fficers

Our second empirical setting examines loan officers making loan application decisions.

IV.A. Loan Officers: Data Description and Institutional Context

We use field experiment data collected by Cole et al. (2015) . 11 The original intent of the experiment was to explore how various incentive schemes affect the quality of loan officers’ screening of loan applications. The framed field experiment was designed to closely match the underwriting process for unsecured small enterprise loans in India. Real loan officers were recruited for the experiment from the active staff of several commercial banks. These loan officers had an average of 10 years of experience in the banking sector. In the field experiment, the loan officers screen real, previously processed loan applications. Each loan file contained all the information available to the bank at the time the loan was first evaluated.

Each loan officer participated in at least one evaluation session. In each session, the loan officer screened six randomly ordered loan files and decided whether to approve or reject the loan application. Because the loan files corresponded to actual loans previously reviewed by banks in India, the files can be classified by the experimenter as performing or nonperforming. Performing loan files were approved and did not default during the actual life of the loan. Nonperforming loans were either rejected by the bank in the loan application process or were approved but defaulted during the actual life of the loan. Loan officers in the experiment were essentially paid based upon their ability to correctly classify the loans as performing (by approving them) or nonperforming (by rejecting them). In our sample, loan officers correctly classified loans approximately 65% of the time. The percentage of performing loans they approved is 78%, while the percentage of nonperforming loans they approved is 62%, which shows they exhibited some ability to sort loans. Overall, the tetrachoric correlation between the binary variables, loan approval, and loan performance (1 = performing, 0 = nonperforming) is 0.29 and significantly different from random chance.

Participants in each session were randomly assigned to one of three incentive schemes, which offered payouts of the form [wP,wD,w¯] , where w P is the payout in rupees for approving a performing loan, w D is the payout for approving a nonperforming loan, and w¯ is the payout for rejecting a loan (regardless of actual loan performance). Beyond direct monetary compensation, participants may have also been motivated by reputational concerns. Loan officers were sent to the experiment by their home bank and the experiment was conducted at a loan officer training college. At the end of the experiment, loan officers received a completion certificate and a document summarizing their overall accuracy rate. The loan officers were told that this summary document would only report their overall accuracy without reporting the ordering of their specific decisions and associated accuracy. Thus, loan officers might have been concerned that their home bank would evaluate these documents and therefore were motivated by factors other than direct monetary compensation. Importantly, however, the approval rate and the ordering of decisions was never reported. Therefore, there was no incentive to negatively autocorrelate decisions for any reason.

In the “flat” incentive scheme, payoffs take the form [20,20,0] , so loan officers had monetary incentives to approve loans regardless of loan quality. However, loan officers may have had reputational concerns that led them to exert effort and reject low quality loan files even within the flat incentive scheme. 12 In the “stronger” incentive scheme, payouts take the form [20,0,10] , so loan officers faced a monetary incentive to reject nonperforming loans. In the “strongest” incentive scheme, payouts take the form [50,100,0] , so approval of nonperforming loans was punished by deducting from an endowment given to the loan officers at the start of the experiment. The payouts across the incentive treatments were chosen to be approximately equal to 1.5 times the hourly wage of the median participant in the experiment.

The loan officers were informed of their incentive scheme. They were also made aware that their decision on the loans would affect their personal payout from the experiment but would not affect actual loan origination (because these were real loan applications that had already been evaluated in the past). Finally, the loan officers were told that the loan files were randomly ordered and that they were drawn from a large pool of loans, of which approximately two-thirds were performing loans. Because the loan officers reviewed loans in an electronic system, they could not review the loans in any order other than the order presented. They faced no time limits or quotas.

Table IV presents summary statistics for our data sample. The data contains information on loan officer background characteristics such as age, education, and the time spent by the loan officer evaluating each loan file. Observations are at the loan officer × loan file level. We consider an observation to correspond to a moderate loan officer if the average approval rate of loans by the loan officer in other sessions (not including the current session) within the same incentive scheme is between 0.3 and 0.7.

Table IV

Loan Officers: Summary Statistics

Full SampleFlat IncentivesStrong IncentivesStrongest Incentives
Loan officer × loan observations9,1681,3326,3361,470
Loan officers1887618189
Sessions (6 loans per session)1,5282221,056245
MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)

Fraction of loans performing0.650.660.650.65
Fraction loans approved0.730.810.720.68
Fraction decisions correct0.640.660.640.64
Fraction performing loans approved0.780.860.770.75
Fraction nonperforming loans approved0.620.720.610.55
Tetrachoric correlation 0.29 ***(0.017) 0.29 ***(0.047) 0.28 ***(0.020) 0.32 ***(0.040)
Fraction moderate0.340.250.360.36
Loan rating (0–1)0.710.160.740.160.700.160.730.15
Fraction grad school education0.290.300.290.26
Time viewed (minutes)3.482.772.842.113.702.963.092.23
Age (years)37.7011.9537.3711.9338.6012.1734.1310.21
Experience in banking (years)9.549.549.679.419.859.768.098.50
Full SampleFlat IncentivesStrong IncentivesStrongest Incentives
Loan officer × loan observations9,1681,3326,3361,470
Loan officers1887618189
Sessions (6 loans per session)1,5282221,056245
MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)

Fraction of loans performing0.650.660.650.65
Fraction loans approved0.730.810.720.68
Fraction decisions correct0.640.660.640.64
Fraction performing loans approved0.780.860.770.75
Fraction nonperforming loans approved0.620.720.610.55
Tetrachoric correlation 0.29 ***(0.017) 0.29 ***(0.047) 0.28 ***(0.020) 0.32 ***(0.040)
Fraction moderate0.340.250.360.36
Loan rating (0–1)0.710.160.740.160.700.160.730.15
Fraction grad school education0.290.300.290.26
Time viewed (minutes)3.482.772.842.113.702.963.092.23
Age (years)37.7011.9537.3711.9338.6012.1734.1310.21
Experience in banking (years)9.549.549.679.419.859.768.098.50

Notes. This table presents summary statistics on the sample of loan officers obtained from Cole et al. (2015) that we use in our decision making analysis. The tetrachoric correlation is the correlation between the loan officer approval decision in the experiment and the indicator for whether the loan is a performing loan. The loan rating represents the continuous quality score loan officers assigned to each loan file during the experiment. This loan rating ranges from 0 to 100 and has been scaled down to be between 0 and 1.

Table IV

Loan Officers: Summary Statistics

Full SampleFlat IncentivesStrong IncentivesStrongest Incentives
Loan officer × loan observations9,1681,3326,3361,470
Loan officers1887618189
Sessions (6 loans per session)1,5282221,056245
MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)

Fraction of loans performing0.650.660.650.65
Fraction loans approved0.730.810.720.68
Fraction decisions correct0.640.660.640.64
Fraction performing loans approved0.780.860.770.75
Fraction nonperforming loans approved0.620.720.610.55
Tetrachoric correlation 0.29 ***(0.017) 0.29 ***(0.047) 0.28 ***(0.020) 0.32 ***(0.040)
Fraction moderate0.340.250.360.36
Loan rating (0–1)0.710.160.740.160.700.160.730.15
Fraction grad school education0.290.300.290.26
Time viewed (minutes)3.482.772.842.113.702.963.092.23
Age (years)37.7011.9537.3711.9338.6012.1734.1310.21
Experience in banking (years)9.549.549.679.419.859.768.098.50
Full SampleFlat IncentivesStrong IncentivesStrongest Incentives
Loan officer × loan observations9,1681,3326,3361,470
Loan officers1887618189
Sessions (6 loans per session)1,5282221,056245
MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)MeanStd. Dev. (Std. Err.)

Fraction of loans performing0.650.660.650.65
Fraction loans approved0.730.810.720.68
Fraction decisions correct0.640.660.640.64
Fraction performing loans approved0.780.860.770.75
Fraction nonperforming loans approved0.620.720.610.55
Tetrachoric correlation 0.29 ***(0.017) 0.29 ***(0.047) 0.28 ***(0.020) 0.32 ***(0.040)
Fraction moderate0.340.250.360.36
Loan rating (0–1)0.710.160.740.160.700.160.730.15
Fraction grad school education0.290.300.290.26
Time viewed (minutes)3.482.772.842.113.702.963.092.23
Age (years)37.7011.9537.3711.9338.6012.1734.1310.21
Experience in banking (years)9.549.549.679.419.859.768.098.50

Notes. This table presents summary statistics on the sample of loan officers obtained from Cole et al. (2015) that we use in our decision making analysis. The tetrachoric correlation is the correlation between the loan officer approval decision in the experiment and the indicator for whether the loan is a performing loan. The loan rating represents the continuous quality score loan officers assigned to each loan file during the experiment. This loan rating ranges from 0 to 100 and has been scaled down to be between 0 and 1.

IV.B. Loan Officers: Empirical Specification Details

We define Y it as an indicator for whether the loan is approved. Loans are ordered within a session. Our sample includes observations for which the lagged loan was viewed in the same session (so we exclude the first loan viewed in each session because we do not expect reliance on the previous decision to necessarily operate across sessions, which are often separated by multiple days). In some specifications, we split the sample by incentive scheme type, that is, flat, strong, or strongest.

We control for heterogeneity in mean approval rates at the loan officer × incentive scheme level using the mean loan officer approval rate within each incentive treatment (calculated excluding the six observations corresponding to the current session).We also include an indicator for whether the loan officer has ever approved all six loans in another session within the same incentive treatment, to control for the fact that these types of loan officers are likely to have particularly high approval rates in the current session. Finally, we include an indicator for whether the current session is the only session attended by the loan officer within the incentive treatment (if so, the first two control variables cannot be calculated and are set to 0). Because the loan officer field experiment data is limited in size and each session consists of only six loan decisions, we do not control for a moving average of each loan officer’s average decision rate over the past five decisions within the session (as we do in the asylum judge setting). In the Online Appendix , we also present results controlling for loan officer fixed effects.

IV.C. Loan Officers: Results

Table V , column (1) shows that loan officers are 8 percentage points less likely to approve the current loan if they approved the previous loan when facing flat incentives. This implies that 2.6% of decisions are reversed due to the sequencing of applications. These effects become much more muted and insignificantly different from zero in the other incentive schemes when loan officers face stronger monetary incentives for accuracy, as shown by the other interaction coefficients in column (1). A test for equality of the coefficients indicate significantly different effects across the three incentive schemes. In column (2), we control for the true quality of the current loan file. Therefore, all reported coefficients represent mistakes on the part of the loan officer. After including this control variable, we find quantitatively similar results, indicating that the negatively autocorrelated decision making results in decision errors.

Table V

Loan Officers: Baseline Results

Approve Loan Dummy
(1)(2)(3)(4)
Lag approve × flat incent −0.0814 ** −0.0712 ** −0.225 *** −0.228 ***
(0.0322)(0.0323)(0.0646)(0.0639)
Lag approve × stronger incent−0.00674−0.00215 −0.0525 ** −0.0484 **
(0.0134)(0.0134)(0.0215)(0.0214)
Lag approve × strongest incent0.01020.0159−0.0530−0.0473
(0.0298)(0.0292)(0.0468)(0.0450)
p -value equality across incentives .0695.0963.0395.0278
Control for current loan qualityNoYesNoYes
SampleAllAllModeratesModerates
N7,6407,6402,6152,615
R20.02570.05360.02470.0544
Approve Loan Dummy
(1)(2)(3)(4)
Lag approve × flat incent −0.0814 ** −0.0712 ** −0.225 *** −0.228 ***
(0.0322)(0.0323)(0.0646)(0.0639)
Lag approve × stronger incent−0.00674−0.00215 −0.0525 ** −0.0484 **
(0.0134)(0.0134)(0.0215)(0.0214)
Lag approve × strongest incent0.01020.0159−0.0530−0.0473
(0.0298)(0.0292)(0.0468)(0.0450)
p -value equality across incentives .0695.0963.0395.0278
Control for current loan qualityNoYesNoYes
SampleAllAllModeratesModerates
N7,6407,6402,6152,615
R20.02570.05360.02470.0544

Notes. This table tests whether the decision to approve the current loan file is related to the decision to approve the previous loan file. Observations are at the loan officer × loan file level and exclude (as a dependent variable) the first loan file evaluated within each session. Columns (1) and (2) use the full sample while columns (3) and (4) restrict the sample to moderate loan officers (an observation is considered moderate if the loan officer’s average approval rate for loans, excluding the current session, is between 0.3 and 0.7 inclusive). Control variables include the loan officer’s mean approval rate within each incentive treatment (calculated excluding the current session), an indicator for whether the loan officer has ever approved all six loans in another session within the same incentive treatment, and an indicator for whether the current session is the only session attended by the loan officer within the incentive treatment (if so, the first two control variables cannot be calculated and are set to 0). Indicator variables for flat incent , strong incent , and strongest incent are also included. Standard errors are clustered by loan officer × incentive treatment. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table V

Loan Officers: Baseline Results

Approve Loan Dummy
(1)(2)(3)(4)
Lag approve × flat incent −0.0814 ** −0.0712 ** −0.225 *** −0.228 ***
(0.0322)(0.0323)(0.0646)(0.0639)
Lag approve × stronger incent−0.00674−0.00215 −0.0525 ** −0.0484 **
(0.0134)(0.0134)(0.0215)(0.0214)
Lag approve × strongest incent0.01020.0159−0.0530−0.0473
(0.0298)(0.0292)(0.0468)(0.0450)
p -value equality across incentives .0695.0963.0395.0278
Control for current loan qualityNoYesNoYes
SampleAllAllModeratesModerates
N7,6407,6402,6152,615
R20.02570.05360.02470.0544
Approve Loan Dummy
(1)(2)(3)(4)
Lag approve × flat incent −0.0814 ** −0.0712 ** −0.225 *** −0.228 ***
(0.0322)(0.0323)(0.0646)(0.0639)
Lag approve × stronger incent−0.00674−0.00215 −0.0525 ** −0.0484 **
(0.0134)(0.0134)(0.0215)(0.0214)
Lag approve × strongest incent0.01020.0159−0.0530−0.0473
(0.0298)(0.0292)(0.0468)(0.0450)
p -value equality across incentives .0695.0963.0395.0278
Control for current loan qualityNoYesNoYes
SampleAllAllModeratesModerates
N7,6407,6402,6152,615
R20.02570.05360.02470.0544

Notes. This table tests whether the decision to approve the current loan file is related to the decision to approve the previous loan file. Observations are at the loan officer × loan file level and exclude (as a dependent variable) the first loan file evaluated within each session. Columns (1) and (2) use the full sample while columns (3) and (4) restrict the sample to moderate loan officers (an observation is considered moderate if the loan officer’s average approval rate for loans, excluding the current session, is between 0.3 and 0.7 inclusive). Control variables include the loan officer’s mean approval rate within each incentive treatment (calculated excluding the current session), an indicator for whether the loan officer has ever approved all six loans in another session within the same incentive treatment, and an indicator for whether the current session is the only session attended by the loan officer within the incentive treatment (if so, the first two control variables cannot be calculated and are set to 0). Indicator variables for flat incent , strong incent , and strongest incent are also included. Standard errors are clustered by loan officer × incentive treatment. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

In columns (3) and (4) of Table V , we repeat the analysis for loan officers with moderate approval rates (estimated using approval rates in other sessions excluding the current session). In the loan officers experimental setting, a potential additional reason why the effect sizes are much larger in the moderate loan officers sample is that some loan officers may have decided to shirk in the experiment and approve almost all loans. Removing these loan officers from the sample leads to much larger effect sizes. Comparing the coefficient estimates with those in the same row in columns (1) and (2), we find that within each incentive treatment, moderate decision makers display much stronger negative autocorrelation in decisions. Under flat incentives, moderate decision makers are 23 percentage points less likely to approve the current loan if they approved the previous loan, implying that 9% of decisions are reversed. Even within the stronger and strongest incentive treatments, loan officers are 5 percentage points less likely to approve the current loan if they approved the previous loan. Overall, these tests suggest that loan officers, particularly moderate ones, exhibit significant negative autocorrelation in decisions, which can be mitigated through the use of strong pay for performance.

Tables A.VII–A.X in the Online Appendix report robustness tests for the loan officer sample. Table A.VII further tests whether the loan officers in the experiment are exerting effort in making accurate decisions and how that effort varies with incentives. We assess whether the loan approval decision is correlated with the ex ante quality of the loan file, as proxied by the fraction of other loan officers who approved the loan file, and the average quality score/rating given by other loan officers for the loan file. We further explore how the correlation between decisions and ex ante loan quality interact with the three incentive schemes. The results show that loan officers are more likely to approve loans that other loan officers approve or rate highly, and that the consensus in decision making increases with stronger incentives.

Table A.VIII shows that the results are similar when using other binary regression models, such as logit and probit. Table A.IX reports results from a specification that includes loan officer fixed effects. The coefficients are directionally similar but more negative than those in Table V . This is expected because the inclusion of fixed effects, particularly in short panels such as in the loan officers’ experimental setting, biases the coefficients downward. Table A.X reports results from a bivariate probit model that adjusts for the bias that, when using a binary dependent variable, the moderate subsample may mechanically exhibit more negative autocorrelation. The results using the bivariate probit model confirm that the negative autocorrelation in decisions is stronger for moderates even after adjusting for this potential bias, and the negative autocorrelation decreases with incentives.

In the remaining analysis, we pool the sample across all three incentive treatments unless otherwise noted. Table VI shows that loan officers with graduate school education and who spend more time reviewing the current loan file display significantly reduced negative autocorrelation in decisions. 13 Older and more experienced loan officers also display significantly reduced negative autocorrelation. These results are similar to our previous findings on asylum judges, and suggest that education, experience, and effort can reduce behavioral biases. Again, because we focus on decisions rather than predictions, our results do not necessarily imply that more educated, experienced, or conscientious loan officers suffer less from cognitive biases. These loan officers may still suffer equally from the gambler’s fallacy but draw, or believe they draw, more precise signals regarding current loan quality, leading them to rely less on their (misinformed and based-on-case-sequence) priors regarding loan quality.

Table VI

Loan Officers: Heterogeneity

Approve Loan Dummy
(1)(2)(3)(4)
Lag approve−0.0247* −0.127 *** −0.376 *** −0.0555 **
(0.0135)(0.0329)(0.136)(0.0250)
Grad school−0.0213
(0.0214)
Lag approve × grad school0.0448*
(0.0245)
Log(time viewed) −0.0968 ***
(0.0202)
Lag approve × log(time viewed) 0.0858 ***
(0.0230)
Log(age)−0.0603*
(0.0329)
Lag approve × log(age) 0.101 ***
(0.0375)
Log(experience)−0.0133
(0.00985)
Lag approve × log(experience)0.0226*
(0.0116)
SampleAllAllAllAll
N7,6407,6407,6407,640
R20.02560.02810.02600.0256
Approve Loan Dummy
(1)(2)(3)(4)
Lag approve−0.0247* −0.127 *** −0.376 *** −0.0555 **
(0.0135)(0.0329)(0.136)(0.0250)
Grad school−0.0213
(0.0214)
Lag approve × grad school0.0448*
(0.0245)
Log(time viewed) −0.0968 ***
(0.0202)
Lag approve × log(time viewed) 0.0858 ***
(0.0230)
Log(age)−0.0603*
(0.0329)
Lag approve × log(age) 0.101 ***
(0.0375)
Log(experience)−0.0133
(0.00985)
Lag approve × log(experience)0.0226*
(0.0116)
SampleAllAllAllAll
N7,6407,6407,6407,640
R20.02560.02810.02600.0256

Notes. This table explores heterogeneity in the correlation between current and lagged decisions. Grad school is an indicator for whether the loan officer has a graduate school education. Time viewed is the number of minutes spent reviewing the current loan file. Age is the age of the loan officer in years. Experience is the loan officer’s years of experience in the banking sector. All other variables are as described in Table V . Standard errors are clustered by loan officer × incentive treatment. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table VI

Loan Officers: Heterogeneity

Approve Loan Dummy
(1)(2)(3)(4)
Lag approve−0.0247* −0.127 *** −0.376 *** −0.0555 **
(0.0135)(0.0329)(0.136)(0.0250)
Grad school−0.0213
(0.0214)
Lag approve × grad school0.0448*
(0.0245)
Log(time viewed) −0.0968 ***
(0.0202)
Lag approve × log(time viewed) 0.0858 ***
(0.0230)
Log(age)−0.0603*
(0.0329)
Lag approve × log(age) 0.101 ***
(0.0375)
Log(experience)−0.0133
(0.00985)
Lag approve × log(experience)0.0226*
(0.0116)
SampleAllAllAllAll
N7,6407,6407,6407,640
R20.02560.02810.02600.0256
Approve Loan Dummy
(1)(2)(3)(4)
Lag approve−0.0247* −0.127 *** −0.376 *** −0.0555 **
(0.0135)(0.0329)(0.136)(0.0250)
Grad school−0.0213
(0.0214)
Lag approve × grad school0.0448*
(0.0245)
Log(time viewed) −0.0968 ***
(0.0202)
Lag approve × log(time viewed) 0.0858 ***
(0.0230)
Log(age)−0.0603*
(0.0329)
Lag approve × log(age) 0.101 ***
(0.0375)
Log(experience)−0.0133
(0.00985)
Lag approve × log(experience)0.0226*
(0.0116)
SampleAllAllAllAll
N7,6407,6407,6407,640
R20.02560.02810.02600.0256

Notes. This table explores heterogeneity in the correlation between current and lagged decisions. Grad school is an indicator for whether the loan officer has a graduate school education. Time viewed is the number of minutes spent reviewing the current loan file. Age is the age of the loan officer in years. Experience is the loan officer’s years of experience in the banking sector. All other variables are as described in Table V . Standard errors are clustered by loan officer × incentive treatment. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table VII examines decisions following streaks of decisions. We find that after approving two applications in a row, loan officers are 7.5 percentage points less likely to approve the next application, relative to when the loan officer denied two applications in a row. The effects are larger and more significant when restricted to moderate loan officers (column (2)). We easily reject that β1 = β2 = β3 and β1 = β3 for the past sequence of decisions. However, “Lag reject-approve” has a less negative coefficient than “Lag approve-reject,” even though a gambler’s fallacy model where recency matters would predict the opposite. The sample size is small, however, and the difference between these two coefficients is insignificant and small in the sample of moderates.

Table VII

Loan Officers: Reactions to Streaks

Approve Loan Dummy
(1)(2)
β1 : Lag approve-approve −0.0751 *** −0.165 ***
(0.0216)(0.0329)
β2 : Lag approve-reject −0.0691 *** −0.0955 ***
(0.0236)(0.0347)
β3 : Lag reject-approve −0.0322 −0.0832 **
(0.0225)(0.0332)
p -value: β1 = β2 = β3.0178.00448
p -value: β1=β2.703.0134
p -value: β1=β3.00493.00300
p -value: β2=β3.0483.688
SampleAllModerates
N6,1122,092
R20.02900.0322
Approve Loan Dummy
(1)(2)
β1 : Lag approve-approve −0.0751 *** −0.165 ***
(0.0216)(0.0329)
β2 : Lag approve-reject −0.0691 *** −0.0955 ***
(0.0236)(0.0347)
β3 : Lag reject-approve −0.0322 −0.0832 **
(0.0225)(0.0332)
p -value: β1 = β2 = β3.0178.00448
p -value: β1=β2.703.0134
p -value: β1=β3.00493.00300
p -value: β2=β3.0483.688
SampleAllModerates
N6,1122,092
R20.02900.0322

Notes. This table tests how loan officers react to streaks in past decisions. Lag approve-approve is an indicator for whether the loan officer approved the two most recent previous loans. Lag approve-reject is an indicator for whether the loan officer rejected the most recent previous loan and approved the loan before that. Lag reject-approve is an indicator for whether the loan officer approved the most recent previous loan and rejected the loan before that. The omitted category is Lag reject-reject , which is an indicator for whether the loan officer rejected the two most recent previous loans. The sample excludes observations corresponding to the first two loans reviewed within each session. All other variables are as described in Table V . Standard errors are clustered by loan officer × incentive treatment. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table VII

Loan Officers: Reactions to Streaks

Approve Loan Dummy
(1)(2)
β1 : Lag approve-approve −0.0751 *** −0.165 ***
(0.0216)(0.0329)
β2 : Lag approve-reject −0.0691 *** −0.0955 ***
(0.0236)(0.0347)
β3 : Lag reject-approve −0.0322 −0.0832 **
(0.0225)(0.0332)
p -value: β1 = β2 = β3.0178.00448
p -value: β1=β2.703.0134
p -value: β1=β3.00493.00300
p -value: β2=β3.0483.688
SampleAllModerates
N6,1122,092
R20.02900.0322
Approve Loan Dummy
(1)(2)
β1 : Lag approve-approve −0.0751 *** −0.165 ***
(0.0216)(0.0329)
β2 : Lag approve-reject −0.0691 *** −0.0955 ***
(0.0236)(0.0347)
β3 : Lag reject-approve −0.0322 −0.0832 **
(0.0225)(0.0332)
p -value: β1 = β2 = β3.0178.00448
p -value: β1=β2.703.0134
p -value: β1=β3.00493.00300
p -value: β2=β3.0483.688
SampleAllModerates
N6,1122,092
R20.02900.0322

Notes. This table tests how loan officers react to streaks in past decisions. Lag approve-approve is an indicator for whether the loan officer approved the two most recent previous loans. Lag approve-reject is an indicator for whether the loan officer rejected the most recent previous loan and approved the loan before that. Lag reject-approve is an indicator for whether the loan officer approved the most recent previous loan and rejected the loan before that. The omitted category is Lag reject-reject , which is an indicator for whether the loan officer rejected the two most recent previous loans. The sample excludes observations corresponding to the first two loans reviewed within each session. All other variables are as described in Table V . Standard errors are clustered by loan officer × incentive treatment. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Last, we discuss why our results are robust to a unique feature of the design of the original field experiment. Within each session, the order of the loans viewed by the loan officers on the computer screen was randomized. However, the original experimenters implemented a balanced session design. Each session consisted of four performing loans and two nonperforming loans. 14 If the loan officers had realized that sessions were balanced, a rational response would have been to reject loans with a greater probability after approving loans within the same session (and vice versa). We believe there are several reasons it is unlikely that loan officers would react to the balanced session design.

First, loan officers were never informed that sessions were balanced. Instead, they were told that the six loans within each session were randomly selected from a large population of loans. Second, if loan officers had “figured out” that sessions were balanced, we would expect that loan officers would be more likely to use this information when subject to stronger pay for performance. In other words, there should have been greater negative autocorrelation within the incentive treatments with stronger pay-for-performance, but this is the opposite of what we find. Also, the better educated may be more likely to deduce that sessions are balanced, so they should display stronger negative autocorrelation, which is again the opposite of what we find.

In columns (1) and (2) of Table A.XI in the Online Appendix , we reproduce the baseline results showing that the negative autocorrelation in decisions is strongest in the flat incentive scheme treatment. In columns (3)–(6), we show that the true performance of the current loan is negatively related to both the lagged decision and the true quality of the lagged loan file, and the negative autocorrelation in true loan quality is approximately similar in magnitude across all three incentive treatments. The results in columns (1) and (2) are inconsistent with loan officers realizing that sessions were balanced. If loan officers had realized that sessions were balanced, we would have expected the opposite result, that is, that the negative autocorrelation in decisions would be equally or more strong under the stronger incentive schemes.

V. B aseball U mpires

Our final empirical setting uses data on called pitches by the home plate umpire in Major League Baseball.

V.A. Baseball Umpires: Data Description and Institutional Context

In MLB, one important job of the home plate umpire is to call a pitch as either a strike or ball, if a batter does not swing. The umpire has to determine if the location of the ball as it passes over home plate is within the strike zone as described and shown in Figure I . If the umpire decides the pitch is within the strike zone, he calls it a strike and otherwise calls it a ball. The boundaries of the strike zone are officially defined as in the caption for Figure I , and are not subject to individual umpire interpretation. However, each umpire is expected to use his “best judgment” when determining the location of the ball relative to the strike zone boundaries. Hence, umpire judgment matters.

Figure I

Baseball Umpires: The Strike Zone

According to Major League Baseball’s “Official Baseball Rules” 2014 Edition, Rule 2.00, “The STRIKE ZONE is that area over home plate the upper limit of which is a horizontal line at the midpoint between the top of the shoulders and the top of the uniform pants, and the lower level is a line at the hollow beneath the kneecap. The Strike Zone shall be determined from the batter’s stance as the batter is prepared to swing at a pitched ball.”

We test whether baseball umpires are more likely to call the current pitch a ball after calling the previous pitch a strike. Of course, pitch quality (e.g., location) is not randomly ordered. For example, a pitcher will adjust his strategy depending on game conditions. An advantage of the baseball umpire data is that it includes precise measures of the trajectory and location of each pitch. Thus, while pitch quality may not be randomly ordered over time, we can control for each pitch’s true location and measure whether mistakes in calls, conditional on a pitch’s true location, are negatively predicted by the previous call.

We use data on umpire calls of pitches from PITCHf/x, a system that tracks the trajectory and location of each pitch with respect to each batter’s strike zone as the pitch crosses in front of home plate. The location measures are accurate to within a square centimeter. The PITCHf/x system was installed in 2006 in every MLB stadium and implemented for part of the 2007 season. Our data cover approximately 3.5 million pitches over the 2008 to 2012 MLB seasons, when the system produced an entire season of pitch data. We restrict our analysis to called pitches, that is, pitches in which the batter does not swing (so the umpire must make a call), excluding the first called pitch in each inning. This sample restriction leaves us with approximately 1.5 million called pitches over 12,564 games by 127 different umpires. In some tests, we further restrict our sample to consecutive called pitches, where the current called pitch and the previous called pitch were not interrupted by another pitch in which the umpire did not make a call (e.g., because the batter took a swing). Consecutive called pitches account for just under 900,000 observations.

Baseball umpires in our sample do not receive immediate feedback regarding whether each call was correct (data on whether each pitch was within the strike zone according to the PITCHf/x system is available after the game). Nevertheless, the umpire likely receives some cues. At the very least, umpires can observe the extent to which others disagreed with the call. First, the umpire knows roughly where the pitch landed and how “close” the call was. A call made on a pitch near the edge of the strike zone is more ambiguous, for instance. More to the point, the umpire also receives cues from the batter’s reaction, the pitcher’s reaction, and the crowd’s reaction to the call. These parties voice their disagreement if they believe the umpire made a mistake. Making an unambiguous erroneous call will likely draw a stronger reaction from at least one of these parties.

Table VIII summarizes our data sample. Approximately 30% of all called pitches are called as strikes (rather than balls). Umpires make the correct call 86.6% of the time. We also categorize pitches by whether they were ambiguous (difficult to call) or obvious (easy to call). Ambiguous pitches fall within ±1.5 inches of the edge of the strike zone; 60% of ambiguous pitches are called correctly. Obvious pitches fall within three inches around the center of the strike zone, or six inches or more outside the edge of the strike zone; 99% of obvious pitches are called correctly.

Table VIII

Baseball Umpires: Summary Statistics

Number of called pitches following a previous called pitch1,536,807
Number of called pitches following a consecutive previous called pitch898,741
Number of games12,564
Number of umpires127
Fraction of pitches called as strike0.3079
Fraction of pitches called correctly0.8664
Fraction of pitches categorized as ambiguous0.1686
Fraction of pitches categorized as obvious0.3731
Fraction of ambiguous pitches called correctly0.6006
Fraction of obvious pitches called correctly0.9924
Number of called pitches following a previous called pitch1,536,807
Number of called pitches following a consecutive previous called pitch898,741
Number of games12,564
Number of umpires127
Fraction of pitches called as strike0.3079
Fraction of pitches called correctly0.8664
Fraction of pitches categorized as ambiguous0.1686
Fraction of pitches categorized as obvious0.3731
Fraction of ambiguous pitches called correctly0.6006
Fraction of obvious pitches called correctly0.9924

Notes. This table presents summary statistics for the sample of MLB umpire calls that we use in our decision making analysis. The sample represents all called pitches by MLB umpires from all games during the 2008 to 2012 seasons, covering 3.5 million pitches in 12,564 games, from 127 different home plate umpires. We restrict the sample to called pitches following a previously called pitch in the same inning. We classify a pitch as ambiguous if the location of the pitch is within 1.5 inches of the boundary of the strike zone. We classify a pitch as obvious if the location of the pitch is within 3 inches of the center of the strike zone, or 6 inches or more outside of the edge of the strike zone.

Table VIII

Baseball Umpires: Summary Statistics

Number of called pitches following a previous called pitch1,536,807
Number of called pitches following a consecutive previous called pitch898,741
Number of games12,564
Number of umpires127
Fraction of pitches called as strike0.3079
Fraction of pitches called correctly0.8664
Fraction of pitches categorized as ambiguous0.1686
Fraction of pitches categorized as obvious0.3731
Fraction of ambiguous pitches called correctly0.6006
Fraction of obvious pitches called correctly0.9924
Number of called pitches following a previous called pitch1,536,807
Number of called pitches following a consecutive previous called pitch898,741
Number of games12,564
Number of umpires127
Fraction of pitches called as strike0.3079
Fraction of pitches called correctly0.8664
Fraction of pitches categorized as ambiguous0.1686
Fraction of pitches categorized as obvious0.3731
Fraction of ambiguous pitches called correctly0.6006
Fraction of obvious pitches called correctly0.9924

Notes. This table presents summary statistics for the sample of MLB umpire calls that we use in our decision making analysis. The sample represents all called pitches by MLB umpires from all games during the 2008 to 2012 seasons, covering 3.5 million pitches in 12,564 games, from 127 different home plate umpires. We restrict the sample to called pitches following a previously called pitch in the same inning. We classify a pitch as ambiguous if the location of the pitch is within 1.5 inches of the boundary of the strike zone. We classify a pitch as obvious if the location of the pitch is within 3 inches of the center of the strike zone, or 6 inches or more outside of the edge of the strike zone.

V.B. Baseball Umpires: Empirical Specification

Our baseline tests explore whether umpires are less likely to call the current pitch a strike after calling the previous pitch a strike, controlling for pitch location, which should be the sole determinant of the call. The sample includes all called pitches except for the first in each game or inning. We define Y it as an indicator for whether the current pitch is called a strike, while Y i,t−1 is an indicator for whether the previous pitch was called a strike.

To attribute negative autocorrelation in decisions to cognitive biases, we assume that the underlying quality of the pitches (e.g., the location of the pitch relative to the strike zone), after conditioning on a set of controls, is not itself negatively autocorrelated. To address this potential concern, we include detailed controls for the characteristics of the current pitch. First we control for the pitch location relative to an absolute point on home plate using indicators for each 3 × 3 inch square. We also control explicitly for whether the current pitch was within the strike zone based on its location, which should be the only characteristic that matters for the call according to MLB rules. Finally, we control for the speed, acceleration, curvature, and spin in the x , y , and z directions of the pitch, which may affect an umpire’s perception. For a complete detailed list of all control variables, please see  Appendix D . Our control variables address the concern that pitch characteristics are not randomly ordered. In addition, the fact that we control for whether the current pitch is actually within the true strike zone for each batter implies that any nonzero coefficients on other variables represent mistakes on the part of the umpire. Nothing else, according to the rules, should matter for the call except the location of the pitch relative to the strike zone. Specifically, any significant coefficient on the lagged umpire decision is evidence of mistakes.

Of course, umpires may be biased in other ways. For example, Parsons et al. (2011) show evidence of discrimination in calls: umpires are less likely to call strikes if the umpire and pitcher differ in race and ethnicity. However, while biases against teams or specific types of players could affect the base rate of called pitches within innings or against certain pitchers, they should not generate high-frequency negative autocorrelation in calls, which is the bias we focus on in this article. 15 In addition, in the Online Appendix , we include umpire, batter, and pitcher fixed effects, which should account for these sorts of biases, and find similar effects. More relevant for our tests, Moskowitz and Wertheim (2014) show that umpires prefer to avoid making calls that result in terminal outcomes or that may determine game outcomes. To differentiate our finding from these other types of biases that may affect the probability of the umpire calling strike versus ball at different points in the game, we control for indicator variables for every possible count combination (number of balls and strikes called so far on the batter), 16 the leverage index (a measure developed by Tom Tango of how important a particular situation is in a baseball game depending on the inning, score, outs, and number of players on base), indicators for the score of the team at bat, indicators for the score of the team in the field, and an indicator for whether the batter belongs to the home team.

In our previous analysis of asylum judges and loan officers, we controlled for heterogeneity in each decision maker’s approval rate using the decision maker’s leave-out-mean approval rate, moving average of the past five decisions, and/or decision maker fixed effects. We also conducted subsample analysis limited to moderate decision makers. These control variables for decision maker heterogeneity are less relevant in the setting of baseball umpires because professional umpires tend to have very homogeneous mean rates of strike calls. 17 Therefore, we present our baseline regression results without including controls for individual heterogeneity (the lack of controls should be a bias against findings of negative autocorrelation), and show in the Online Appendix that the results are very similar if we control for umpire fixed effects or a moving average of the past five decisions.

Finally, we use the sample of called pitches, that is, pitches in which the batter chose not to swing. Whether the batter chooses to swing is unlikely to be random and may depend on various game conditions, which is partly why we add all of the controls above. However, endogenous sample selection of this form should also not bias our results toward finding spurious negative autocorrelation in umpire calls. We test, within the sample of called pitches, whether umpires tend to make mistakes in the opposite direction of the previous decision, after controlling for the true quality (location) of the current pitch. We also show that, insofar as pitch quality is not randomly ordered, it tends to be slightly positively autocorrelated within this sample, which is a bias against our findings of negative autocorrelation.

V.C. Baseball Umpires: Results

Table IX , column (1) shows that umpires are 0.9 percentage points less likely to call a pitch a strike if the most recent previously called pitch was called a strike. Column (2) shows that the negative autocorrelation is stronger following streaks. Umpires are 1.3 percentage points less likely to call a pitch a strike if the two most recently called pitches were also called strikes. Further, umpires are less likely to call the current pitch a strike if the most recent pitch was called a strike and the pitch before that was called a ball than if the ordering of the last two calls were reversed. In other words, extreme recency matters. We easily reject that β1 = β2 = β3 in favor of β1 < β2 < β3 < 0. These findings are consistent with the gambler’s fallacy and less consistent with a quotas explanation (in addition, umpires do not face explicit quotas). The results are also less consistent with a learning model about where to set a quality cutoff bar because there is an objective quality bar (the official strike zone) that, according to the rules, should not move depending on the quality of the previous pitch.

Table IX

Baseball Umpires: Baseline Results

Strike Full Sample
Consecutive Pitches
(1)(2)(3)(4)
Lag strike −0.00924 *** −0.0146 ***
(0.000591)(0.000972)
β1 : Lag strike-strike −0.0133 *** −0.0208 ***
(0.00104)(0.00269)
β2 : Lag ball-strike −0.0100 *** −0.0188 ***
(0.000718)(0.00157)
β3 : Lag strike-ball −0.00276 *** −0.00673 ***
(0.000646)(0.00155)
p -value: β1 = β2 = β31.49e-315.17e-22
p -value: β1=β2.000423.414
p -value: β1=β34.71e-253.07e-08
p -value: β2=β33.79e-241.62e-21
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N1,536,8071,331,399898,741428,005
R20.6690.6680.6650.669
Strike Full Sample
Consecutive Pitches
(1)(2)(3)(4)
Lag strike −0.00924 *** −0.0146 ***
(0.000591)(0.000972)
β1 : Lag strike-strike −0.0133 *** −0.0208 ***
(0.00104)(0.00269)
β2 : Lag ball-strike −0.0100 *** −0.0188 ***
(0.000718)(0.00157)
β3 : Lag strike-ball −0.00276 *** −0.00673 ***
(0.000646)(0.00155)
p -value: β1 = β2 = β31.49e-315.17e-22
p -value: β1=β2.000423.414
p -value: β1=β34.71e-253.07e-08
p -value: β2=β33.79e-241.62e-21
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N1,536,8071,331,399898,741428,005
R20.6690.6680.6650.669

Notes. This table tests whether the decision to call the current pitch a strike is related to the decision to call the previous pitch(es) a strike. Observations are at the umpire × pitch level and exclude (as a dependent variable) the first pitch within each game. Columns (1) and (2) use the sample of all called pitches while columns (3) and (4) restrict the sample to consecutive called pitches that are not interrupted by a pitch in which the umpire did not make a call (e.g., because the batter swung at the ball). Note that the sample size falls further in column (4) because we require that the current pitch, previous pitch, and previous pitch before those are all consecutive. Control variables include the pitch location (indicators for each 3×3 inch square), an indicator for whether the current pitch was within the strike zone, the speed, acceleration, and spin in the x , y , and z directions of the pitch, break angle characteristics, indicators for every possible count combination ( # of balls and strikes called so far for the batter), the leverage index, indictors for the score of the team at bat and indicators for the score of the team in the field, and an indicator for whether the batter belongs to the home team. For a complete detailed list of control variables, please see  Appendix D . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table IX

Baseball Umpires: Baseline Results

Strike Full Sample
Consecutive Pitches
(1)(2)(3)(4)
Lag strike −0.00924 *** −0.0146 ***
(0.000591)(0.000972)
β1 : Lag strike-strike −0.0133 *** −0.0208 ***
(0.00104)(0.00269)
β2 : Lag ball-strike −0.0100 *** −0.0188 ***
(0.000718)(0.00157)
β3 : Lag strike-ball −0.00276 *** −0.00673 ***
(0.000646)(0.00155)
p -value: β1 = β2 = β31.49e-315.17e-22
p -value: β1=β2.000423.414
p -value: β1=β34.71e-253.07e-08
p -value: β2=β33.79e-241.62e-21
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N1,536,8071,331,399898,741428,005
R20.6690.6680.6650.669
Strike Full Sample
Consecutive Pitches
(1)(2)(3)(4)
Lag strike −0.00924 *** −0.0146 ***
(0.000591)(0.000972)
β1 : Lag strike-strike −0.0133 *** −0.0208 ***
(0.00104)(0.00269)
β2 : Lag ball-strike −0.0100 *** −0.0188 ***
(0.000718)(0.00157)
β3 : Lag strike-ball −0.00276 *** −0.00673 ***
(0.000646)(0.00155)
p -value: β1 = β2 = β31.49e-315.17e-22
p -value: β1=β2.000423.414
p -value: β1=β34.71e-253.07e-08
p -value: β2=β33.79e-241.62e-21
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N1,536,8071,331,399898,741428,005
R20.6690.6680.6650.669

Notes. This table tests whether the decision to call the current pitch a strike is related to the decision to call the previous pitch(es) a strike. Observations are at the umpire × pitch level and exclude (as a dependent variable) the first pitch within each game. Columns (1) and (2) use the sample of all called pitches while columns (3) and (4) restrict the sample to consecutive called pitches that are not interrupted by a pitch in which the umpire did not make a call (e.g., because the batter swung at the ball). Note that the sample size falls further in column (4) because we require that the current pitch, previous pitch, and previous pitch before those are all consecutive. Control variables include the pitch location (indicators for each 3×3 inch square), an indicator for whether the current pitch was within the strike zone, the speed, acceleration, and spin in the x , y , and z directions of the pitch, break angle characteristics, indicators for every possible count combination ( # of balls and strikes called so far for the batter), the leverage index, indictors for the score of the team at bat and indicators for the score of the team in the field, and an indicator for whether the batter belongs to the home team. For a complete detailed list of control variables, please see  Appendix D . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

All analysis in this and subsequent tables includes detailed controls for the actual location, speed, and curvature of the pitch. In addition, because we control for an indicator for whether the current pitch actually fell within the strike zone, all reported non-zero coefficients reflect mistakes on the part of the umpires (if the umpire always made the correct call, all coefficients other than the coefficient on the indicator for whether the pitch fell within the strike zone should equal zero).

In columns (3) and (4) of Table IX , we repeat the analysis but restrict the sample to pitches that were called consecutively (so both the current and most recent pitch received umpire calls of strike or ball) without any interruption. In the consecutive sample, the umpire’s recent previous calls may be more salient because they are not separated by uncalled pitches. We find that the magnitude of the negative autocorrelation increases substantially in this sample. Umpires are 2.1 percentage points less likely to call the current pitch a strike if the previous two pitches were called strikes. This represents a 6.8% decline relative to the base rate of strike calls. We test whether the differences in magnitudes between the full sample and the consecutive called pitches sample are significant and find that they are, with p -values below .001. In all subsequent analysis, unless otherwise noted, we restrict the sample to consecutive called pitches.

Tables A.XII–A.XVI in the Online Appendix report robustness tests of these results. Table shows similar results with batter, pitcher, and umpire fixed effects. Table A.XII shows similar results with batter, pitcher, and umpire fixed effects. Table A.XIII reports similar results using the moving average of the umpire's past five calls as a control variable. Table A.XIV shows similar effects and economic magnitudes using logit and probit models, and Table A.XV shows similar results using a bivariate probit model. Table A.XVI shows similar results if we exclude control variables for the count (number of balls and strikes called so far on the batter).

Since in this setting we are particularly concerned that the “quality,” that is, the location, of the pitch will also react to the umpire’s previous call, we control for each pitch’s true location (plus the other controls described in  Appendix D ) and measure whether mistakes in calls conditional on a pitch’s true location are negatively predicted by the previous call. If our location and other controls are mismeasured or inadequate, however, then autocorrelation in the quality of pitches could still be an issue. To assess how concerning this issue might be, we also reestimate the regression by replacing the dependent variable of whether a pitch is called a strike with an indicator for whether the pitch is actually a true strike. We also estimate a version of the analysis where the dependent variable is replaced with the distance of the pitch from the center of the strike zone. We then test whether these proxies for the true location of the pitch depend on whether the lagged pitch was called a strike. In other words, how does the actual quality of the pitch respond to the previous call?

Table A.XVII in the Online Appendix shows that the negative autocorrelation in umpire calls is unlikely to be caused by changes in the actual location of the pitch. We continue to restrict the sample to consecutive called pitches and repeat the analysis using the current pitch’s true location as our dependent variable (to identify the effect of previous calls on the location of the current pitch, we exclude location controls). Columns (1) and (2), which use an indicator for whether the current pitch was within the strike zone as the dependent variable, show that pitchers are more likely to throw another strike after the previous pitch was called a strike, resulting in positive, rather than negative, coefficients on the previous call. Hence, autocorrelation in the quality of pitches biases us against our finding of negatively autocorrelated decision making. In columns (3) and (4), we use the distance of the pitch in inches from the center of the strike zone as the dependent variable. If pitchers are more likely to throw true balls (more distant from the center of the strike zone) after the previous pitch was called a strike, we should find significant positive coefficients on lagged strike calls; again, we find the opposite. In other words, endogenous changes in pitch location as a response to previous calls should lead to positive rather than negative autocorrelation in umpire calls because the quality of pitches is slightly positively autocorrelated. Finally, in columns (5) and (6), we include the same set of detailed pitch location controls (dummies for each 3 × 3 inch square) as in our baseline specifications, and find that all coefficients on lagged calls become small and insignificant, suggesting that our controls effectively remove any autocorrelation in the quality of pitches and account for the pitcher’s endogenous responses to previous calls.

Table X shows that the negative autocorrelation in decisions is reduced when umpires receive more informative signals about the quality of the current pitch. Columns (1) and (2) restrict the analysis to observations in which the current pitch is ambiguous—pitches located close to the boundary of the strike zone, where it is difficult to make a correct strike or ball call. Columns (3) and (4) restrict the analysis to observations in which the current pitch is likely to be obvious—pitches located close to the center of the strike zone (“obvious” strikes) or far from the edge of the strike zone (“obvious” balls). We find that the magnitude of negative autocorrelation coefficients are 10 to 15 times larger when the current pitch is ambiguous relative to when the current pitch is obvious. We can confidently reject equality of the estimates for ambiguous and obvious pitches in columns (1) and (3) with p -values well below .001. This is consistent with the gambler’s fallacy model that the decision maker’s prior beliefs about case quality will have less impact on the decision when the signal about current case quality is more informative.

Table X

Baseball Umpires: Ambiguous versus Obvious Calls

Strike Current Pitch Ambiguous
Current Pitch Obvious
(1)(2)(3)(4)
Lag strike −0.0347 *** −0.00226 ***
(0.00378)(0.000415)
β1 : Lag strike-strike −0.0479 *** −0.00515 ***
(0.0113)(0.00101)
β2 : Lag ball-strike −0.0324 *** −0.00442 ***
(0.00566)(0.000773)
β3 : Lag strike-ball −0.000838 −0.00283 ***
(0.00563)(0.000841)
p -value: β1 = β2 = β31.74e-11.00573
p -value: β1 = β2.148.395
p -value: β1 = β3.0000205.0104
p -value: β2 = β35.02e-11.00507
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N151,50173,820335,318153,996
R20.3170.3160.8910.896
Strike Current Pitch Ambiguous
Current Pitch Obvious
(1)(2)(3)(4)
Lag strike −0.0347 *** −0.00226 ***
(0.00378)(0.000415)
β1 : Lag strike-strike −0.0479 *** −0.00515 ***
(0.0113)(0.00101)
β2 : Lag ball-strike −0.0324 *** −0.00442 ***
(0.00566)(0.000773)
β3 : Lag strike-ball −0.000838 −0.00283 ***
(0.00563)(0.000841)
p -value: β1 = β2 = β31.74e-11.00573
p -value: β1 = β2.148.395
p -value: β1 = β3.0000205.0104
p -value: β2 = β35.02e-11.00507
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N151,50173,820335,318153,996
R20.3170.3160.8910.896

Notes. This table tests how our results differ depending on whether the current pitch is ambiguous or obvious. The sample is restricted to consecutive called pitches. Columns (1) and (2) restrict the sample to observations in which the current pitch is ambiguous (the location of the pitch is within 1.5 inches of the boundary of the strike zone). Columns (3) and (4) restrict the sample to observations in which the current pitch is obvious (the location of the pitch is within 3 inches of the center of the strike zone or 6 inches or more outside of the edge of the strike zone). All control variables are as described in Table IX . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table X

Baseball Umpires: Ambiguous versus Obvious Calls

Strike Current Pitch Ambiguous
Current Pitch Obvious
(1)(2)(3)(4)
Lag strike −0.0347 *** −0.00226 ***
(0.00378)(0.000415)
β1 : Lag strike-strike −0.0479 *** −0.00515 ***
(0.0113)(0.00101)
β2 : Lag ball-strike −0.0324 *** −0.00442 ***
(0.00566)(0.000773)
β3 : Lag strike-ball −0.000838 −0.00283 ***
(0.00563)(0.000841)
p -value: β1 = β2 = β31.74e-11.00573
p -value: β1 = β2.148.395
p -value: β1 = β3.0000205.0104
p -value: β2 = β35.02e-11.00507
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N151,50173,820335,318153,996
R20.3170.3160.8910.896
Strike Current Pitch Ambiguous
Current Pitch Obvious
(1)(2)(3)(4)
Lag strike −0.0347 *** −0.00226 ***
(0.00378)(0.000415)
β1 : Lag strike-strike −0.0479 *** −0.00515 ***
(0.0113)(0.00101)
β2 : Lag ball-strike −0.0324 *** −0.00442 ***
(0.00566)(0.000773)
β3 : Lag strike-ball −0.000838 −0.00283 ***
(0.00563)(0.000841)
p -value: β1 = β2 = β31.74e-11.00573
p -value: β1 = β2.148.395
p -value: β1 = β3.0000205.0104
p -value: β2 = β35.02e-11.00507
Pitch locationYesYesYesYes
Pitch trajectoryYesYesYesYes
Game conditionsYesYesYesYes
N151,50173,820335,318153,996
R20.3170.3160.8910.896

Notes. This table tests how our results differ depending on whether the current pitch is ambiguous or obvious. The sample is restricted to consecutive called pitches. Columns (1) and (2) restrict the sample to observations in which the current pitch is ambiguous (the location of the pitch is within 1.5 inches of the boundary of the strike zone). Columns (3) and (4) restrict the sample to observations in which the current pitch is obvious (the location of the pitch is within 3 inches of the center of the strike zone or 6 inches or more outside of the edge of the strike zone). All control variables are as described in Table IX . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

It is also important to note that the stronger negative autocorrelation for ambiguous pitches is not merely a consequence of these pitches being more difficult to call. We expect umpire accuracy to decline for these pitches, but an unbiased umpire should not be more likely to make mistakes in the opposite direction from the previous call. That is, overall accuracy may be lower but there is no expectation that calls should alternate and be negatively autocorrelated.

In Table XI , we explore heterogeneity with respect to game conditions and umpire characteristics. Column (1) shows that an increase in leverage (the importance of a particular game situation for determining the game outcome) leads to significantly stronger negative autocorrelation in decisions. However, the magnitude of the effect is small: a one standard deviation increase in game leverage leads to less than a 10% increase in the extent of negative autocorrelation. Column (2) shows that umpires who are more accurate (calculated as the fraction of pitches correctly called by the umpire in other games excluding the current game) are also less susceptible to negatively autocorrelated decision making. A one standard deviation increase in umpire accuracy reduces negative autocorrelation by 25%. Finally, column (3) tests whether the magnitude of the negative autocorrelation varies by game attendance. We divide game attendance into quintiles and compare the highest and lowest quintiles to the middle three quintiles (which represent the omitted category). We do not find any significant differences in behavior by game attendance except in the highest quintile, where the negative autocorrelation increases by 18%. However, this difference in behavior is only marginally significant. The marginally stronger negative autocorrelation effects for high leverage situations and high attendance games may be consistent with umpires worrying about appearing biased in more heavily scrutinized environments, where fans, analysts, and the media may suffer from the gambler’s fallacy.

Table XI

Baseball Umpires: Heterogeneity

(1)(2)(3)
Lag strike −0.0146 *** −0.0146 *** −0.0143 ***
(0.000972)(0.000972)(0.00108)
Leverage0.000330
(0.000390)
Lag strike × leverage −0.00140 **
(0.000625)
Umpire accuracy −0.00406 ***
(0.000451)
Lag strike × umpire accuracy 0.00353 ***
(0.000621)
High attendance 0.00441 ***
(0.00115)
Low attendance −0.00330 ***
(0.00117)
Lag strike × high attendance−0.00270*
(0.00157)
Lag strike × low attendance0.00123
(0.00164)
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741898,154894,779
R20.6650.6650.665
(1)(2)(3)
Lag strike −0.0146 *** −0.0146 *** −0.0143 ***
(0.000972)(0.000972)(0.00108)
Leverage0.000330
(0.000390)
Lag strike × leverage −0.00140 **
(0.000625)
Umpire accuracy −0.00406 ***
(0.000451)
Lag strike × umpire accuracy 0.00353 ***
(0.000621)
High attendance 0.00441 ***
(0.00115)
Low attendance −0.00330 ***
(0.00117)
Lag strike × high attendance−0.00270*
(0.00157)
Lag strike × low attendance0.00123
(0.00164)
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741898,154894,779
R20.6650.6650.665

Notes. This table tests how our results differ depending on game conditions or umpire characteristics. The sample is restricted to consecutive called pitches. Leverage and umpire accuracy are represented as z -scores. Leverage is a measure developed by Tom Tango of how important a particular situation is in a baseball game depending on the inning, score, outs, and number of players on base. Umpire accuracy is the fraction of pitches correctly called by the umpire, calculated excluding observations corresponding to the current game. High and low attendance are indicator variables for whether game attendance is in the highest and lowest quintiles of attendance, respectively (the omitted category consists of the middle three quintiles). All control variables are as described in Table IX . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table XI

Baseball Umpires: Heterogeneity

(1)(2)(3)
Lag strike −0.0146 *** −0.0146 *** −0.0143 ***
(0.000972)(0.000972)(0.00108)
Leverage0.000330
(0.000390)
Lag strike × leverage −0.00140 **
(0.000625)
Umpire accuracy −0.00406 ***
(0.000451)
Lag strike × umpire accuracy 0.00353 ***
(0.000621)
High attendance 0.00441 ***
(0.00115)
Low attendance −0.00330 ***
(0.00117)
Lag strike × high attendance−0.00270*
(0.00157)
Lag strike × low attendance0.00123
(0.00164)
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741898,154894,779
R20.6650.6650.665
(1)(2)(3)
Lag strike −0.0146 *** −0.0146 *** −0.0143 ***
(0.000972)(0.000972)(0.00108)
Leverage0.000330
(0.000390)
Lag strike × leverage −0.00140 **
(0.000625)
Umpire accuracy −0.00406 ***
(0.000451)
Lag strike × umpire accuracy 0.00353 ***
(0.000621)
High attendance 0.00441 ***
(0.00115)
Low attendance −0.00330 ***
(0.00117)
Lag strike × high attendance−0.00270*
(0.00157)
Lag strike × low attendance0.00123
(0.00164)
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741898,154894,779
R20.6650.6650.665

Notes. This table tests how our results differ depending on game conditions or umpire characteristics. The sample is restricted to consecutive called pitches. Leverage and umpire accuracy are represented as z -scores. Leverage is a measure developed by Tom Tango of how important a particular situation is in a baseball game depending on the inning, score, outs, and number of players on base. Umpire accuracy is the fraction of pitches correctly called by the umpire, calculated excluding observations corresponding to the current game. High and low attendance are indicator variables for whether game attendance is in the highest and lowest quintiles of attendance, respectively (the omitted category consists of the middle three quintiles). All control variables are as described in Table IX . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

VI. A ddressing A lternative E xplanations

Across all three of our settings, we find consistent evidence of negatively autocorrelated decision making. We believe our results are best explained by decision makers suffering from the gambler’s fallacy. Other explanations for negatively autocorrelated decisions such as quotas, learning, or preferences to treat all parties fairly, are less consistent with the evidence, though we cannot completely rule out sequential contrast effects as an alternative explanation.

VI.A. Sequential Contrast Effects

Perhaps the most difficult alternative story to distinguish—and one that we will not be able to fully reject empirically—is sequential contrast effects. Under SCE, negative autocorrelation in decisions can arise if agents view the current case in contrast to the preceding case. SCE imply that lagged case quality affects the perception of the quality of the current case. Under Rabin’s (2002) original model, where agents react to past binary outcomes, we could, in principle, distinguish between agents responding to past decisions (the gambler’s fallacy) versus lagged quality (SCE), where the former is a binary outcome and the latter is continuous. Specifically, we could estimate
Yit=β0+β1Yi,t1+β2Qualityi,t1+Controls+ϵit.
If SCE drives our findings, then we expect to find that β2 < 0, holding constant the previous discrete decision Y i,t– 1 . The idea is, holding constant the previous discrete decision, SCE predicts that decision makers should be more likely to reject the current case if the previous case was of high quality, as measured continuously using Quality i,t–1 . However, in a more general model of the gambler’s fallacy, such as that proposed in Rabin and Vayanos (2010) , agents may react more negatively to the previous decision if they are more certain that the previous case was a true 1 (0) because it was very high (low) in quality. Such a model would also predict that β2 < 0, and hence the two theories make identical predictions.

Tables A.XVIII and A.XIX in the Online Appendix estimate the above equation for the asylum judges and loan officers samples, respectively, and find that, using a continuous predicted quality measure for asylum cases and loan officer’s quality scores for loans, the current decision is negatively correlated with the previous decision, but not reliably related to the previous case’s quality. This is consistent with a simple gambler’s fallacy model as in Rabin (2002) and less consistent with SCE or a more general model of the gambler’s fallacy in which agents react negatively to the continuous quality of the previous case. However, the test cannot fully reject SCE because we may measure the true quality of the previous case with error. If unobserved quality is better captured by the binary decision rather than the observed continuous quality measure, then both coefficients represent quality and are consistent with both SCE and the gambler’s fallacy.

Likewise, we cannot completely rule out SCE in baseball. In principle, SCE may simply be less likely to occur in the context of baseball because there is a well-defined quality metric: the regulated strike zone. Therefore, quality is established by rule. However, there still may be room for SCE to affect perceptions of quality, at least at the margin. Further, as shown later in Table XII , umpires are slightly more likely to reverse the next call when the previous pitch was an obvious strike, that is, high quality, which is consistent with SCE.

Table XII

Baseball Umpires: Treating Teams “Fairly”

Strike Full Sample
Following Ambiguous/ Obvious
(1)(2)(3)
Lag strike × prev call correct −0.0177 ***
(0.00101)
Lag strike × prev call incorrect −0.00663 ***
(0.00130)
Lag strike × prev call obvious −0.0180 *** −0.0175 ***
(0.00189)(0.00216)
Lag strike × prev call ambiguous −0.0120 ***
(0.00123)
Lag strike × prev call not ambiguous/obvious −0.0150 ***
(0.00103)
Lag strike × prev call ambiguous and correct −0.0140 ***
(0.00175)
Lag strike × prev call ambiguous and incorrect −0.00821 ***
(0.00188)
p -value: equality 6.70e-22.00158.0000736
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741895,733476,819
R20.6650.6650.666
Strike Full Sample
Following Ambiguous/ Obvious
(1)(2)(3)
Lag strike × prev call correct −0.0177 ***
(0.00101)
Lag strike × prev call incorrect −0.00663 ***
(0.00130)
Lag strike × prev call obvious −0.0180 *** −0.0175 ***
(0.00189)(0.00216)
Lag strike × prev call ambiguous −0.0120 ***
(0.00123)
Lag strike × prev call not ambiguous/obvious −0.0150 ***
(0.00103)
Lag strike × prev call ambiguous and correct −0.0140 ***
(0.00175)
Lag strike × prev call ambiguous and incorrect −0.00821 ***
(0.00188)
p -value: equality 6.70e-22.00158.0000736
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741895,733476,819
R20.6650.6650.666

Notes. This table tests whether our results are driven by umpires reversing previous marginal or incorrect calls. Columns (1) and (2) use the sample of all consecutive called pitches. Column (3) restricts the sample to pitches following a consecutive called pitch that was either obvious or ambiguous. Prev call correct and prev call incorrect are indicator variables for whether the umpire’s previous call of strike or ball was correct or incorrect as measured by PITCHf/x. Prev call obvious in an indicator variable for whether the location of the previous called pitch was within 3 inches of the center of the strike zone or 6 inches or more outside of the edge of the strike zone. Prev call ambiguous is an indicator variable for whether the location of the previous pitch was within 1.5 inches of boundary of the strike zone. Prev call not ambiguous/obvious is an indicator equal to one if the previous pitch was neither obvious nor ambiguous. Column (3) further divides previous ambiguous calls by whether they were called correctly. This is not done for previous obvious calls because almost all (99.3%) obvious calls are called correctly as compared to 60.3% of ambiguous calls. In all columns, the reported interactions fully segment the regression sample. For example, the coefficient on “lag strike × prev call correct” represents the autocorrelation conditional on the previous call being correct and the coefficient on “lag strike × prev call incorrect” represents the autocorrelation conditional on the previous call being incorrect. p -values report tests for the equality of the reported coefficients. All control variables are as described in Table IX . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Table XII

Baseball Umpires: Treating Teams “Fairly”

Strike Full Sample
Following Ambiguous/ Obvious
(1)(2)(3)
Lag strike × prev call correct −0.0177 ***
(0.00101)
Lag strike × prev call incorrect −0.00663 ***
(0.00130)
Lag strike × prev call obvious −0.0180 *** −0.0175 ***
(0.00189)(0.00216)
Lag strike × prev call ambiguous −0.0120 ***
(0.00123)
Lag strike × prev call not ambiguous/obvious −0.0150 ***
(0.00103)
Lag strike × prev call ambiguous and correct −0.0140 ***
(0.00175)
Lag strike × prev call ambiguous and incorrect −0.00821 ***
(0.00188)
p -value: equality 6.70e-22.00158.0000736
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741895,733476,819
R20.6650.6650.666
Strike Full Sample
Following Ambiguous/ Obvious
(1)(2)(3)
Lag strike × prev call correct −0.0177 ***
(0.00101)
Lag strike × prev call incorrect −0.00663 ***
(0.00130)
Lag strike × prev call obvious −0.0180 *** −0.0175 ***
(0.00189)(0.00216)
Lag strike × prev call ambiguous −0.0120 ***
(0.00123)
Lag strike × prev call not ambiguous/obvious −0.0150 ***
(0.00103)
Lag strike × prev call ambiguous and correct −0.0140 ***
(0.00175)
Lag strike × prev call ambiguous and incorrect −0.00821 ***
(0.00188)
p -value: equality 6.70e-22.00158.0000736
Pitch locationYesYesYes
Pitch trajectoryYesYesYes
Game conditionsYesYesYes
N898,741895,733476,819
R20.6650.6650.666

Notes. This table tests whether our results are driven by umpires reversing previous marginal or incorrect calls. Columns (1) and (2) use the sample of all consecutive called pitches. Column (3) restricts the sample to pitches following a consecutive called pitch that was either obvious or ambiguous. Prev call correct and prev call incorrect are indicator variables for whether the umpire’s previous call of strike or ball was correct or incorrect as measured by PITCHf/x. Prev call obvious in an indicator variable for whether the location of the previous called pitch was within 3 inches of the center of the strike zone or 6 inches or more outside of the edge of the strike zone. Prev call ambiguous is an indicator variable for whether the location of the previous pitch was within 1.5 inches of boundary of the strike zone. Prev call not ambiguous/obvious is an indicator equal to one if the previous pitch was neither obvious nor ambiguous. Column (3) further divides previous ambiguous calls by whether they were called correctly. This is not done for previous obvious calls because almost all (99.3%) obvious calls are called correctly as compared to 60.3% of ambiguous calls. In all columns, the reported interactions fully segment the regression sample. For example, the coefficient on “lag strike × prev call correct” represents the autocorrelation conditional on the previous call being correct and the coefficient on “lag strike × prev call incorrect” represents the autocorrelation conditional on the previous call being incorrect. p -values report tests for the equality of the reported coefficients. All control variables are as described in Table IX . Standard errors are clustered by game. *, ** , and *** indicate significance at the 10%, 5%, and 1% levels, respectively.

Theoretically, the main distinction between the gambler’s fallacy and SCE lies in when the subject makes a quality assessment. Under the gambler’s fallacy, a decision maker who just reviewed a high quality case would predict that the next case is less likely to be high quality (because two high quality cases in a row are unlikely to occur) even before seeing the next case, whereas under SCE, the decision maker will make a relative comparison after seeing both cases. While the laboratory setting may be able to separate these two biases, they are observationally equivalent when looking at only decision outcomes since we cannot observe what is inside a decision maker’s head. Complicating matters further, contrast effects bias could potentially arise from the gambler’s fallacy. After reviewing a high-quality case and deciding in the affirmative, a decision maker may believe that the next case is less likely to be of high quality, and this makes her perceive the next case as indeed having lower quality, resulting in a contrast effect.

VI.B. Quotas and Learning

In all three of our empirical settings, agents do not face explicit quotas. For example, loan officers are paid based upon accuracy and are explicitly told that they do not face quotas. However, one may be concerned that decision makers self-impose quotas. Even without a self-imposed quota, decision makers may believe that the correct fraction of affirmative decisions should be some level of θ . Under a learning model, the decision maker may be unsure of where to set the quality bar to achieve an affirmative target rate of θ , and learn over time. We show that self-imposed quotas or targets are unlikely to explain our results by controlling for the fraction of the previous N decisions that were made in the affirmative, where N equals 2 or 5, and testing whether the previous single decision still matters. We find that, holding constant the fraction of the previous two or five decisions decided in the affirmative, the previous single decision negatively predicts the next decision (see Tables II, IX , and A.XIII). The only exception is the loan officers setting in which we do not find, controlling for the fraction of the past two decisions made in the affirmative, that loan officers react more negatively to the most recent decision. However, the results are less precisely estimated because the field experiment data offers a shorter panel and smaller sample size. 18 In general, this behavior is consistent with models of the gambler’s fallacy, and largely inconsistent with self-imposed quotas unless the decision maker has a very limited memory and cannot remember beyond the most recent decision. Likewise, decision makers in our three settings are highly experienced and should have a standard of quality calibrated from many years of experience. These decision makers are probably not learning much from their most recent decision. Therefore, a learning model would not predict a strong negative reaction to the most recent decision, especially if we also control for their history of recent decisions using the fraction of recent decisions decided in the affirmative. In addition, baseball umpires should make decisions according to an objective quality standard (the officially defined strike zone) rather than according to a target affirmative decision rate.

VI.C. External Perceptions and Preferences for Alternation and Fairness

Finally, we discuss two additional possible explanations for negatively autocorrelated decisions that are closely related to variants of our gambler’s fallacy hypothesis. The first is that the decision maker fully understands random processes, but cares about the opinions of others, for example, promotion committees or voters, who are fooled by randomness. These rational decision makers will choose to make negatively autocorrelated decisions, even if they know they are wrong, to avoid the appearance of being too lenient or too harsh. Concerns about external perceptions could be an important driver of decisions. However, they are unlikely to drive the results in the context of loan approval, which is an experimental setting where payouts depend only on accuracy and the ordering of decisions, and their associated accuracy are never reported to participants or their home banks.

The second related explanation is that agents may prefer to alternate being “mean” and “nice” over short time horizons. We cannot rule out this preference for mixing entirely. However, the desire to avoid being mean two times in a row, holding the recent fraction of negative decisions constant, could actually originate from the gambler’s fallacy. A decision maker who desires to be fair may over-infer that she is becoming too harsh and negative from a short sequence of “mean” decisions. Moreover, a preference to alternate mean and nice is again unlikely to drive behavior in the loan-approval setting where loan officer decisions in the experiment do not affect real loan origination (so there is no sense of being mean or nice).

An important and related consideration that is specific to the baseball setting is that umpires may have a preference to be equally nice or “fair” to two opposing teams. The desire to be fair to two opposing teams is unlikely to drive results in the asylum judges and loan officers settings because the decision maker reviews a sequence of independent cases, and the cases are not part of any teams. However, in baseball, the umpire makes sequential calls on the same team at bat. Fairness motives may lead umpires to undo a previous marginal or mistaken call, which could result in negative autocorrelation. After calling a marginal pitch a strike, the umpire may choose to balance out his calls by calling the next pitch a ball. While we do not seek to completely rule out these types of situations, we show that “make-up” calls and preferences for fairness appear unlikely to drive our estimates for baseball umpires. 19

In Table XII , column (1) shows that the negative autocorrelation is stronger following a previous correct call than following a previous incorrect call, which is inconsistent with a fairness motive because umpires concerned with fairness should be more likely to reverse the previous call if it was incorrect. Column (2) of the same table shows that the negative autocorrelation remains equally strong or stronger when the previous call was obvious. In these cases, the umpire is less likely to feel guilt about making a particular call because the umpire could not have called it any other way (e.g., he, and everyone else, knew it was the right call to make). Nevertheless, we find strong negative autocorrelation following these obvious calls, suggesting that a desire to undo marginal calls is not the sole driver of our results. Finally, in column (3) of Table XII , we restrict the sample to called pitches following previous calls that were either obvious or ambiguous. We further divide previous ambiguous calls into those that were called correctly (60%) and those that were called incorrectly (40%). If fairness concerns drive the negative autocorrelation in calls, the negative autocorrelation should be strongest following previous ambiguous and incorrect calls. We find the opposite. The negative autocorrelation is stronger following obvious calls (of which 99% are called correctly) and also following previous ambiguous calls that were called correctly. These results suggest that fairness concerns and a desire to be equally nice to two opposing teams are unlikely to explain our results.

VII. C onclusion

We document strong negative autocorrelation by decision makers, unrelated to the quality of cases, in the following three high-stakes contexts: refugee asylum courts, loan application reviews, and professional baseball umpire calls. We find consistent evidence with many common links across the three independent settings. This negative autocorrelation is stronger among more moderate and less experienced decision makers, following longer streaks of decisions in one direction, when the current and previous cases share similar characteristics or occur close in time, and when decision makers face weaker incentives for accuracy. We show that the negative autocorrelation in decision making is most consistent with the gambler’s fallacy inducing decision makers to erroneously alternate decisions because they mistakenly believe that streaks of affirmative or negative decisions are unlikely to occur by chance. We cannot rule out that sequential contrast effects also help explain these findings, but we show that the results are unlikely to be driven by other alternative explanations such as quotas, learning, or preferences to treat parties fairly.

Beyond the three settings we study, negatively autocorrelated decision making could have broader implications. For example, financial auditors, human resource interviewers, medical doctors, and policy makers all make sequences of decisions under substantial uncertainty. Our results suggest that misperceptions of what constitutes a fair process can perversely lead to unfair or incorrect decisions in many situations.

Appendix A. Calculation of Reversal and Mistake Rates

In this section we discuss how to interpret regression coefficients as approximate reversal or mistake rates. Consider the simple regression Yt=β0+β1Yt1+ϵt . Taking expectations, P(Y=1)=β01β1 . Let aβ01β1 be the rate of affirmative decisions in the data. Suppose that, absent the bias toward negative autocorrelation in decisions, the rate of affirmative decisions would still equal a . If the previous decision was a negative, then the negative autocorrelation causes the current decision to be too likely to be an affirmative by the amount (β0a) . If the previous decision was an affirmative, then the current decision is not likely enough to be an affirmative by the amount (a(β0+β1)) . Therefore, the fraction of decisions that are reversed due to the negative autocorrelation is R(β0a)·P(Yt1=0)+(a(β0+β1))·P(Yt1=1) . To simplify, substitute β0=a(1β1) , so that the previous equation simplifies to R2β1a(1a) , which is positive since β1 < 0.

If the correct decision is known, we can also estimate the fraction of decisions that are mistakes caused by the negative autocorrelation in decisions. Consider the alternative simple regression Yt=β~0+β~1Yt1+γYt,true+et . Let τE[Yt,true] be the rate of affirmative decisions in the data if all decisions were correct. Let δE[1{Y=Ytrue}] be the accuracy rate in the data. Taking expectations, P(Y=1)=β~0+γτ1β~1 . Let a~β~0+γτ1β~1 be the rate of affirmative decisions in the data. Suppose that, absent the bias toward negative autocorrelation in decisions, the rate of affirmative decisions would still equal a~ . If the previous decision was a negative, then the negative autocorrelation causes the current decision to be too likely to be an affirmative by the amount (β~0+γτa~) . If the previous decision was an affirmative, then the current decision is not likely enough to be an affirmative by the amount (a~(β~0+γτ+β~1)) . Therefore, the fraction of decisions that are reversed due to the negative autocorrelation is R~(β~0+γτa~)·P(Yt1=0)+(a~(β~0+γτ+β~1))·P(Yt1=1) . To simplify, substitute β~0+γτ=a~(1β~1) , so that the previous equation simplifies to R~2β~1a~(1a~) , which is positive since β~1<0 .

The fraction of decisions that are mistakes caused by the negative autocorrelation is approximately M=R~(δ0)R~(1δ0), where δ0=δ+M is the accuracy rate if there were no negative autocorrelation in decisions. The mistake rate is the sum of the fraction of decisions that would have been accurate but are reversed due to the negative autocorrelation in decisions, minus the fraction of decisions that would have been inaccurate but are reversed due to the negative autocorrelation in decisions. Note that, in extreme situations where the decision maker is wrong more than half the time (e.g., δ < 0.5), reversals can increase accuracy. Solving yields a mistake rate of M=(2δ1)R~12R~ .

Appendix B. A Model of Decision making under the Gambler’s Fallacy

To motivate why the gambler’s fallacy may lead to negatively correlated decision making, we present a simple extension of the Rabin (2002) model of the gambler’s fallacy and belief in the law of small numbers. In the Rabin model, agents who suffer from the gambler’s fallacy believe that, within short sequences, black (1) and white (0) balls are drawn from an imaginary urn of finite size without replacement. Therefore, a draw of a black ball increases the odds of the next ball being white. As the size of the imaginary urn approaches infinity, the biased agent behaves like the rational thinker.

We extend the model to decision making by assuming that before assessing each case, agents hold a prior belief about the probability that the case will be a black ball. This prior belief is shaped by the same mechanics as the behavioral agent’s beliefs in the Rabin model. However, the agent also receives a noisy signal about the quality of the current case, so the agent’s ultimate decision is a weighted average of her prior belief and the noisy signal.

Model Setup

Suppose an agent makes 0/1 decisions for a randomly ordered series of cases. The true case quality is an i.i.d. sequence {yt}t=1M where yt={0,1},P(yt=1)=α(0,1) , and ytyt1t .

The agent’s prior about the current case is
PtP(yt=1{yτ}τ=1t1).
For simplicity, we assume that the decision maker believes the true case quality for all cases prior to t is equal to the decision made (e.g., if the agent decided the ball was black, she believes it is black). 20
The agent also observes an i.i.d. signal about current case quality St{0,1} , which is accurate with probability μ and uninformative with probability 1 – μ . By Bayes's rule, the agent’s belief after observing S t is
P(yt=1|St,{yτ}τ=1t1)=[μSt+(1μ)α]Ptα.
The agent then imposes a threshold decision rule and makes a decision Dt{0,1} such that
Dt=1{[μSt+(1μ)α]PtαX¯}.
We then compare the prior beliefs and decisions of a rational agent to those of an agent who suffers from the gambler’s fallacy. The rational agent understands that the y t are i.i.d. Therefore, her priors are independent of history:
PtR=P(yt=1{yτ}τ=1t1)=P(yt=1)=α.
By Bayes's rule, the rational agent’s belief after observing S t is
P(yt=1|St=1,{yτ}τ=1t1)=μSt+(1μ)α.
It is straightforward to see that the rational agent’s decision on the current case should be uncorrelated with her decisions in previous cases, conditional on α .

Following Rabin (2002) , we assume that an agent who suffers from the gambler’s fallacy believes that for rounds 1, 4, 7, … cases are drawn from an urn containing N cases, αN of which are 1’s (and the remainder are 0’s). For rounds 2, 5, 8, … cases are drawn from an urn containing N – 1 cases, αN y t– 1 of which are 1’s. Finally, for rounds 3, 6, 9, … cases are drawn from an urn containing N – 2 cases, αNyt1yt2 of which are 1’s. The degree of belief in the law of small numbers is indexed by NN and we assume N6 . As N , the biased agent behaves likes the rational thinker.

Model Predictions

The simple model generates the following testable predictions for decision makers who suffer from the gambler’s fallacy:

  • Decisions will be negatively autocorrelated as long as the signal of case quality is not perfectly informative. This occurs because decisions depend on prior beliefs that are negatively related to the previous decision.

  • “Moderate” decision makers, defined as those with α close to 0.5, will make more unconditionally negatively autocorrelated decisions than extreme decision makers, defined as those with α close to 0 or 1. This follows immediately from Rabin (2002) .

  • The negative autocorrelation will be stronger following a streak of two or more decisions in the same direction. This follows from an extension of Rabin (2002) where the decision maker believes that he is making the first, second, or third draw from the urn, each with a probability of one-third.

  • The negative autocorrelation in decisions is stronger when the signal about the quality of the current case is less informative. This follows directly from the threshold decision rule defined above.

Appendix C. Additional Background on Asylum Judges

Immigration Courts Overview

The immigration judges are part of the Executive Office for Immigration Review (EOIR), an agency of the Department of Justice ( Political Asylum Immigration Representation Project 2014 ). At present, there are over 260 immigration judges in 59 immigration courts. In removal proceedings, immigration judges determine whether an individual from a foreign country (an alien) should be allowed to enter or remain in the United States or should be removed. Immigration judges are responsible for conducting formal court proceedings and act independently in deciding the matters before them. The judges also have jurisdiction to consider various forms of relief from removal. In a typical removal proceeding, the immigration judge may decide whether an alien is removable (formerly called deportable) or inadmissible under the law, then may consider whether that alien may avoid removal by accepting voluntary departure or by qualifying for asylum, cancellation of removal, adjustment of status, protection under the United Nations Convention Against Torture, or other forms of relief (Executive Office for Immigration Review 2014).

Immigration Judges

The immigration judges are attorneys appointed by the U.S. Attorney General as administrative judges, and are subject to the supervision of the Attorney General, but otherwise exercise independent judgment and discretion in considering and determining the cases before them. See INA sec. 101(b)(4) (8 U.S.C. 1101(b)(4)); 8 CFR 1003.10(b), (d). Decisions of the immigration judges are subject to review by the Board pursuant to 8 CFR 1003.1(a)(1) and (d)(1); in turn, the Board’s decisions can be reviewed by the Attorney General, as provided in 8 CFR 1003.1(g) and (h). Decisions of the Board and the Attorney General are subject to judicial review (Executive Office for Immigration Review 2014).

In our own data collection of immigration judge biographies, many previously worked as immigration lawyers or at the Immigration and Naturalization Service (INS) for some time before they were appointed. The average tenure of active immigration judges, as of 2007, was approximately 11 to 12 years. Since 2003 the annual attrition rate has averaged approximately 5%, with the majority of departures being due to retirement ( TRAC Immigration 2008 ).

Proceedings before Immigration Courts

There are two ways an applicant arrives to the Immigration Court. First, the asylum seeker can affirmatively seek asylum by filing an application. In the event that the Asylum Office did not grant the asylum application 21 and referred it to Immigration Court, the asylum seeker can now pursue his or her asylum claim as a defense against removal in Immigration Court. Second, if the asylum seeker never filed for asylum with the Asylum Office but rather the government started removal proceedings against him or her for some other reason, he or she can now pursue an asylum case in Immigration Court ( Political Asylum Immigration Representation Project 2014 ). This latter group is classified as defensive applicants and includes defendants picked up in immigration raids.

Families

We treat multiple family members as a single case because family members almost always receive the same asylum decision (based on Ramji-Nogales et al. 2007 and verified through conversations with several asylum judges). Following Ramji-Nogales et al. (2007) , we infer shared family status if cases share a hearing date, nationality, court, judge, decision, representation status, and case type (affirmative or defensive). Because our data contains some fields previously unavailable in the Ramji-Nogales et al. (2007) data, we also require family members to have the same lawyer identity code and to be heard during the same or consecutive hearing start time.

A potential concern with inferring that two applicants belong to the same family case using the criteria above is that family members must have, among the many other similarities, similar decision status. Therefore, sequential cases inferred to belong to different families will tend to have different decisions. This may lead to spurious measures of negative autocorrelation in decisions that are caused by error in the inference of families. We address this concern in two ways. First, we are much more conservative in assigning cases to families than Ramji-Nogales et al. (2007) . In addition to their criteria, we also require family members to have the same identity for their lawyer and the same or consecutive hearing start time. This will lead to under-inference of families if some family members are seen during nonconsecutive clock times or the data fails to record lawyer identity, both of which occur in the data according to conversations with TRAC data representatives. Since family members tend to have the same decision, under-inference of families should lead to biases against our findings of negative autocorrelation in decisions. Second, we find evidence of significant and strong negative autocorrelation when the current and previous case do not correspond to the same nationality. This type of negative autocorrelation is extremely unlikely to be generated by errors in the inference of families because family members will almost always have the same nationality.

Appendix D. MLB Control Variables

The empirical tests for baseball umpire decisions include the following control variables unless otherwise noted. All controls are introduced as linear continuous variables unless otherwise specified below.

  • Indicator variables for each 3 × 3 inch square for the ( x , y ) location of the pitch as it passed home plate, with (0, 0) being the lowest left box from the umpire’s perspective.

  • Indicator for whether the batter belongs to the home team.

  • Indicator for each possible pitch count combination (number of balls and strikes prior to current pitch).

  • Acceleration of the pitch, in feet per second, in the x -, y -, and z -direction measured at the initial release point (three continuous variables).

  • Break angle: The angle, in degrees, from vertical to the straight line path from the release point to where the pitch crossed the front of home plate, as seen from the catcher’s/umpire’s perspective.

  • Break length: The measurement of the greatest distance, in inches, between the trajectory of the pitch at any point between the release point and the front of home plate, and the straight line path from the release point and the front of home plate.

  • The distance in feet from home plate to the point in the pitch trajectory where the pitch achieved its greatest deviation from the straight line path between the release point and the front of home plate.

  • End speed: The pitch speed in feet per second measured as it crossed the front of home plate.

  • The horizontal movement, in inches, of the pitch between the release point and home plate, as compared to a theoretical pitch thrown at the same speed with no spin-induced movement.

  • The vertical movement, in inches, of the pitch between the release point and home plate, as compared to a theoretical pitch thrown at the same speed with no spin-induced movement.

  • The left/right distance, in feet, of the pitch from the middle of the plate as it crossed home plate (the PITCHf/x coordinate system is oriented to the catcher’s/umpire’s perspective, with distances to the right being positive and to the left being negative).

  • The height of the pitch in feet as it crossed the front of home plate.

  • The direction, in degrees, of the ball’s spin. A value of 0 indicates a pitch with no spin. A value of 180 indicates the pitch was spinning from the bottom.

  • Spin rate: The angular velocity of the pitch in revolutions per minute.

  • The velocity of the pitch, in feet per second, in the x , y , and z dimensions, measured at the initial point (three continuous variables).

  • The left/right distance, in feet, of the pitch, measured at the initial point.

  • The height, in feet, of the pitch, measured at the initial point.

  • Proportion of previous pitches to the batter during the given game that were either in the dirt or were a hit-by-pitch.

  • Proportion of previous pitches to the batter during the given game that were put into play.

  • Proportion of previous pitches to the batter during the game that were described as either swinging strike, missed bunt or classified as strike.

  • Proportion of previous pitches to the batter during the game that were described as either intentional ball, pitchout, automatic ball, or automatic strike.

  • Proportion of previous pitches to the batter during the game described as foul tip, foul, foul bunt, foul (runner going) or foul pitchout.

  • Proportion of previous pitches to the batter during the game described as “ball.”

  • Proportion of previous pitches to the batter during the game described as “called strike.”

  • Indicator variable for whether the pitch should have been called a strike based on the objective definition of the strike zone.

  • A measure developed by Tom Tango of how important a particular situation is in a baseball game depending on the inning, score, outs, and number of players on base.

  • Indicator variables for each possible score of the team at bat.

  • Indicator variables for each possible score of the team in the field.

1. Simonsohn and Gino (2013) also examine decisions in a real-world setting by looking at the scoring of MBA admissions interviews. These authors focus on narrow bracketing (dividing continuous flows of judgments into daily subsets), although they discuss the gambler’s fallacy as a potential mechanism behind their findings. While Simonsohn and Gino examine scores on a 1–5 scale, we study binary sequences of decisions, which may be a closer fit to simple binary models of the gambler’s fallacy. In addition, and importantly, we emphasize differences in reactions to the ordering of recent decisions while Simonsohn and Gino test a general narrow bracketing model in which agents react to the average score assigned previously within the same day, regardless of ordering. As we highlight, the sequencing and ordering of cases is a key distinguishing feature of the gambler’s fallacy.

2. Following Gilovich et al.’s (1985) seminal work, a number of papers have found evidence of hot hand beliefs in sports settings, although some results have been challenged in recent work, for example, Miller and Sanjurjo (2014) and Green and Zwiebel (2015) .

3. Except for the regressions with decision maker fixed effects, we never include the current observation in the calculation of averages for control variables since that could lead to a spurious negative estimated relationship between the current and previous decisions in finite panels.

4. While we will present specific solutions to the possibility that case quality is not randomly ordered in later sections, we note that most types of nonrandom ordering are likely to correspond to positive autocorrelation (e.g., slow-moving trends in refugee quality), which would bias against finding negative autocorrelation in decisions.

5. See http://www.uscis.gov/humanitarian/refugees-asylum/asylum/obtaining-asylum-united-states for more details regarding the asylum application process and defensive versus affirmative applications.

6. Exceptions to the first-in-first-out rule occur when applicants file applications on additional issues or have closures made other than grant or deny (e.g., closures may occur if the applicant does not show up, if the applicant chooses to withdraw, or for miscellaneous rare reasons encoded in the “other” category). Since these violations of first-in-first-out are likely driven by applicant behaviors often several months prior to the recent set of decisions, they are likely uncorrelated with the judge’s previous decision, which often occurs in the same or previous day. To test this, we also examine autocorrelation in proxies for case quality in the Online Appendix to assess whether deviations from the rule drive negative autocorrelation in decisions. We find nothing in this regard.

7. Following Ramji-Nogales et al. (2007) , we use the decision on the asylum application for our baseline analysis. If the judge denies asylum but grants withholding of removal or protection under CAT, the asylum applicant receives much more limited benefits than she would if she were granted asylum. In such cases, applicants face employment limitations and are only granted withholding of removal to the particular country where they may be persecuted but may be moved to a safe third country (and such protections are person-specific rather than applying to spouses or children). Therefore, it is not obvious whether a denial of asylum accompanied by a grant of withholding or protection under CAT is a positive or negative decision. Further, while the evidentiary standard for qualifying for withholding of removal or protection under CAT is much higher than those for the asylum application, the judge also exercises less subjective discretion in the determination of the former two applications, which are classified as mandatory if the applicants meet the high evidentiary standard. This is relevant for cases in which the applicant has committed crimes or assisted in the persecution of others (which disqualify her for asylum) but remains eligible for withholding of removal or protection under CAT.

8. Results, reported in the Online Appendix , are qualitatively similar using these two sets of cutoffs.

9. We also have data on decisions in which we do not know the ordering of the current case with respect to the previous case because two or more cases are considered within a single session with a single time stamp. These observations are excluded from the regression sample but are used to create control variables relating to judge average grant rates.

10. To identify the effect of experience within judges over time, we include judge fixed effects in column (4). In general, we avoid including judge fixed effects except in tables in the Online Appendix because judge fixed effects bias the coefficient on Lag grant downward. However, the coefficient on Lag grant × experienced judge remains informative, which we focus on in column (4).

11. For a detailed description of the data, we refer the interested reader to Cole et al. (2015) ; our data sample consists of a subset of the data described in their paper. This subsample was chosen by the original authors and given to us before any tests of serial correlation in decision making were conducted. Therefore, differences between the subsample and full sample should not bias the analysis in favor of our findings.

12. The incentives in the “flat” scheme may at first seem surprisingly weak, but the authors of the original experiment used this incentive condition to mimic the relatively weak incentives faced by real loan officers in India. As shown in the next table, the overall approval rate within the flat incentive scheme is only 10 percentage points higher than the approval rates under the two other incentive schemes and loan officers were still more likely to approve performing than nonperforming loans. This suggests that loan offers still chose to reject many loans and may have experienced some other intrinsic or reputational motivation to accurately screen loans.

13. The sum of the coefficients on Lag approve and Lag approve × grad approve is positive, leading to the puzzling implication that loan officers with graduate school education engage in positively autocorrelated decision making. However, our sample size is limited and the sum of the two coefficients is insignificantly different from zero.

14. Note that the fraction of loans performing is not exactly 67%, implying that the original experiment did not implement an exactly balanced session design for every session. Cole et al. (2015) initially balanced each session to have exactly four performing loans, according to early performance data given to them by the bank that originally processed the loans. However, the bank then sent the researchers a revised categorization of the loan files. The researchers used the revised data to categorize the data but did not reassign loans to sessions. This led to 85% of the sessions in our data having exactly four performing loans, 11% of the sessions having three performing loans, 3% of the sessions having five performing loans, and 1% of sessions having two performing loans.

15. Along the same lines, umpires may potentially be misled by catcher framing, in which catchers strategically try to catch a pitch close to the chest, so that the pitch appears closer to the center of the strike zone than it actually was. In general, deceptive maneuvers such as catcher framing may alter the overall rate of called strikes within a game or inning, but should not affect our results, which measure high-frequency negative autocorrelation. We test whether the current mistake in umpire decisions is negatively related to the previous call. Catcher framing should not affect negative autocorrelation in calls because catchers do not have incentives to frame more following a previous call of ball.

16. In Table A.XVI in the Online Appendix , we find qualitatively similar coefficients on Y i,t– 1 if we do not control for count.

17. Among umpires who have made more than 500 calls, the standard deviation in the mean rate of calling strikes is 0.01, potentially because extreme umpires would be much less accurate and umpire accuracy can be judged relative to the PITCHf/x system.

18. While we cannot show that loan officers react negatively to the most recent decision controlling for the fraction of recent decisions made in the affirmative, other results appear inconsistent with a quotas or learning explanation. The loan officers are paid for accuracy and they should be more likely to self-impose quotas or learn how to implement a target decision rate when they face stronger monetary incentives.

19. We also tested the effect of the last called pitch for the previous team at bat on the first called pitch for the opposing team at bat. Fairness to two teams would suggest that, if an umpire called a pitch one way or made an error in one direction against one team, then he would make that same call on the opposing team to balance it out. This implies positive autocorrelation in calls when the inning changes. We find no evidence consistent with this prediction.

20. In this simple model of the gambler’s fallacy in decision making, agents form priors based upon previous decisions. In a more general model of the gambler’s fallacy, along the lines of the model in Rabin and Vayanos (2010) , agents may react more negatively to previous decisions if they are more certain that the previous decision was correct. Such a model would yield similar predictions to those of a SCE model in which agents are more likely to reverse previous decisions if the previous case was very low or high in quality, measured continuously.

References

Angrist
Joshua D.
Pischke
Jörn-Steffen
,
Mostly Harmless Econometrics: An Empiricist’s Companion
(
Princeton, NJ: Princeton University Press
,
2008
).

Asparouhova
Elena
Hertzel
Michael
Lemmon
Michael
,
“Inference from Streaks in Random Outcomes: Experimental Evidence on Beliefs in Regime Shifting and the Law of Small Numbers,”
Management Science
,
55
(
2009
),
1766
1782
.

Ayton
Peter
Fischer
Ilan
,
“The Hot Hand Fallacy and the Gambler’s Fallacy: Two Faces of Subjective Randomness?,”
Memory & Cognition
,
32
(
2004
),
1369
1378
.

Bar-Hillel
Maya
Wagenaar
Willem A
,
“The Perception of Randomness,”
Advances in Applied Mathematics
,
12
(
1991
),
428
454
.

Benjamin
Daniel
Moore
Don
Rabin
Matthew
, “Misconceptions of Chance: Evidence from an Integrated Experiment,” Working Paper,
Harvard University
:
Cambridge, MA
.
2013
.

Bhargava
Saurabh
Fisman
Ray
,
“Contrast Effects in Sequential Decisions: Evidence from Speed Dating,”
Review of Economics and Statistics
,
96
(
2014
),
444
457
.

Bordalo
Pedro
Gennaioli
Nicola
Shleifer
Andrei
,
“Salience Theory of Judicial Decisions,”
Journal of Legal Studies
, forthcoming.

Clotfelter
Charles T.
Cook
Philip J.
,
“The ‘Gambler’s Fallacy’ in Lottery Play,”
Management Science
,
39
(
1993
),
1521
1525
.

Cole
Shawn
Kanz
Martin
Klapper
Leora
,
“Incentivizing Calculated Risk-Taking: Evidence from an Experiment with Commercial Bank Loan Officers,”
Journal of Finance
,
70
(
2015
),
537
575
.

Croson
Rachel
Sundali
James
,
“The Gambler’s Fallacy and the Hot Hand: Empirical Data from Casinos,”
Journal of Risk and Uncertainty
,
30
(
2005
),
195
209
.

Danziger
Shai
Levav
Jonathan
Avnaim-Pesso
Liora
,
“Extraneous Factors in Judicial Decisions,”
Proceedings of the National Academy of Sciences
,
108
(
2011
),
6889
6892
.

Executive Office for Immigration Review
, Office of the Chief Immigration Judge, http://www.justice.gov/eoir/ocijinfo.htm .

Gilovich
Thomas
Vallone
Robert
Tversky
Amos
,
“The Hot Hand in Basketball: On the Misperception of Random Sequences,”
Cognitive Psychology
,
17
(
1985
),
295
314
.

Gold
E.
Hester
G.
, “The Gambler’s Fallacy and a Coin’s Memory,” in
Rationality and Social Responsibility: Essays in Honor of Robyn Mason Dawes
,
Krueger
Joachim I.
, ed. (
New York
:
Psychology Press
,
2008
).

Green
Brett S.
Zwiebel
Jeffrey
, “The Hot-Hand Fallacy: Cognitive Mistakes or Equilibrium Adjustments? Evidence from Major League Baseball,” Working paper, University of California, Berkeley and Stanford Graduate School of Business,
2015
.

Guthrie
Chris
Rachlinski
Jeffrey J.
Wistrich
Andrew J.
,
“Inside the Judicial Mind,”
Cornell Law Review
,
86
(
2000
),
777
830
.

Hartzmark
Samuel M.
Shue
Kelly
, “A Tough Act to Follow: Contrast Effects in Financial Markets,” Working Paper,
University of Chicago
,
Chicago, IL
,
2016
.

Miller
Joshua Benjamin
Sanjurjo
Adam
, “A Cold Shower for the Hot Hand Fallacy,” IGIER Working Paper 518, Innocenzo Gasparini Institute for Economic Research,
2014
.

Moskowitz
Tobias
Jon Wertheim
L.
,
Scorecasting: The Hidden Influences Behind How Sports Are Played and Games Are Won,
(
New York
:
Crown
,
2014
).

Mullainathan
Sendhil
Schwartzstein
Joshua
Shleifer
Andrei
,
“Coarse Thinking and Persuasion,”
Quarterly Journal of Economics
,
123
(
2008
),
577
619
.

Parsons
Christopher
Sulaeman
Johan
Yates
Michael C.
Hamermesh
Daniel S.
,
“Strike Three: Discrimination, Incentives, and Evaluation,”
American Economic Review
,
101
(
2011
),
1410
1435
.

Pepitone
Albert
DiNubile
Mark
,
“Contrast Effects in Judgments of Crime Severity and the Punishment of Criminal Violators,”
Journal of Personality and Social Psychology
,
33
(
1976
),
448
459
.

Political Asylum Immigration Representation Project
, “Appearing at a Master Calendar Hearing in Immigration Court,” Boston, MA
2014
.

Rabin
Matthew
,
“Inference by Believers in the Law of Small Numbers,”
Quarterly Journal of Economics
,
117
(
2002
),
775
816
.

Rabin
Matthew
Vayanos
Dmitri
,
“The Gambler’s and Hot-Hand Fallacies: Theory and Applications,”
Review of Economic Studies
,
77
(
2010
),
730
778
.

Ramji-Nogales
Jaya
Schoenholtz
Andrew I.
Schrag
Philip G
,
“Refugee Roulette: Disparities in Asylum Adjudication,”
Stanford Law Review
,
60
(
2010
),
295
411
.

Rapoport
Amnon
Budescu
David V.
,
“Generation of Random Series in Two-Person Strictly Competitive Games,”
Journal of Experimental Psychology: General
,
121
(
1992
),
352
363
.

Simonsohn
Uri
,
“New Yorkers Commute More Everywhere: Contrast Effects in the Field,”
The Review of Economics and Statistics
,
88
(
2006
),
1
9
.

Simonsohn
Uri
Gino
Francesca
,
“Daily Horizons Evidence of Narrow Bracketing in Judgment From 10 Years of MBA Admissions Interviews,”
Psychological Science
,
24
(
2013
),
219
224
.

Simonsohn
Uri
Loewenstein
George
,
“Mistake #37: The Effect of Previously Encountered Prices on Current Housing Demand,”
Economic Journal
,
116
(
2006
),
175
199
.

Suetens
Sigrid
Galbo-Jorgensen
Claus B.
Tyran
Jean-Robert
,
“Predicting Lotto Numbers,”
Journal of the European Economic Association
(2015), doi: 10.1111/jeea.12147.

Terrell
Dek
,
“A Test of the Gambler’s Fallacy—Evidence from Pari-Mutuel Games,”
Journal of Risk and Uncertainty
,
8
(
1994
),
309
317
.

TRAC Immigration
, “Improving the Immigration Courts: Effort to Hire More Judges Falls Short,”
2008
. http://trac.syr.edu/immigration/reports/189/ .

Tversky
Amos
Kahneman
Daniel
,
“Belief in the Law of Small Numbers,”
Psychological Bulletin
,
76
(
1971
),
105
110
.

Tversky
Amos
Kahneman
Daniel
,
“Judgment under Uncertainty: Heuristics and Biases,”
Science
,
185
(
1974
),
1124
1131
.

Author notes

*We thank Dan Benjamin, John Campbell, Kent Daniel, Stefano Dellavigna, Andrea Frazzini, Radha Gopalan, Emir Kamenica, Adrien Matray, Sendhil Mullainathan, Josh Schwartzstein, Dick Thaler, Jeff Zwiebel, three anonymous referees, and Andrei Shleifer (the editor) for helpful comments and suggestions. We also thank seminar participants at the AEA, ANU, Conference on Behavioral and Experimental Economics, Conference on Empirical Legal Studies, Cornell, Cubist, Dartmouth, Econometric Society, Gerzensee ESSFM, Indiana University, ISNIE, McGill, NBER Behavioral Economics, Northeastern, Norwegian School of Economics, Red Rock Finance Conference, Rice, Rochester, SITE, Texas Finance Festival, University of Chicago, University of Oklahoma, University of Washington, UNSW, Yale Summer School in Behavioral Finance, and Zurich for helpful comments. Further, we thank Alex Bennett, Luca Braghieri, Leland Bybee, Sarah Eichmeyer, Chattrin Laksanabunsong, and Kaushik Vasudevan for excellent research assistance, and Sue Long for helpful discussions about the asylum court data.

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License ( http://creativecommons.org/licenses/by-nc/4.0/ ), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact [email protected]

Supplementary data