Notices of the American Mathematical Society

Welcome to the current issue of the Notices of the American Mathematical Society.
With support from AMS membership, we are pleased to share the journal with the global mathematical community.


The Piranha Problem: Large Effects Swimming in a Small Pond

Christopher Tosh
Philip Greengard
Ben Goodrich
Andrew Gelman
Aki Vehtari
Daniel Hsu

Communicated by Notices Associate Editor Richard Levine

1. Background

In this work, we discuss an inevitable consequence of having a stable system in which many explanatory variables have large effects: these variables must have large interactions which will be unlikely to cancel out each other to the extent required for general stability or predictability. We call this type of result a “piranha theorem” 7, the analogy being the folk belief that if one has a large number of piranhas (representing large effects) in a single fish tank, then one will soon be left with far fewer piranhas. If there is some outcome for which studies find large and consistent effects of many different inputs, then we can conclude that some of these effects are smaller than claimed or that multiple explanatory variables are essentially measuring the same phenomenon.

Identifying and measuring the effects of explanatory variables are central problems in statistics and drive much of the world’s scientific research. Despite the substantial effort spent on these tasks, there has been comparatively little work on addressing a related question: how many explanatory variables can have large effects on an outcome? The present work follows up on 3 and 5, considering quantitative constraints on the effects of additional variables.

Consider, for example, the problem of explaining voters’ behaviors and choices. Researchers have identified and tested the effects of internal factors such as fear, hope, pride, anger, anxiety, depression, and menstrual cycles, as well as external factors such as droughts, shark attacks, and the performance of local college football teams. Many of these findings have been questioned on methodological grounds, but they remain in the public discourse. Beyond the details of these particular studies, it is natural to ask if all of these effects can be real in the sense of representing patterns that will consistently appear in the future.

The implication of the published and well-publicized claims regarding ovulation and voting, shark attacks and voting, college football and voting, etc., is not merely that some voters are superficial and fickle. No, this literature claims that seemingly trivial or irrelevant factors have large and consistent effects, and this runs into the problem of interactions. For example, the effect on your vote of the local college football team losing could depend crucially on whether there’s been a shark attack lately, or on what’s up with your hormones on election day. Or the effect could be positive in an election with a female candidate and negative in an election with a male candidate. Or the effect could interact with your parents’ socioeconomic status, or whether your child is a boy or a girl, or the latest campaign ad, or any of the many other factors that have been studied in the evolutionary psychology and political psychology literatures. If such effects are large, it is necessary to consider their interactions, because the average effect of a factor in any particular study will depend on the levels of all the other factors in that environment. Similarly, 13 has argued against naive assumptions of causal identification in economics, where there is a large literature considering rainfall as an instrumental variable, without accounting for the implication that these many hypothesized causal pathways would, if taken seriously, represent violations of the assumptions of the model. Even if a particular experiment or observational study analyzes only one causal factor, the existence of potential interactions (as indeed are implied if one were to take the social science literature at face value) destroys the implicit assumption that an effect measured under some particular set of conditions can be interpreted as a general or persistent effect.

These concerns are particularly relevant in social science, where the search for potential causes is open-ended. Our work here is partly motivated by the replication crisis, which refers to the difficulties that many have had in trying to independently verify established findings in social and biological sciences. Many of the explanations for the crisis have focused on various methodological issues, such as low power and unrecognized researcher degrees of freedom 15. Beyond the criticisms of practice and suggested fixes, these works have also provided much-needed statistical intuition. Groups of studies that claim to have found a variety of important explanatory variables for a single outcome should be scrutinized, particularly when the dependencies among the explanatory variables have not been investigated.

This article collects several mathematical results regarding the distributions of correlations or coefficients, with the aim of fostering further work on statistical models for environments with a multiplicity of effects. What is novel in this paper is not the theorems themselves but rather viewing them in the context of trying to make sense of clusters of research studies that claim to have found large effects.

There are many ways to capture the dependence among random variables, and thus we should expect there to be a correspondingly large collection of piranha theorems. We formalize and prove piranha theorems for correlation, regression, and mutual information in Section 4. These theorems illustrate the general phenomena at work in any setting with multiple causal or explanatory variables, with implications for the replication crisis in social science.

2. Piranhas and Butterflies

A fundamental tenet of social psychology and behavioral economics, at least how it is presented in the news media, and taught and practiced in many business schools, is that small “nudges,” often the sorts of things that we might not think would affect us at all, can have big effects on behavior.

The model of the world underlying these claims is not just the “butterfly effect” that small changes can have big effects; rather, it’s that small changes can have big and predictable effects, a sort of “button-pushing” model of social science, the idea that if you do , you can expect to see .

In response to this attitude, we present the piranha argument, which states that there can be some large and predictable effects on behavior, but not a lot, because, if there were, then these different effects would interfere with each other, a “hall of mirrors” of interactions 4 that would make it hard to detect any consistent effects of anything in observational data.

In a similar vein, 2 writes:

The butterfly effect is the semi-serious claim that a butterfly flapping its wings can cause a tornado half way around the world. It’s a poetic way of saying that some systems show sensitive dependence on initial conditions, that the slightest change now can make an enormous difference later …Once you think about these things for a while, you start to see nonlinearity and potential butterfly effects everywhere. There are tipping points everywhere waiting to be tipped!

But, Cook continues, it’s not so simple:

A butterfly flapping its wings usually has no effect, even in sensitive or chaotic systems. You might even say especially in sensitive or chaotic systems. Sensitive systems are not always and everywhere sensitive to everything. They are sensitive in particular ways under particular circumstances and can otherwise be resistant to influence…. The lesson that many people draw from their first exposure to complex systems is that there are high-leverage points, if only you can find them and manipulate them. They want to insert a butterfly at just the right time and place to bring about a desired outcome. Instead, we should humbly evaluate to what extent it is possible to steer complex systems at all. We should evaluate what aspects can be steered and how well they can be steered. The most effective intervention may not come from tweaking the inputs but from changing the structure of the system.

Effects in social science vary across people and scenarios and over time, and they can be represented by probability distributions. Cook’s advice to think about “the structure of the system” echoes recommendations from the literature on statistical quality control that system-level variation puts a limit on what can be learned about the average effects of particular interventions. In the presence of possible interactions, there is no reason to expect stability of treatment effects.

3. Example: Hypothesized Effect Sizes in Social Priming

We demonstrate the possibility of quantitative analysis of the piranha problem using the example of an influential experiment from 1996 in which participants were given a scrambled-sentence task and then were surreptitiously timed when walking away from the lab. Students whose sentences included elderly-related words such as “worried,” “Florida,” “old,” and “lonely” walked an average of 13% more slowly than students in the control condition, and the difference was statistically significant.

This experimental claim is of historical interest in psychology in that, despite its implausibility, it was taken seriously for many years and received thousands of citations, but it failed to replicate and is no longer generally believed to represent a real effect; for background see 19. Now we understand such apparently statistically-significant findings as the result of selection with many researcher degrees of freedom 15. Here, though, we will take the published claim at face value and also work within its larger theoretical structure, under which weak indirect stimuli can produce large effects.

An effect of 13% on walking speed is not in itself huge; the difficulty comes when considering elderly-related words as just one of many potential stimuli. Here are just some of the factors that have been published in the social priming and related literatures as having large effects on attitudes and behavior: hormones, subliminal images, news of football games and shark attacks, a chance encounter with a stranger, parental socioeconomic status, weather, the last digit of one’s age, the sex of a hurricane name, the sexes of siblings, the position in which a person is sitting, and many others. See 8 for references to these claims, along with other papers criticizing or refuting them. A common feature of these examples is that the stimuli have no clear direct effect on the measured outcomes, and in many cases the experimental subject is not even aware of the manipulation. Based on these examples, one can come up with dozens of other potential stimuli that fit the pattern. In addition to walking speed being affected by elderly-related words, one could also consider word length (with longer words corresponding to slower movement), sounds of words (with smooth sibilance motivating faster walking), subject matter (sports-related words as compared to sedentary words), affect (happy compared to sad words, or calm compared to angry), words related to travel (inducing faster walking) or invoking adhesives such as tape or glue (inducing slower walking), and so on. Similarly, one can consider different sorts of incidental events, not just encounters with strangers but also a ringing phone or knocking at the door, the presence of a male or female lab assistant (which could have a main effect or interact with the participant’s sex), a newspaper or magazine on a nearby table, ad infinitum.

Now we can invoke the piranha principle. Imagine 100 possible stimuli, each with an effect of 13% on walking speed, all of which could arise in a real-world setting where we encounter many sources of text, news, and internal and external stimuli. If each stimulus corresponds to two equally probable states with effects of on log walking speed, and these effects are independent in the wild, then the sum of these will be approximately normally distributed with standard deviation . Thus walking speed could easily be multiplied or divided by based on a collection of arbitrary stimuli that are imperceptible to the person being affected. And this factor of 1.8 could be made arbitrarily large by simply increasing the number of potential primes.

It is outrageous to think that walking speed could be doubled or halved based on a random collection of unnoticed and essentially irrelevant stimuli—but that is the implication of the embodied cognition literature. It is basically a Brownian motion model in which the individual inputs are too large to work out.

We can think of five ways to avoid the ridiculous conclusion. The first possibility is that the different factors could interact or interfere in some way so that the variance of the total effect is less than the sum of the variances of the individual components. Second, effects could be much smaller. Change those 13% effects to 1% effects and you can get to more plausible totals, in the same way that real-world Brownian oscillations are tolerable because the impact of each individual molecule in the liquid is so small. Third, one could reduce the total number of possible influences. If there were only 10 possible stimuli rather than 100 or 1000 or more, then the total effect could fall within the range of plausibility. Fourth, there could be a distribution of effects with a few large influences and a long tail of relatively unimportant factors, so that, when correctly translated to standardized population effect sizes, most treatment effects are already small, and the infinite sum has a reasonable bound. Fifth, multiple explanatory variables could be essentially measuring the same phenomenon.

All these options have major implications for the study of social priming and, more generally, for causal inference in an open-ended setting with large numbers of potential influences. If large interactions are possible, this suggests that stable individual treatment effects might be impossible to find: a 13% effect of a particular intervention in one particular experiment might be in another context or in the presence of some other unnoticed factor, and this in turn raises questions about the relevance of any particular study. If effects are much smaller than reported, this suggests that existing studies are extremely underpowered, so that published estimates are drastically overestimated and often in the wrong direction 9, thus essentially noise. At the same time, a restriction of the universe of potential stimuli would require an overhaul of the underlying theoretical framework in which just about any stimulus can cause a noticeable change. For example, if we think there cannot be more than five or ten large effects on walking speed, it would seem a stretch that unnoticed words in a sentence scrambling test would be one of these special factors. Finally, if the distribution of average effects is represented by a long series, most of whose elements are tiny (either because of small individual effects or because any large effects occur infrequently in the general population, as with rare diseases or short-term interventions with rapidly decaying effects), this implies a prior distribution on average effect sizes with a spike near zero, which in turn would result in near-zero estimated population effect sizes in most cases. Our point is not that all effects are zero but rather that in a world of essentially infinitely many possible causal factors, some external structure must be applied in order to identify stable effects from finite samples.

4. Piranha Theorems

In this section, we present piranha theorems for linear and nonlinear effects. We consider two different ways of measuring linear effects. We first show that it is impossible for a large number of explanatory variables to be correlated with some outcome variable unless they are highly correlated with each other. Second, we show that if a set of explanatory random variables is plugged into a linear regression, the -norm of the least-squares coefficient vector can be bounded above in terms of the eigenvalues of the second-moment matrix of the predictors. Thus, there can only be so many individual coefficients with a large magnitude. Finally, we consider a general nonlinear form of dependency, mutual information, and present a corresponding piranha theorem for that measure.

4.1. Correlation

The first type of pattern we consider is correlation. In particular, we will show that if all the covariates are highly correlated with some outcome variable, then there must be a reasonable amount of correlation among the covariates themselves. This is formalized in the following theorem, which is known as Van der Corput’s inequality 18. We offer a proof here for completeness.

Theorem 1 (Van der Corput’s inequality).

If are real-valued random variables with finite nonzero variance, then

In particular, if for each , then .

Proof.

Without loss of generality, we may assume that have mean zero and unit variance. Define by

Thus and for each . By Cauchy-Schwarz,

Therefore,

Rearranging gives us the theorem statement.

A direct consequence of Theorem 1 is that if are independent (or uncorrelated) random variables and each has correlation at least with , then .

In some situations, the outcome may change from study to study, for example a program evaluation in economics might look at employment, income, or savings; a political intervention might target turnout or vote choice; or an education experiment might look at several tests. Although the different outcomes in a study are not exactly the same, we might reasonably expect them to be highly correlated. However, if we have mean-zero and unit-variance random variables satisfying and , then

and, by Cauchy-Schwarz,

Thus, . This gives the following corollary of Theorem 1.

Corollary 2.

Suppose are real-valued random variables with finite nonzero variance. If and for , then .

The bound in Theorem 1 is essentially tight for large . To see this, pick any and take to be mean-zero random variables with covariance matrix given by

If , then for each ,

One drawback of Theorem 1 is that the upper bound depends on a coarse measure of interdependence of the covariates, namely the sum of all pairwise correlations . One might hope that if we have a finer-grained control on the correlation matrix, we should be able to get a stronger result. This is accomplished by the following piranha theorem, which shows that we can instead get an upper bound that depends on the largest eigenvalue of the correlation matrix. However, this comes at the expense of bounding the sum of squared correlations , rather than the sum of their absolute values.

Theorem 3.

If are real-valued random variables with finite nonzero variance, then

where is the maximum eigenvalue of the correlation matrix .

Consider again the case where are uncorrelated, and each has correlation at least with . In this case, the correlation matrix will be the identity matrix, whose largest eigenvalue is 1, and Theorem 3 implies that , which was the same conclusion provided by Theorem 1. However, in general Theorems 1 and 3 are incomparable since , but

As an example of when these theorems can produce different conclusions, one can give a randomized construction of a correlation matrix , where the columns of are drawn from the uniform distribution over the hypersphere . In this case, if each covariate has correlation at least with , then with high probability the conclusion of Theorem 1 is that is bounded above on the order of , while Theorem 3 gives a much tighter bound on the order of .

The proof of Theorem 3 relies on the following technical lemma, essentially a consequence of orthogonality.

Lemma 4.

If are real-valued random variables with mean zero and unit variance such that for all , then .

Proof.

Denote the covariance matrix of the random vector as

where for . Define the vector . Then

where the inequality follows from the fact that is a covariance matrix and hence positive semidefinite. We conclude that .

With the above in hand, we turn to the proof of Theorem 3.

Proof of Theorem 3.

Assume without loss of generality that have mean zero and unit variance. Denote the eigendecomposition of as , where and is orthogonal.

Let , where . Then is a mean-zero random vector whose covariance matrix is . For with , we have almost surely. We then apply Lemma 4 to get

Then,

where we have used the fact that is orthogonal.

4.2. Linear regression

We next turn to showing that least squares linear regression solutions cannot have too many large coefficients. Specifically, letting denote the regression coefficients of least squared error,

we bound the number of ’s that can have large magnitude. This is formalized in our next piranha theorem.

Theorem 5.

Suppose are real-valued random variables with mean zero and unit variance. If satisfies equation 1, then the squared norm of satisfies

where is the minimum eigenvalue of the second-moment matrix of .

Consider again the setting where are standardized and uncorrelated. In this case, the second-moment matrix will be the identity matrix, and its minimum eigenvalue will be . Thus, Theorem 5 states for independent covariates, there may be at most regression coefficients with magnitude larger than . In general, cannot get small without the explanatory variables having sizeable correlations with each other.

Proof of Theorem 5.

The case where is trivial. Thus, assume . In this case, the second-moment matrix is invertible, its inverse has eigenvalues bounded above by , and

Define , so is a vector of mean-zero and unit-variance random variables with for all . By Lemma 4,

Therefore,

where the first inequality uses the upper bound of on the eigenvalues of .

4.3. Mutual information

Though many statistical analyses hinge on discovering linear relations among variables, not all do. Thus, we turn to a more general form of dependency for random variables, mutual information. Our mutual information piranha theorem will be of a similar form as the previous results, namely that if many covariates share information with a common variable, then they must share information among themselves.

To simplify our analysis, we assume that all the random variables we consider in this section take values in discrete spaces. For two random variables and , their mutual information is defined as

where and denote entropy and conditional entropy, respectively. These are defined as

where (resp. ) is the range of (resp. ), is the joint probability mass function of and , and is the marginal probability mass function of .

We use the following facts about entropy and conditional entropy.

Fact (Chain rule of entropy).

For random variables ,

Moreover, for any other random variable ,

Fact (Conditioning reduces entropy).

For random variables ,

Using these facts, we can prove the following piranha theorem about mutual information.

Theorem 6.

Given random variables and , we have

where .

Proof.

Using the definition of mutual information, we have

Since conditioning reduces entropy, this implies

Then, by the chain rule of entropy,

The chain rule of entropy combined with the fact that conditioning reduces entropy implies

Plugging equations 2 and 3 into our formula for gives

Now we can also write,

Rearranging yields the theorem.

One corollary of Theorem 6 is that for any random variable , there can be at most random variables that (a) are mutually independent and (b) satisfy .

5. Correlations in a Finite Sample

We now turn our focus back to correlations, this time in a finite sample. Suppose we conduct a survey with data on predictors and one outcome of interest on a random sample of people, and then we evaluate the correlations between the outcome and each of the predictors.

We collect the data in an matrix with , where each of the columns of has mean zero and unit norm, and we will use for (neither in the span of the all-ones vector ) to denote the sample correlation:

where and .

An application of Theorem 3 tells us that any non-constant vector satisfies

where denote the singular values of . Moreover, it is not hard to see there exists a vector that achieves the upper bound, namely the top singular vector of . This analysis shows a worst-case piranha theorem: a bound on the number of large correlations with all possible response vectors. Stronger results can be obtained if we consider average behavior. Here, we consider a stochastic piranha theorem in which we assume that is uniformly distributed on the unit sphere in . Our result will hold for any choice of radially symmetric random vector that is independent of , but we state it for the uniform distribution over the unit sphere for concreteness. We choose a radially symmetric distribution because we have no reason to give preference to one direction over another. Recall the value of studying average as well as worst-case behavior in areas such as random matrix theory.

The following theorem demonstrates this principle, showing that the maximum sum of squared correlations, an quantity in , is generally much larger than the expected sum of square correlations. Specifically, the following theorem shows that the expected sum of squared correlations decays like .

Theorem 7.

Let be uniformly distributed on the unit sphere in . Then

If is uniformly distributed on the unit sphere in , then for large , the distribution of is well approximated by the -dimensional multivariate Gaussian with mean zero and covariance . In particular, is spherically symmetric, and

As a consequence, for large , the distribution of sum of squared correlations is well approximated by a linear combination of independent random variables, each with one degree of freedom: .

Combining this observation with Theorems 3 and 7, for any matrix (or sample of data) , if a vector is distributed according to a spherically symmetric distribution, then is supported on , has expectation , and for large has variance.

6. Discussion and Directions for Future Work

The piranha problem is a practical issue: as discussed in the references in Sections 1 and 3, it has interfered with research in fields including social priming, evolutionary psychology, economics, and voting behavior. An understanding of the piranha problem can be a helpful step in recognizing fundamental limitations of research in these fields along with related areas of application such as marketing and policy nudges 17. We suspect that a naive interpretation of the butterfly effect has led many researchers and policymakers to believe that there can be many large and persistent effects; thus, there is value in exploring the statistical reasons why this is not likely. In this way, the piranha problem resembles certain other statistical phenomena such as regression to the mean and the birthday coincidence problem, in that there is a regularity in the world that surprises people, and this regularity can be understood as a mathematical result. This motivates us to seek theorems that capture some of this regularity in a rigorous way. We are not all the way there, but this seems to us to be a valuable research direction.

6.1. Bridging between deterministic and probabilistic piranha theorems

Are there connections between the worst-case bounds in Section 4, constraints on main effects and interactions 14, the probabilistic bounds in Section 5, priors for the effective number of nonzero coefficients, and models such as the parameterization of linear regression as proposed by 20? We can consider two directions. The first is to consider departures from parametric models such as the multivariate normal and and work out their implications for correlations and regression coefficients. The second approach is to obtain limiting results in high dimensions (that is, large numbers of predictors), by analogy to central limit theorems of random matrices. The idea here would be to consider a matrix and then pull out one of the columns at random and consider it as the outcome, , with the other columns being the predictors, . One should also be able to connect this with work such as 6 and 12 on how regression coefficients change when new predictors are added to a model.

6.2. Regularization, sparsity, and Bayesian prior distributions

There has been research from many directions on regularization methods that provide soft constraints on models with large numbers of parameters. By “soft constraints,” we mean that the parameters are not literally constrained to fall within any finite range, but the estimates are pulled toward zero and can only take on large values if the data provide strong evidence in that direction.

Examples of regularization in non-Bayesian statistics include wavelet shrinkage, lasso regression, estimates for overparameterized image analysis and deep learning networks, and models that grow in complexity with increasing sample size. In a Bayesian context, regularization can be implemented using weakly informative prior distributions 11 or more informative priors that can encode the assumed sparsity 1. Classical regularization is motivated by the goal of optimizing long-run frequency performance, and Bayesian priors represent additional information about parameters, coded as probability distributions. The various piranha theorems correspond to different constraints on these priors and thus even weakly informative priors should start by encoding these constraints.

From a different direction is the idea that any given data might allow only some small number of effects or, more generally, a low-dimensional structure, to be reliably learned. More generally, models such as the horseshoe 1 assume a distribution of effect sizes with a sharp peak near zero and a long tail, which represent a solution to the piranha problem by allowing a large number of predictors without overflowing variance.

6.3. Nonlinear models

So far we have discussed linear regression, with theorems capturing different aspects of the constraint that the total cannot exceed 1 (or some bound less than 1, if some of the variation is by its nature unexplainable because it comes from a random process). We can make similar arguments for nonlinear regression.

For example, consider a model of binary data with 20 causal inputs, each of which is supposed to have an independent effect of 0.5 on the logistic scale. Aligning these factors in the same direction would give an effect of 10, enough to change the probability from 0.01 to 0.99, which would be unrealistic in applied fields ranging from marketing to voting where no individual behavior can be predicted to that level of accuracy. One way to avoid these sorts of extreme probabilities would be to suppose the predictors are highly negatively correlated with each other, but in practice, input variables in social science tend to be positively, not negatively correlated (consider, for example, conservative political ideology, Republican Party identification, and various issue attitudes that predict Republican vote choice and have positive correlations among the population of voters). The only other alternative that allows one to keep the large number of large effects is for the model to include strong negative interactions, but then the effects of the individual inputs would no longer be stable, and any effect would depend very strongly on the conditions of the experiment in which it is studied. It should be possible to express this reasoning more formally, perhaps in a way similar to long-range dependence models in time series and spatial processes.

6.4. Implications for social science research

Although we cannot directly apply these piranha theorems to data, we see them as providing some relevance to social science reasoning.

As noted at the beginning of this article, there has been a crisis in psychology, economics, and other areas of social science, with prominent findings and apparently strong effects that do not appear in attempted replications by outside research groups; see, for example, 10. The replication crisis involves many challenges, including estimating its scale and scope, identifying the statistical errors and questionable research practices that have led researchers to systematically overestimate effect sizes and be overconfident in their findings, and studying the incentives of the scientific publication process that can allow entire subfields to get lost in the interpretation of noise. Even when individual effects are large, they can apply just to a small subset of the population or just for a short period of time, not leaving persistent effects.

The research reviewed in the present article is related to, but different from, the cluster of ideas corresponding to multiple comparisons, false discovery rates, and multilevel models. Those theories correspond to statistical inference in the presence of some specified distribution of effects, possibly very few nonzero effects (the needle-in-a-haystack problem) or possibly an entire continuous distribution, but without necessarily any concern about how these effects interact.

The present article goes in a different direction, asking the theoretical question: under what conditions is it possible for many large and persistent effects to coexist in a multivariate system? In different ways, our results rule out or make extremely unlikely the possibility of multiple large effects or “piranhas” among a set of random variables. These theoretical findings do not directly call into question any particular claimed effect, but they raise suspicions about a model of social interactions in which many large effects are swimming around, just waiting to be captured by researchers who cast out the net of a quantitative study.

To more directly connect our theorems with social science would require modeling predictor and outcome variables in a subfield, similar to multiverse analysis 16. Bounds can be strengthened with reference to empirical correlations among predictors being considered. When conducting systematic reviews of evidence, it could be appropriate to consider the potential interactions among various hypothesized causal factors, rather than attempting to combine separate estimates using meta-analysis. Any general implications for social science would only become clear after consideration of particular research areas.

A. Proof of Theorem 7

For any such that for all (i.e., is not in the span of ), we write to denote the “standardized” vector given by the formula,

The unit vector in is orthogonal to . Using this notation, we have

for any not in the span of .

Write the singular value decomposition of as

where are orthonormal left singular vectors of , are orthonormal right singular vectors of , and are the singular values of .

Recall that we assume satisfy and for all . This implies the following lemma.

Lemma 8.

for all , and for all .

Proof.

The assumption on implies that for each . Moreover, the assumptions imply that the all-ones vector is orthogonal to the range of , which is spanned by . Hence for each as well.

We then take advantage of the following lemma for expressing the sum of squared correlations.

Lemma 9.

For any vector such that for all ,

Proof.

By direct computation:

Proof of Theorem 7.

By Lemma 8, the vectors are orthogonal to the unit vector . We extend the collection of orthonormal vectors with orthonormal unit vectors to obtain an orthonormal basis for . With probability , the random vector is not in the span of . Hence, is well defined and can be written uniquely as a linear combination of the aforementioned basis vectors:

where

and

(since is a unit vector). In particular, , which by symmetry implies that for each . Then, by Lemma 9,

Since are the eigenvalues of and the columns of have unit norm,

because the trace of is equal to the sum of its eigenvalues.

Acknowledgments

We thank Lauren Kennedy, the editor, and four reviewers for helpful discussion. Christopher Tosh and Daniel Hsu acknowledge NSF grant CCF-1740833, a JP Morgan Faculty Award, and a Sloan Research Fellowship. Much of this work was done while Christopher Tosh was at Columbia University. Andrew Gelman acknowledges ONR grant N000142212648. Ben Goodrich acknowledges NSF grants 2051246 and 2153019. Aki Vehtari acknowledges Research Council of Finland project 340721.

References

[1]
C. M. Carvalho, N. G. Polson, and J. G. Scott, The horseshoe estimator for sparse signals, Biometrika 97 (2010), 465–480.
[2]
J. Cook, The other butterfly effect, John D. Cook Consulting, 2018, https://www.johndcook.com/blog/2018/08/07/the-other-butterfly-effect/.
[3]
J. Cornfield, W. Haenszel, E. C. Hammond, A. M. Lilienfeld, M. B. Shimkin, and E. L. Wynder, Smoking and lung cancer: Recent evidence and a discussion of some questions, Journal of the National Cancer Institute 22 (1959), 173–203.
[4]
L. J. Cronbach, Beyond the two disciplines of scientific psychology, American Psychologist 30 (1975), 116–127.
[5]
Peng Ding and Tyler J. VanderWeele, Generalized Cornfield conditions for the risk difference, Biometrika 101 (2014), no. 4, 971–977, DOI 10.1093/biomet/asu030. MR3286930,
Show rawAMSref \bib{ding:2014}{article}{ label={Ding and Vanderweele(2014)}, author={Ding, Peng}, author={VanderWeele, Tyler J.}, title={Generalized Cornfield conditions for the risk difference}, journal={Biometrika}, volume={101}, date={2014}, number={4}, pages={971--977}, issn={0006-3444}, review={\MR {3286930}}, doi={10.1093/biomet/asu030}, }
[6]
K. A. Frank, Impact of a confounding variable on a regression coefficient, Sociological Methods and Research 29 (2002).
[7]
A. Gelman, The piranha problem in social psychology / behavioral economics: The “take a pill” model of science eats itself, 2017, https://statmodeling.stat.columbia.edu/2017/12/15/eats/.
[8]
A. Gelman, Here are just some of the factors that have been published in the social priming and related literatures as having large effects on behavior, 2023, https://statmodeling.stat.columbia.edu/2023/11/25/some/.
[9]
A. Gelman and J. B. Carlin, Beyond power calculations: Assessing type S (sign) and type M (magnitude) errors, Perspectives on Psychological Science 9 (2014), 641–651.
[10]
M. Gordon, D. Viganola, M. Bishop, Y. Chen, A. Dreber, B. Goldfedder, F. Holzmeister, M. Johannesson, Y. Liu, C. Twardy, J. Wang, and T. Pfeiffer, Are replication rates the same across academic fields? Community forecasts from the DARPA SCORE programme, Royal Society Open Science 7 (2020), 200566.
[11]
Sander Greenland and Mohammad Ali Mansournia, Penalization, bias reduction, and default priors in logistic and related categorical and survival regressions, Stat. Med. 34 (2015), no. 23, 3133–3143, DOI 10.1002/sim.6537. MR3402580,
Show rawAMSref \bib{greenland2015}{article}{ label={Greenland and Mansournia(2015)}, author={Greenland, Sander}, author={Mansournia, Mohammad Ali}, title={Penalization, bias reduction, and default priors in logistic and related categorical and survival regressions}, journal={Stat. Med.}, volume={34}, date={2015}, number={23}, pages={3133--3143}, issn={0277-6715}, review={\MR {3402580}}, doi={10.1002/sim.6537}, }
[12]
Samson Abramsky, Logic and geometry of agents in agent-based modeling, Complex social and behavioral systems—game theory and agent-based models, Encycl. Complex. Syst. Sci., Springer, New York, [2020] ©2020, pp. 705–723, DOI 10.1007/978-3-642-27737-5_313-4. MR4365626,
Show rawAMSref \bib{knaeble2020}{article}{ label={Knaeble et~al.(2020)Knaeble, Osting, and Abramson}, author={Abramsky, Samson}, title={Logic and geometry of agents in agent-based modeling}, conference={ title={Complex social and behavioral systems---game theory and agent-based models}, }, book={ series={Encycl. Complex. Syst. Sci.}, publisher={Springer, New York}, }, date={[2020] \copyright 2020}, pages={705--723}, review={\MR {4365626}}, doi={10.1007/978-3-642-27737-5\_313-4}, }
[13]
J. Mellon, Rain, rain, go away: 137 potential exclusion-restriction violations for studies using weather as an instrumental variable, 2020, https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3715610.
[14]
W. M. Rogers, Theoretical and mathematical constraints of interactive regression models, Organizational Research Methods 5 (2002).
[15]
J. P. Simmons, L. D. Nelson, and U. Simonsohn, False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant, Psychological Science 22 (2011), 1359–1366.
[16]
S. Steegen, F. Tuerlinckx, A. Gelman, and W. Vanpaemel, Increasing transparency through a multiverse analysis, Perspectives on Psychological Science 11 (2016), 702–712.
[17]
B. Szszi, A. B. C. Anthony C. Higney, A. Gelman, I. Ziano, B. Aczel, D. G. Goldstein, D. S. Yeager, and E. Tipton, No reason to expect large and consistent effects of nudge interventions, Proceedings of the National Academy of Sciences 119 (2022), e2200732119.
[18]
T. Tao, When is correlation transitive?, What’s New, 2014, https://terrytao.wordpress.com/2014/06/05/when-is-correlation-transitive.
[19]
E. Wagenmakers, R. Wetzels, D. Borsboom, R. Kievit, and H. L. J. van der Maas, A skeptical eye on psi, in Extrasensory Perception: Support, Skepticism, and Science, Praeger, Santa Barbara, CA, 2015, 153–176.
[20]
Yan Dora Zhang, Brian P. Naughton, Howard D. Bondell, and Brian J. Reich, Bayesian regression using a prior on the model fit: the R2-D2 shrinkage prior, J. Amer. Statist. Assoc. 117 (2022), no. 538, 862–874, DOI 10.1080/01621459.2020.1825449. MR4436318,
Show rawAMSref \bib{zhang2020r2d2}{article}{ label={Zhang et~al.(2020)Zhang, Naughton, Bondell, and Reich}, author={Zhang, Yan Dora}, author={Naughton, Brian P.}, author={Bondell, Howard D.}, author={Reich, Brian J.}, title={Bayesian regression using a prior on the model fit: the R2-D2 shrinkage prior}, journal={J. Amer. Statist. Assoc.}, volume={117}, date={2022}, number={538}, pages={862--874}, issn={0162-1459}, review={\MR {4436318}}, doi={10.1080/01621459.2020.1825449}, }

Credits

Photo of Christopher Tosh is courtesy of Christopher Tosh.

Photo of Philip Greengard is courtesy of Philip Greengard.

Photo of Ben Goodrich is courtesy of Ben Goodrich.

Photo of Andrew Gelman is courtesy of Andrew Gelman.

Photo of Aki Vehtari is courtesy of Lasse Lecklin, Aalto University.

Photo of Daniel Hsu is courtesy of Jaime Quan.