This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\DeclareLabeldate

date origdate n.d.

Causality in cognitive neuroscience: concepts, challenges, and distributional robustness

Sebastian Weichwald & Jonas Peters
(July 2020)
Abstract

While probabilistic models describe the dependence structure between observed variables, causal models go one step further: they predict, for example, how cognitive functions are affected by external interventions that perturb neuronal activity. In this review and perspective article, we introduce the concept of causality in the context of cognitive neuroscience and review existing methods for inferring causal relationships from data. Causal inference is an ambitious task that is particularly challenging in cognitive neuroscience. We discuss two difficulties in more detail: the scarcity of interventional data and the challenge of finding the right variables. We argue for distributional robustness as a guiding principle to tackle these problems. Robustness (or invariance) is a fundamental principle underlying causal methodology. A causal model of a target variable generalises across environments or subjects as long as these environments leave the causal mechanisms intact. Consequently, if a candidate model does not generalise, then either it does not consist of the target variable’s causes or the underlying variables do not represent the correct granularity of the problem. In this sense, assessing generalisability may be useful when defining relevant variables and can be used to partially compensate for the lack of interventional data.

1 Introduction

Cognitive neuroscience aims to describe and understand the neuronal underpinnings of cognitive functions such as perception, attention, or learning. The objective is to characterise brain activity and cognitive functions, and to relate one to the other. The submission guidelines for the Journal of Cognitive Neuroscience, for example, state: “The Journal will not publish research reports that bear solely on descriptions of function without addressing the underlying brain events, or that deal solely with descriptions of neurophysiology or neuroanatomy without regard to function.” We think that understanding this relation requires us to relate brain events and cognitive function in terms of the cause-effect relationships that govern their interplay. A causal model could, for example, describe how cognitive functions are affected by external interventions that perturb neuronal activity (cf. Section 2.1). [86] argue that “the ultimate phenomenon of theoretical interest in all FC [functional connectivity] research is understanding the causal interaction among neural entities”.

Causal inference in cognitive neuroscience is of great importance and perplexity. This motivates our discussion of two pivotal challenges. First, the scarcity of interventional data is problematic as several causal models may be equally compatible with the observed data while making conflicting predictions only about the effects of interventions (cf. Section 3.1). Second, the ability to understand how neuronal activity gives rise to cognition depends on finding the right variables to represent the neuronal activity (cf. Section 3.2). Our starting point is the well-known observation that causal models of a target (or response) variable are distributionally robust and thus generalise across environments, subjects, and interventional shifts [37, 1, 71]. Models that do not generalise are either based upon the wrong variables that do not represent causal entities or include variables that are not causes of the target variable. We thus propose to pursue robust (or invariant) models. That way, distributional robustness may serve as a guiding principle towards a causal understanding of cognitive function and may help us tackle both challenges mentioned above.

1.1 Running examples

We consider the following simplified examples. Assume that the consumption of alcohol affects reaction times in a cognitive task. In a randomised controlled trial we find that drinking alcoholic (versus non-alcoholic) beer results in slowed reaction times hereinafter. Therefore, we may write ‘alcoholreaction time\text{alcohol}\to\text{reaction time}’ and call alcohol a cause of reaction time and reaction time an effect of alcohol. Intervening on the cause results in a change in the distribution of the effect. In our example, prohibiting the consumption of any alcoholic beers results in faster reaction times.

In cognitive neuroscience one may wish to describe how the neuronal activity is altered upon beer consumption and how this change in turn affects the reaction time. For this, we additionally require a measurement of neuronal activity, say a functional magnetic resonance imaging (fMRI) scan and voxel-wise blood-oxygen-level dependent (BOLD) signals, that can serve as explanans in a description of the phenomenon ‘alcoholneuronal activityreaction time\text{alcohol}\to\text{neuronal activity}\to\text{reaction time}’. We distinguish the following two scenarios:

Running Example A, illustrated in Figure 1(a).

A so-called treatment or stimulus variable TT (say, consumption of alcohol) affects neuronal activity as measured by a dd-dimensional feature vector 𝐗=[X1,,Xd]\mathbf{X}=[X_{1},\dots,X_{d}]^{\top} and the target variable YY reflects a cognitive function (say, reaction time). We may concisely write T𝐗YT\to\mathbf{X}\to Y for a treatment that affects neuronal activity which in turn maintains a cognitive function [[, this is analogous to the ‘stimulus \to brain activity \to response’ set-up considered in]]weichwald2014causal,weichwald2015causal.

Running Example B, illustrated in Figure 1(b).

We may wish to describe how neuronal entities cause one another and hence designate one such entity as the target variable YY. In this example, we consider a target variable corresponding to a specific brain signal or region instead of a behavioural or cognitive response.

Refer to caption
TTHHX1X_{1}X2X_{2}X3X_{3}YY???
(a) Running Example A
Refer to caption
TTHHX1X_{1}X2X_{2}YY??
(b) Running Example B
Figure 1: Illustration of two scenarios in cognitive neuroscience where we seek a causal explanation focusing on a target variable YY that either resembles (a) a cognitive function, or (b) a neuronal entity. The variables TT, 𝐗=[X1,,Xd]\mathbf{X}=[X_{1},\dots,X_{d}]^{\top}, and HH represent treatment, measurements of neuronal activity, and an unobserved variable, respectively.

1.2 Existing work on causality in the context of cognitive neuroscience

Several methods such as Granger causality or constraint-based methods have been applied to the problem of inferring causality from cognitive neuroscience data. We describe these methods in Section 2.3. In addition, there are ongoing conceptual debates that revolve around the principle of causality in cognitive neuroscience, some of which we now mention. [65] raise concerns about the “lure of causal statements” and expound the problem of confounding when interpreting functional connectivity. Confounders are similarly problematic for multi-voxel pattern analyses [106, 122]. The causal interpretation of encoding and decoding (forward and backward, univariate and multivariate) models has received much attention as they are common in the analysis of neuroimaging data: [25] examine the differences between the model types, [38] point out that the weights of linear backward models may be misleading, and [119] extend the latter argument to non-linear models and clarify which causal interpretations are warranted from either model type. Feature relevance in mass-univariate and multivariate models can be linked to marginal and conditional dependence statements that yield an enriched causal interpretation when both are combined [119]; this consideration yields refined results in neuroimaging analyses [45, 3, 112] and explains improved functional connectivity results when combining bivariate and partial linear dependence measures [89]. Problems such as indirect measurements and varying temporal delays complicate causal Bayesian network approaches for fMRI [83, 68]. [99] present a simulation study evaluating several methods for estimating brain networks from fMRI data and demonstrate that identifying the direction of network links is difficult. The discourse on how to leverage connectivity analyses to understand mechanisms in brain networks is ongoing [110, 115, 98, 66]. Many of the above problems and findings are related to the two key challenges that we discuss in Section 3.

1.3 Structure of this work

We begin Section 2 by formally introducing causal concepts. In Section 2.1, we outline why we believe there is a need for causal models in cognitive neuroscience by considering what types of questions could be answered by an OraCle Modelling (OCM) approach. We discuss the problem of models that are observationally equivalent yet make conflicting predictions about the effects of interventions in Section 2.2. In Section 2.3, we review different causal discovery methods and their underlying assumptions. We focus on two challenges for causality in cognitive neuroscience that are expounded in Section 3: (1) the scarcity of interventional data and (2) the challenge of finding the right variables. In Section 4, we argue that one should seek distributionally robust variable representations and models to tackle these challenges. Most of our arguments in this work are presented in an i.i.d. setting and we briefly discuss the implications for time-dependent data in Section 4.5. We conclude in Section 5 and outline ideas that we regard as promising for future research.

2 Causal models and causal discovery

In contrast to classical probabilistic models, causal models induce not only an observational distribution but also a set of so-called interventional distributions. That is, they predict how a system reacts under interventions. We present an introduction to causal models that is based on pioneer work by [71] and [100]. Our exposition is inspired by [116, Chapter 2], which provides more introductory intuition into causal models viewed as structured sets of interventional distributions. For both simplicity and focus of exposition, we omit a discussion of counterfactual reasoning and other akin causality frameworks such as the potential outcomes formulation of causality [47]. We phrase this article within the framework and terminology of Structural Causal Models (SCMs) [9, 71].

An SCM over variables 𝐙=[Z1,,Zd]\mathbf{Z}=[Z_{1},\dots,Z_{d}]^{\top} consists of

structural equations

that relate each variable ZkZ_{k} to its parents 𝐏𝐀(Zk){Z1,,Zd}{\mathbf{PA}(Z_{k})}\subseteq\{Z_{1},\dots,Z_{d}\} and a noise variable NkN_{k} via a function fkf_{k} such that Zk:=fk(𝐏𝐀(Zk),Nk)Z_{k}:=f_{k}({\mathbf{PA}(Z_{k})},N_{k}), and a

noise distribution 𝐏𝐍\mathbf{P}_{\mathbf{N}}

of the noise variables 𝐍=[N1,,Nd]\mathbf{N}=[N_{1},\dots,N_{d}]^{\top}.

We associate each SCM with a directed causal graph where the nodes correspond to the variables Z1,,ZdZ_{1},\dots,Z_{d} and we draw an edge from ZiZ_{i} to ZjZ_{j} whenever ZiZ_{i} appears on the right hand side of the equation Zj:=fj(𝐏𝐀(Zj),Nj)Z_{j}:=f_{j}({\mathbf{PA}(Z_{j})},N_{j}). That is, if Zi𝐏𝐀(Zj)Z_{i}\in{\mathbf{PA}(Z_{j})} the graph contains the edge ZiZjZ_{i}\to Z_{j}. Here, we assume that this graph is acyclic. The structural equations and noise distributions together induce the observational distribution 𝐏𝐙\mathbf{P_{Z}} of Z1,,ZdZ_{1},\dots,Z_{d} as simultaneous solution to the equations. ([10] formally define SCMs when the graph includes cycles.)

The following is an example of a linear Gaussian SCM:

Z1\displaystyle Z_{1} :=f1(𝐏𝐀(Z1),N1)=Z2+N1\displaystyle:=f_{1}({\mathbf{PA}(Z_{1})},N_{1})=Z_{2}+N_{1}
Z2\displaystyle Z_{2} :=f1(𝐏𝐀(Z2),N2)=N2\displaystyle:=f_{1}({\mathbf{PA}(Z_{2})},N_{2})=N_{2}
Z3\displaystyle Z_{3} :=f3(𝐏𝐀(Z3),N3)=Z1+5Z2+N3\displaystyle:=f_{3}({\mathbf{PA}(Z_{3})},N_{3})=Z_{1}+5\cdot Z_{2}+N_{3}

with mutually independent standard-normal noise variables N1,N2,N3N_{1},N_{2},N_{3}. The corresponding graph is

Z1Z_{1}Z2Z_{2}Z3Z_{3}

and the SCM induces the observational distribution 𝐏𝐙\mathbf{P_{Z}}, which is the multivariate Gaussian distribution

(Z1Z2Z3)𝐏𝐙=𝒩((000),(2171167638)).\begin{pmatrix}Z_{1}\\ Z_{2}\\ Z_{3}\end{pmatrix}\sim\mathbf{P_{Z}}=\mathcal{N}\left(\begin{pmatrix}0\\ 0\\ 0\end{pmatrix},\begin{pmatrix}2&1&7\\ 1&1&6\\ 7&6&38\end{pmatrix}\right)\text{.} (1)

In addition to the observational distribution, an SCM induces interventional distributions. Each intervention denotes a scenario in which we fix a certain subset of the variables to a certain value. For example, the intervention do(Z2:=0,Z3:=5){\operatorname{do}(Z_{2}:=0,Z_{3}:=5)} denotes the scenario where we force Z2Z_{2} and Z3Z_{3} to take on the values 0 and 55, respectively. The interventional distributions are obtained by (a) replacing the structural equations of the intervened upon variables by the new assignment, and (b) considering the distribution induced by the thus obtained new set of structural equations. For example, the distribution under intervention do(Z1:=a){\operatorname{do}(Z_{1}:=a)} for aa\in\mathbb{R}, denoted by 𝐏𝐙do(Z1:=a)\mathbf{P_{Z}}^{\operatorname{do}(Z_{1}:=a)}, is obtained by changing the equation Z1:=f1(𝐏𝐀(Z1),N1)Z_{1}:=f_{1}({\mathbf{PA}(Z_{1})},N_{1}) to Z1:=aZ_{1}:=a. In the above example, we find

𝐏𝐙do(Z1:=a)=𝒩((a0a),(0000150526)),\mathbf{P_{Z}}^{\operatorname{do}(Z_{1}:=a)}=\mathcal{N}\left(\begin{pmatrix}a\\ 0\\ a\end{pmatrix},\begin{pmatrix}0&0&0\\ 0&1&5\\ 0&5&26\end{pmatrix}\right)\text{,}

where X𝒩(a,0)X\sim\mathcal{N}(a,0) if and only if 𝐏(X=a)=1\mathbf{P}(X=a)=1. Analogously, for bb\in\mathbb{R} and intervention on Z2Z_{2} we have

𝐏𝐙do(Z2:=b)=𝒩((bb6b),(101000102)).\mathbf{P_{Z}}^{\operatorname{do}(Z_{2}:=b)}=\mathcal{N}\left(\begin{pmatrix}b\\ b\\ 6\cdot b\end{pmatrix},\begin{pmatrix}1&0&1\\ 0&0&0\\ 1&0&2\end{pmatrix}\right)\text{.}

The distribution of Z1Z_{1} differs between the observational distribution and the interventional distribution, that is, 𝐏Z1𝐏Z1do(Z2:=b)\mathbf{P}_{Z_{1}}\neq\mathbf{P}_{Z_{1}}^{\operatorname{do}(Z_{2}:=b)}. We call a variable XX an (indirect) cause of a variable YY if there exists an intervention on XX under which the distribution of YY is different from its distribution in the observational setting. Thus, Z2Z_{2} is a cause of Z1Z_{1}. The edge Z2Z1Z_{2}\to Z_{1} in the above causal graph reflects this cause-effect relationship. In contrast, Z2Z_{2} remains standard-normally distributed under all interventions do(Z1:=a){\operatorname{do}(Z_{1}:=a)} on Z1Z_{1}. Because the distribution of Z2Z_{2} remains unchanged under any intervention on Z1Z_{1}, Z1Z_{1} is not a cause of Z2Z_{2}.

In general, interventional distributions do not coincide with the corresponding conditional distributions. In our example we have 𝐏𝐙Z1=a𝐏𝐙do(Z1:=a)\mathbf{P}_{\mathbf{Z}\mid Z_{1}=a}\neq\mathbf{P_{Z}}^{\operatorname{do}(Z_{1}:=a)} while 𝐏𝐙Z2=b=𝐏𝐙do(Z2:=b)\mathbf{P}_{\mathbf{Z}\mid Z_{2}=b}=\mathbf{P_{Z}}^{\operatorname{do}(Z_{2}:=b)}. We further have that the conditional distribution 𝐏Z3Z2,Z1\mathbf{P}_{Z_{3}\mid Z_{2},Z_{1}} of Z3Z_{3} given its parents Z1Z_{1} and Z2Z_{2} is invariant under interventions on variables other than Z3Z_{3}. We call a model of Z3Z_{3} based on Z1,Z2Z_{1},Z_{2} invariant (cf. Section 4.1).

We have demonstrated how an SCM induces a set of observational and interventional distributions. The interventional distributions predict observations of the system upon intervening on some of its variables. As such, a causal model holds additional content compared to a common probabilistic model that amounts to one distribution to describe future observations of the same unchanged system. Sometimes we are only interested in modelling certain interventions or cannot perform others as there may be no well-defined corresponding real-world implementation. For example, we cannot intervene on a person’s gender. In these cases it may be helpful to explicitly restrict ourselves to a set of interventions of interest. Furthermore, the choice of an intervention set puts constraints on the granularity of the model [*][cf. Section 3.2 and Rubenstein & Weichwald et al.,][]rubenstein2017causal.

2.1 When are causal models important?

We do not always need causal models to answer our research question. For some scientific questions it suffices to consider probabilistic, that is, observational models. For example, if we wish to develop an algorithm for early diagnosis of Alzheimer’s disease from brain scans, we need to model the conditional distribution of Alzheimer’s disease given brain activity. Since this can be computed from the joint distribution, a probabilistic model suffices. If, however, we wish to obtain an understanding that allows us to optimally prevent progression of Alzheimer’s disease by, for example, cognitive training or brain stimulation, we are in fact interested in a causal understanding of the Alzheimer’s disease and require a causal model.

Distinguishing between these types of questions is important as it informs us about the methods we need to employ in order to answer the question at hand. To elaborate upon this distinction, we now discuss scenarios related to our running examples and the relationship between alcohol consumption and reaction time (cf. Section 1.1). Assume we have access to a powerful OraCle Modelling (OCM) machinery that is unaffected by statistical problems such as model misspecification, multiple-testing, or small sample sizes. By asking ourselves, what queries must be answered by OCM for us to ‘understand’ the cognitive function, the difference between causal and non-causal questions becomes apparent.

Assume, firstly, we ran the reaction task experiment with multiple subjects, fed all observations to our OCM machinery, and have Kim visiting our lab today. Since OCM yields us the exact conditional distribution of reaction times 𝐏YT=t\mathbf{P}_{Y\mid T=t} for Kim having consumed T=tT=t units of alcoholic beer, we may be willing to bet against our colleagues on how Kim will perform in the reaction task experiment they are just about to participate in. No causal model for brain activity is necessary.

Assume, secondly, that we additionally record BOLD responses 𝐗=[X1,,Xd]\mathbf{X}=[X_{1},\dots,X_{d}]^{\top} at certain locations and times during the reaction task experiment. We can query OCM for the distribution of BOLD signals that we are about to record, that is, 𝐏𝐗T=t\mathbf{P}_{\mathbf{X}\mid T=t}, or the distribution of reaction times given we measure Kim’s BOLD responses 𝐗=𝐱\mathbf{X}=\mathbf{x}, that is, 𝐏YT=t,𝐗=𝐱\mathbf{P}_{Y\mid T=t,\mathbf{X}=\mathbf{x}}. As before, we may bet against our colleagues on how Kim’s BOLD signals will look like in the upcoming reaction task experiment or bet on their reaction time once we observed the BOLD activity 𝐗=𝐱\mathbf{X}=\mathbf{x} prior to a reaction cue. Again, no causal model for brain activity is required.

In both of the above situations, we have learned something useful. Given that the data were obtained in an experiment in which alcohol consumption was randomised, we have learned, in the first situation, to predict reaction times after an intervention on alcohol consumption. This may be considered an operational model for alcohol consumption and reaction time. In the second situation, we have learned how the BOLD signal responds to alcohol consumption. Yet, in none of the above situations have we gained understanding of the neuronal underpinnings of the cognitive function and the reaction times. Knowing the conditional distributions 𝐏YT=t\mathbf{P}_{Y\mid T=t} and 𝐏YT=t,𝐗=𝐱\mathbf{P}_{Y\mid T=t,\mathbf{X}=\mathbf{x}} for any tt yields no insight into any of the following questions. Which brain regions maintain fast reaction times? Where in the brain should we release drugs that excite neuronal activity in order to counterbalance the effect of alcohol? How do we need to update our prediction if we learnt that Kim just took a new drug that lowers blood pressure in the prefrontal cortex? To answer such questions, we require causal understanding.

If we had a causal model, say in form of an SCM, we could address the above questions. An SCM offers an explicit way to model the system under manipulations. Therefore, a causal model can help to answer questions about where to release an excitatory drug. It may enable us to predict whether medication that lowers blood pressure in the prefrontal cortex will affect Kim’s reaction time; in general, this is the case if the corresponding variables appear in the structural equations for YY or any of YY’s ancestors.

Instead of identifying conditional distributions, one may formulate the problem as a regression task with the aim to learn the conditional mean functions t𝔼[𝐗|T=t]t\to\mathbb{E}[\mathbf{X}|T=t] and (t,𝐱)𝔼[Y|T=t,𝐗=𝐱](t,\mathbf{x})\to\mathbb{E}[Y|T=t,\mathbf{X}=\mathbf{x}]. These functions are then parameterised in terms of tt or tt and 𝐱\mathbf{x}. We argue in Section 2.2, point (2), that such parameters do not carry a causal meaning and thus do not help to answer the questions above.

Promoted by slogans such as ‘correlation does not imply causation’ careful and associational language is sometimes used in the presentation of cognitive neuroscience studies. We believe, however, that a clear language that states whether a model should be interpreted causally (that is, as an interventional model) or non-causally (that is, as an observational model) is needed. This will help to clarify both the real world processes the model can be used for and the purported scientific claims.

Furthermore, causal models may generalise better than non-causal models. We expect systematic differences between subjects and between different trials or recording days of the same subject. These different situations, or environments, are presumably not arbitrarily different. If they were, we could not hope to gain any scientific insight from such experiments. The apparent question is, which parts of the model we can expect to generalise between environments. It is well-known that causal models capture one such invariance property, which is implicit in the definition of interventions. An intervention on one variable leaves the assignments of the other variables unaffected. Therefore, the conditional distributions of these other variables, given their parents, are also unaffected by the intervention [37, 1]. Thus, causal models may enable us to formulate more clearly which mechanisms we assume to be invariant between subjects. For example, we may assume that the mechanism how alcohol intake affects brain activity differs between subjects, whereas the mechanism from signals in certain brain regions to reaction time is invariant. We discuss the connection between causality and robustness in Section 4.

2.2 Equivalences of models

Causal models entail strictly more information than observational models. We now introduce the notion of equivalence of models [71, 76, 10]. This notion allows us to discuss the falsifiability of causal models, which is important when assessing candidate models and their ability to capture cause-effect relationships that govern a cognitive process under investigation.

We call two models observationally equivalent if they induce the same observational distribution. Two models are said to be interventionally equivalent if they induce the same observational and interventional distributions. As discussed above, for some interventions there may not be a well-defined corresponding experiment in the real world. We therefore also consider interventional equivalence with respect to a restricted set of interventions.

One reason why learning causal models from observational data is difficult is the existence of models that are observationally but not interventionally equivalent. Such models agree in their predictions about the observed system yet disagree in their predictions about the effects of certain interventions. We continue the example from Section 2 and consider the following two SCMs:

Z1\displaystyle Z_{1} :=Z2+N1\displaystyle:=Z_{2}+N_{1} Z1\displaystyle{Z}_{1} :=2N1\displaystyle:=\sqrt{2}\cdot N_{1}
Z2\displaystyle Z_{2} :=N2\displaystyle:=N_{2} Z2\displaystyle{Z}_{2} :=1/2Z1+1/2N2\displaystyle:=\nicefrac{{1}}{{2}}\cdot{Z}_{1}+\nicefrac{{1}}{{\sqrt{2}}}\cdot N_{2}
Z3\displaystyle Z_{3} :=Z1+5Z2+N3\displaystyle:=Z_{1}+5\cdot Z_{2}+N_{3} Z3\displaystyle{Z}_{3} :=Z1+5Z2+N3\displaystyle:={Z}_{1}+5\cdot{Z}_{2}+N_{3}

where in both cases N1,N2,N3N_{1},N_{2},N_{3} are mutually independent standard-normal noise variables. The two SCMs are observationally equivalent as they induce the same observational distribution, the one shown in Equation (1). The models are not interventionally equivalent, however, since 𝐏Z2do(Z1:=3)=𝒩(0,1)\mathbf{P}_{Z_{2}}^{\operatorname{do}(Z_{1}:=3)}=\mathcal{N}(0,1) and 𝐏Z2do(Z1:=3)=𝒩(3/2,1/2)\mathbf{P}_{Z_{2}}^{\operatorname{do}(Z_{1}:=3)}=\mathcal{N}(\nicefrac{{3}}{{2}},\nicefrac{{1}}{{2}}) for the left and right model, respectively. The two models can be told apart when interventions on Z1Z_{1} or Z2Z_{2} are considered. They are interventionally equivalent with respect to interventions on Z3Z_{3}.

The existence of observationally equivalent models that are not interventionally equivalent has several implications. (1) Without assumptions, it is impossible to learn causal structure from observational data. This is not exclusive to causal inference from data and an analogous statement holds true for regression [36]. The regression problem is solvable only under certain simplicity assumptions, for example, on the smoothness of the regression function, which have been proven useful in real world applications. Similarly, there are several assumptions that can be exploited for causal discovery. We discuss some of these assumptions in Section 2.3. (2) As a consequence, without further restrictive assumptions on the data generating process, the estimated parameters do not carry any causal meaning. For example, given any finite sample from the observational distribution, both of the above SCMs yield exactly the same likelihood. Therefore, the above structures cannot be told apart by a method that employs the maximum likelihood estimation principle. Instead, which SCM and thus which parameters are selected in such a situation may depend on starting values, optimisation technique, or numerical precision. (3) Assume that we are given a probabilistic (observational) model of a data generating process. To falsify it, we may apply a goodness-of-fit test based on an observational sample from that process. An interventional model cannot be falsified based on observational data alone and one has to also take into account the outcome of interventional experiments. This requires that we are in agreement about how to perform the intervention in practice (see also Section 3.2). Interventional data may be crucial in particular for rejecting some of the observationally equivalent models (cf. the example above). The scarcity of interventional data therefore poses a challenge for causality in cognitive neuroscience (cf. Section 3.1).

2.3 Causal discovery

The task of learning a causal model from observational (or a combination of observational and interventional) data is commonly referred to as causal discovery or causal structure learning. We have argued in the preceding section that causal discovery from purely observational data is impossible without any additional assumptions or background knowledge. In this section, we discuss several assumptions that render (parts of) the causal structure identifiable from the observational distribution. In short, assumptions concern how causal links manifest in observable statistical dependences, functional forms of the mechanisms, certain invariances under interventions, or the order of time. We briefly outline how these assumptions can be exploited in algorithms. Depending on the application at hand, one may be interested in learning the full causal structure as represented by its graph or in identifying a local structure such as the causes of a target variable YY. The methods described below cover either of the two cases. We keep the description brief focussing on the main ideas and intuition, while more details can be found in the respective references.

Randomisation.

The often called ‘gold standard’ to establishing whether TT causes YY is to introduce controlled perturbations, that is, targeted interventions, to a system. Without randomisation, a dependence between TT and YY could stem from a confounder between TT and YY or from a causal link from YY to TT. If TT is randomised it is no further governed by the outcome of any other variable or mechanism. Instead, it only depends on the outcome of a randomisation experiment, such as the roll of a die. If we observe that under the randomisation, YY depends on TT, say the higher TT the higher YY, then there must be a (possibly indirect) causal influence from TT to YY. In our running examples, this allows us to conclude that the amount of alcoholic beer consumed causes reaction times (cf. Section 1.1). When falsifying interventional models, it suffices to consider randomised experiments as interventions [[]Proposition 6.48]peters2017elements. In practice, however, performing randomised experiments is often infeasible due to cost or ethical concerns, or impossible as, for example, we cannot randomise gender nor fully control neuronal activity in the temporal lobe. While it is sometimes argued that the experiment conducted by James Lind in 1747 to identify a treatment for scurvy is among the first randomised controlled trials, the mathematical theory and methodology was popularised by Ronald A. Fisher in the early 20th century [22].

Constraint-based methods.

Constraint-based methods rely on two assumptions that connect properties of the causal graph with conditional independence statements in the induced distribution. The essence of the first assumption is sometimes described as Reichenbach’s common cause principle [85]: If XX and YY are dependent, then there must be some cause-effect structure that explains the observed dependence, that is, either XX causes YY, or YY causes XX, or another unobserved variable HH causes both XX and YY, or some combination of the aforementioned. This principle is formalised by the Markov condition [[, see for example]]Lauritzen1996. This assumption is considered to be mild. Any distribution induced by an acyclic SCM satisfies the Markov condition with respect to the corresponding graph [52, 71]. The second assumption (often referred to as faithfulness), states that any (conditional) independence between random variables is implied by the graph structure [100]. For example, if two variables are independent, then neither does cause the other nor do they share a common cause. Both assumptions together establish a one-to-one correspondence between conditional independences in the distribution and graphical separation properties between the corresponding nodes.

The back-bone of the constraint-based causal discovery algorithms such as the PC algorithm is to test for marginal and conditional (in)dependences in observed data and to find all graphs that encode the same list of separation statements [100, 71]. This allows us to infer a so-called Markov equivalence class of graphs: all of its members encode the same set of conditional independences. It has been shown that two directed acyclic graphs (assuming that all nodes are observed) are Markov equivalent if and only if they have the same skeleton and v-structures \to{\circ}\leftarrow [113]. Allowing for hidden variables, as done by the FCI algorithm, for example, enlarges the class of equivalent graphs and the output is usually less informative [100].

The following example further illustrates the idea of a constraint-based search. For simplicity, we assume a linear Gaussian setting, so that (conditional) independence coincides with vanishing (partial) correlation. Say we observe XX, YY, and ZZ. Assume that the partial correlation between XX and ZZ given YY vanishes while none of the other correlations and partial correlations vanish. Under the Markov and faithfulness assumptions there are multiple causal structures that are compatible with those constraints, such as XYZX\to Y\to Z, XYZX\leftarrow Y\leftarrow Z, XYZX\leftarrow Y\to Z, or

HHXXYYZZ

,   or   HHXXYYZZ,

where HH is unobserved. Still, the correlation pattern rules out certain other causal structures. For example, neither XYZX\to Y\leftarrow Z nor XHYZX\leftarrow H\to Y\leftarrow Z can be the correct graph structure since either case would imply that XX and ZZ are uncorrelated (and XZYX\mbox{${}\perp\mkern-11.0mu\perp{}$}Z\mid Y is not satisfied).

Variants of the above setting were considered in neuroimaging where a randomised experimental stimulus or time-ordering was used to further disambiguate between the remaining possible structures [35, 117, 118, 63]. Constraint-based causal inference methodology also clarifies the interpretation of encoding and decoding analyses in neuroimaging and has informed a refined understanding of the neural dynamics of probabilistic reward prediction and an improved functional atlas [119, 3, 112].

Direct applications of this approach in cognitive neuroscience are difficult, not only due to the key challenges discussed in Section 3, but also due to indirect and spatially smeared neuroimaging measurements that effectively spoil conditional independences. In the linear setting, there are recent advances that explicitly tackle the problem of inferring the causal structure between latent variables, say the neuronal entities, based on observations of recorded variables [96]. Further practical challenges include the difficulty of testing for non-parametric conditional independence [94] and near-faithfulness violations [109].

Score-based methods.

Instead of directly exploiting the (conditional) independences to inform our inference about the causal graph structure, score-based methods assess different graph structures by their ability to fit observed data [[, see for example]]Chickering2002. This approach is motivated by the idea that graph structures that encode the wrong (conditional) independences will also result in bad model fit. Assuming a parametric model class, we can evaluate the log-likelihood of the data and score different candidate graph structures by the Bayesian Information Criterion, for example. The number of possible graph structures to search over grows super-exponentially. That combinatorial difficulty can be dealt with by applying greedy search procedures that usually, however, do not come with finite sample guarantees. Alternatively, [127] exploit an algebraic characterisation of graph structures to maximise a score over acyclic graphs by solving a continuous optimisation problem. The score-based approach relies on correctly specifying the model class. Furthermore, in the presence of hidden variables, the search space grows even larger and model scoring is complicated by the need to marginalise over those hidden variables [48].

Restricted structural causal models.

Another possibility is to restrict the class of functions in the structural assignments and the noise distributions. Linear non-Gaussian acyclic models [95], for example, assume that the structural assignments are linear and the noise distributions are non-Gaussian. As for independent component analysis, identifiability of the causal graph follows from the Darmois-Skitovich theorem [24, 97]. Similar results hold for nonlinear models with additive noise [44, 126, 77, 12] or linear Gaussian models when the error variances of the different variables are assumed to be equal [73]. The additive noise assumption is a powerful, yet restrictive, assumption that may be violated in practical applications.

Dynamic causal modelling (DCM).

We may have prior beliefs about the existence and direction of some of the edges. Incorporating these by careful specification of the priors is an explicit modelling step in DCM [110]. Given such a prior, we may prefer one model over the other among the two observationally equivalent models presented in Section 2.2, for example. Since the method’s outcome relies on this prior information, any disagreement on the validity of that prior information necessarily yields a discourse about the method’s outcome [56]. Further, a simulation study raised concerns regarding the validity of the model selection procedure in DCM [32, 56, 33, 11, 57].

Granger causality.

Granger causality is among the most popular approaches for the analysis of connectivity between time-evolving processes. It exploits the existence of time and the fact that causes precede their effects. Together with its non-linear extensions it has been considered for the analysis of neuroimaging data with applications to electro-encephalography (EEG) and fMRI data [61, 62, 103, 104]. The idea is sometimes wrongly described as follows: If including the past of YtY_{t} improves our prediction of XtX_{t} compared to a prediction that is only based on the past of XtX_{t} alone, then YY Granger-causes XX. [34] himself put forward a more careful definition that includes a reference to all the information in the universe: If the prediction of XtX_{t} based on all the information in the universe up to time tt is better than the prediction where we use all the information in the universe up to time tt apart from the past of YtY_{t}, then YY Granger-causes XX. In practice, we may instead resort to a multivariate formulation of Granger causality. If all relevant variables are observed (often referred to as causal sufficiency), there is a close correspondence between Granger causality and the constraint-based approach [76, Chapter 10.3.3]. Observing all relevant variables, however, is a strong assumption which is most likely violated for data sets in cognitive neuroscience. While Granger causality may be combined with a goodness-of-fit test to at least partially detect the existence of confounders [75], it is commonly applied as a computationally efficient black box approach that always outputs a result. In the presence of instantaneous effects (for example, due to undersampling) or hidden variables, these results may be erroneous [[, see, for example,]]sanchez2019estimating.

Inferring causes of a target variable.

We now consider a problem that is arguably simpler than inferring the full causal graph: identifying the causes of some target variable of interest. As outlined in the running examples in Section 1.1, we assume that we have observations of the variables T,Y,X1,,XdT,Y,X_{1},\ldots,X_{d}, where YY denotes the target variable. Assume that there is an unknown structural causal model that includes the variables T,Y,X1,,XdT,Y,X_{1},\ldots,X_{d} and that describes the data generating process well. To identify the variables among X1,,XdX_{1},\ldots,X_{d} that cause YY, it does not suffice to regress YY on X1,,XdX_{1},\ldots,X_{d}. The following example of an SCM shows that a good predictive model for YY is not necessarily a good interventional model for YY. Consider

X1\displaystyle X_{1} :=N1\displaystyle:=N_{1}
Y\displaystyle Y :=X1+NY\displaystyle:=X_{1}+N_{Y}
X2\displaystyle X_{2} :=10Y+N2\displaystyle:=10\cdot Y+N_{2}
X1X_{1}X2X_{2}YY

where N1,N2,NYN_{1},N_{2},N_{Y} are mutually independent standard-normal noise variables. X2X_{2} is a good predictor for YY, but X2X_{2} does not have any causal influence on YY: the distribution of YY is unchanged upon interventions on X2X_{2}.

Recently, causal discovery methods have been proposed that aim to infer the causal parents of YY if we are given data from different environments, that is, from different experimental conditions, repetitions, or different subjects. These methods exploit a distributional robustness property of causal models and are described in Section 4.

Cognitive function versus brain activity as the target variable.

When we are interested in inferring direct causes of a target variable YY, it can be useful to include background knowledge. Consider our Running Example A (cf. Section 1.1 and Figure 1(a)) with reaction time as the target variable and assume we are interested in inferring which of the variables measuring neuronal activity are causal for the reaction time YY. We have argued in the preceding paragraph that if a variable XjX_{j} is predictive of YY, it does not necessarily have to be causal for YY. Assuming, however, that we can exclude that the cognitive function ‘reaction time’ causes brain activity (for example, because of time ordering), we obtain the following simplification: every XjX_{j} that is predictive of YY, must be an indirect or direct cause of YY, confounded with YY, or a combination of both. This is different if our target variable is a neuronal entity as in Running Example B (cf. Figure 1(b)). Here, predictive variables can be either ancestors of YY, confounded with YY, descendants of YY, or some combination of the aforementioned (these statements follow from the Markov condition).

3 Two challenges for causality in cognitive neuroscience

Performing causal inference on measurements of neuronal activity comes with several challenges, many of which have been discussed in the literature (cf. Section 1.2). In the following two subsections we explicate two challenges that we think deserve special attention. In Section 4, we elaborate on how distributional robustness across environments, such as different recording sessions or subjects, can serve as a guiding principle for tackling those challenges.

3.1 Challenge 1: The scarcity of targeted interventional data

In Section 2.2 we discussed that different causal models may induce the same observational distribution while they make different predictions about the effects of interventions. That is, observationally equivalent models need not be interventionally equivalent. This implies that some models can only be refuted when we observe the system under interventions which perturb some specific variables in our model. In contrast to broad perturbations of the system, we call targeted interventions those for which the intervention target is known and for which we can list the intervened-upon variables in our model, say “X1,X3,X8X_{1},X_{3},X_{8} have been intervened upon.” Even if some targeted interventions are available, there may still be multiple models that are compatible with all observations obtained under those available interventions. In the worst case, a sequence of up to dd targeted interventional experiments may be required to distinguish between the possible causal structures over dd observables X1,,XdX_{1},\dots,X_{d} when the existence of unobserved variables cannot be excluded while assuming Markovianity, faithfulness, and acyclicity [28]. In general, the more interventional scenarios are available to us, the more causal models we can falsify and the further we can narrow down the set of causal models compatible with the data.

Therefore, the scarcity of targeted interventional data is a barrier to causal inference in cognitive neuroscience. Our ability to intervene on neural entities such as the BOLD level or oscillatory bandpower in a brain region is limited and so is our ability to either identify the right causal model from interventional data or to test causal hypotheses that are made in the literature. One promising avenue are non-invasive brain stimulation techniques such as transcranial magnetic or direct/alternating current stimulation which modulate neural activity by creating a field inside the brain [69, 40, 7, 50]. Since the stimulation acts broadly and its neurophysiological effects are not yet fully understood, transcranial stimulation cannot be understood as targeted intervention on some specific neuronal entity in our causal model [2, 114]. The inter-individual variability in response to stimulation further impedes its direct use for probing causal pathways between brain regions [58]. [6] review the obstacles to inferring causality from non-invasive brain stimulation studies and provide guidelines to attenuate the aforementioned. Invasive stimulation techniques, such as deep brain stimulation relying on electrode implants [64], may enable temporally and spatially more fine-grained perturbations of neural entities. [26] exemplify how to revise causal structures inferred form observational neuroimaging data on a larger cohort through direct stimulation of specific brain regions and concurrent fMRI on a smaller cohort of neurosurgical epilepsy patients. In non-human primates, concurrent optogenetic stimulation with whole-brain fMRI had been used to map the wiring of the medial prefrontal cortex [54, 53]. Yet, there are ethical barriers to large-scale invasive brain stimulation studies and it may not be exactly clear how an invasive stimulation corresponds to an intervention on, say, the BOLD response measured in some voxels. We thus believe that targeted interventional data will remain a scarcity due to physical and ethical limits to non-invasive and invasive brain stimulation.

Consider the following variant of our Running Example B (cf. Section 1.1). Assume that (a) the consumption of alcoholic beer TT slows neuronal activity in the brain regions X1X_{1}, X2X_{2}, and YY, (b) X2X_{2} is a cause of X1X_{1}, and (c) X2X_{2} is a cause of YY. Here, (a) could have been established by randomising TT, whereas (b) and (c) may be background knowledge. Nothing is known, however, about the causal relationship between X1X_{1} and YY (apart from the confounding effect of X2X_{2}). The following graph summarises these causal relationships between the variables:

[Uncaptioned image]
TTHHX1X_{1}X2X_{2}YY?

Assume we establish on observational data that there is a dependence between X1X_{1} and YY and that we cannot render these variables conditionally independent by conditioning on any combination of the remaining observable variables TT and X2X_{2}. Employing the widely accepted Markov condition, we can conclude that either X1YX_{1}\to Y, X1YX_{1}\leftarrow Y, X1HYX_{1}\leftarrow H\to Y for some unobserved variable HH, or some combination of the aforementioned settings. Without any further assumptions, however, these models are observationally equivalent. That is, we cannot refute any of the above possibilities based on observational data alone. Even randomising TT does not help: The above models are interventionally equivalent with respect to interventions on TT. We could apply one of the causal discovery methods described in Section 2.3. All of these methods, however, employ further assumptions on the data generating process that go beyond the Markov condition. We may deem some of those assumptions implausible given prior knowledge about the system. Yet, in the absence of targeted interventions on X1X_{1}, X2X_{2} or YY, we can neither falsify candidate models obtained by such methods nor can we test all of the underlying assumptions. In Section 4.2, we illustrate how we may benefit from heterogeneity in the data, that is, from interventional data where the intervention target is unknown.

3.2 Challenge 2: Finding the right variables

Causal discovery often starts by considering observations of some variables Z1,,ZdZ_{1},\dots,Z_{d} among which we wish to infer cause-effect relationships, thereby implicitly assuming that those variables are defined or constructed in a way that they can meaningfully be interpreted as causal entities in our model. This, however, is not necessarily the case in neuroscience. Without knowing how higher-level causal concepts emerge from lower levels, for example, it is hard to imagine how to make sense and use of a causal model of the 8686 billion neurons in a human brain [39]. One may hypothesise that a model of averaged neuronal activity in distinct functional brain regions may be pragmatically useful to reason about the effect of different treatments and to understand the brain. For such an approach we need to find the right transformation of the high-dimensional observed variables to obtain the right variables for a causal explanation of the system.

The problem of relating causal models with different granularity and finding the right choice of variable transformations that enable causal reasoning has received attention in the causality literature also outside of neuroscience applications. [29] fleshes out an instructive two-variable example that demonstrates that the choice of variables for causal modelling may be underdetermined even if interventions were available. For a wrong choice of variables our ability to causally reason about a system breaks. An example of this is the historic debate about whether a high cholesterol diet was beneficial or harmful with respect to heart disease. It can be partially explained by an ambiguity of how exactly total cholesterol is manipulated. Today, we know that low-density lipoproteins and high-density lipoproteins have opposing effects on heart disease risk. Merging these variables together to total cholesterol does not yield a variable with a well-defined intervention: Referring to an intervention on total cholesterol does not specify what part of the intervention is due to a change in low-density lipoproteins (LDL) versus high-density lipoproteins (HDL). As such, only including total cholesterol instead of LDL and HDL may therefore be regarded as a too coarse-grained variable representation that breaks a model’s causal semantics, that is, the ability to map every intervention to a well-defined interventional distribution [101, 102, 107].

Yet, we may sometimes prefer to transform micro variables into macro variables. This can result in a concise summary of the causal information that abstracts away detail, is easier to communicate and operationalise, and more effectively represents the information necessary for a certain task [42, 41, 116]; for example, a causal model over 8686 billion neurons may be unwieldy for a brain surgeon aiming to identify and remove malignant brain tissue guided by the cognitive impairments observed in a patient. [88] formalise a notion of exact transformations that ensures causally consistent reasoning between two causal models where the variables in one model are transformations of the variables in the other. Roughly speaking, two models are considered causally consistent if the following two ways to reason about how the distribution of the macro-variables changes upon a macro-level intervention agree with one another: (a) find an intervention on the micro-variables that corresponds to the considered macro-level intervention, and consider the macro-level distribution implied by the micro-level intervention, and (b) obtain the interventional distribution directly within the macro-level structural causal model sidestepping any need to refer to the micro-level. If the two resulting distributions agree with one another for all (compositions of) interventions, then the two models are said to be causally consistent and we can view the macro-level as an exact transformation of the micro-level causal model that preserves its causal semantics. A formal exposition of the framework and its technical subtleties can be found in the aforementioned work. Here, we revisit a variant of the cholesterol example for an illustration of what it entails for two causal models to be causally consistent and illustrate a failure mode: Consider variables LL (LDL), HH (HDL), and DD (disease), where D:=HL+NDD:=H-L+N_{D} for L,H,NdL,H,N_{d} mutually independent random variables. Then a model based on the transformed variables T=L+HT=L+H and DDD\equiv D is in general not causally consistent with the original model: For (l1,h1)(l2,h2)(l_{1},h_{1})\neq(l_{2},h_{2}) with l1+h1=l2+h2l_{1}+h_{1}=l_{2}+h_{2} the interventional distributions induced by the micro-level model corresponding to setting L:=l1L:=l_{1} and H:=h1H:=h_{1} or alternatively L:=l2L:=l_{2} and H:=h2H:=h_{2} do in general not coincide due to the differing effects of LL and HH on DD. Both interventions correspond to the same level of TT and the intervention setting T:=tT:=t with t=l1+h1=l2+h2t=l_{1}+h_{1}=l_{2}+h_{2} in the macro-level model. Thus, the distributions obtained from reasoning (a) and (b) above do not coincide. If, on the other hand, we had D~:=H+L+ND\widetilde{D}:=H+L+N_{D}, then we could indeed use a macro-level model where we consider T=H+LT=H+L to reason about the distribution of D~\widetilde{D} under the intervention do(T:=t){\operatorname{do}(T:=t)} without running into conflict with the interventional distributions implied by all corresponding interventions in the micro-level model. This example can analogously be considered in the context of our running examples (cf. Section 1.1): Instead of LDL, HDL, and disease one could alternatively think of some neuronal activity (LL) that delays motor response, some neuronal activity (HH) that increases attention levels, and the detected reaction time (DD) assessed by subjects performing a button press; the scenario then translates into how causal reasoning about the cause of slowed reaction times is hampered once we give up on considering HH and LL as two separate neural entities and instead try to reason about the average activity TT. [49] observe similar problems for causal reasoning when aggregating variables and show that the observational and interventional stationary distributions of a bivariate autoregressive processes cannot in general be described by a two-variable causal model. A recent line of research focuses on developing a notion of approximate transformations of causal models [5, 4]. While there exist first approaches to learn discrete causal macro-variables from data [16, 15], we are unaware of any method that is generally applicable and learns causal variables from complex high-dimensional data.

In cognitive neuroscience, we commonly treat large-scale brain networks or brain systems as causal entities and then proceed to infer interactions between those [123, 82]. [99] demonstrate that this should be done with caution: Network identification is strongly susceptible to slightly wrong or different definitions of the regions of interest (ROIs) or the so-called atlas. Analyses based on Granger causality depend on the level of spatial aggregation and were shown to reflect the intra-areal properties instead of the interactions among brain regions if an ill-suited aggregation level is considered [17]. Currently, there does not seem to be consensus as to which macroscopic entities and brain networks are the right ones to (causally) reason about cognitive processes [108]. Furthermore, the observed variables themselves are already aggregates: A single fMRI voxel or the local field potential at some cortical location reflects the activity of thousands of neurons [55, 30]; EEG recordings are commonly considered a linear superposition of cortical electromagnetic activity which has spurred the development of blind source separation algorithms that try to invert this linear transformation to recover the underlying cortical variables [70].

4 Causality and leveraging robustness

4.1 Robustness of causal models

The concept of causality is linked to invariant models and distributional robustness. Consider again the setting with a target variable YY and covariates X1,,XdX_{1},\ldots,X_{d}, as described in the running examples in Section 1.1. Suppose that the system is observed in different environments. Suppose further that the generating process can be described by an SCM, that 𝐏𝐀(Y){X1,,Xd}{\mathbf{PA}(Y)}\subseteq\{X_{1},\ldots,X_{d}\} are the causal parents of YY, and that the different environments correspond to different interventions on some of the covariates, while we neither (need to) know the interventions’ targets nor its precise form. In our reaction time example, the two environments may represent two subjects (say, a left-handed subject right after having dinner and a trained race car driver just before a race) that differ in the mechanisms for X1X_{1}, X3X_{3}, and X7X_{7}. Then the joint distribution over Y,X1,,XdY,X_{1},\ldots,X_{d} may be different between the environments and also the marginal distributions may vary. Yet, if the interventions do not act directly on YY the causal model is invariant in the following sense: the conditional distribution of Y|𝐗𝐏𝐀(Y)Y\,|\,\mathbf{X}_{{\mathbf{PA}(Y)}} is the same in all environments. In the reaction time examples this could translate to the neuronal causes that facilitate fast (versus slow) reaction times to be the same across subjects. This invariance can be formulated in different ways. For example, we have for all kk and \ell, where kk and \ell denote the indices of two environments, and for almost all xx

(Yk|𝐗𝐏𝐀(Y)k=x)=(Y|𝐗𝐏𝐀(Y)=x)in distribution.(Y^{k}\,|\,\mathbf{X}^{k}_{{\mathbf{PA}(Y)}}=x)=(Y^{\ell}\,|\,\mathbf{X}^{\ell}_{{\mathbf{PA}(Y)}}=x)\qquad\text{in distribution}. (2)

Equivalently,

EY|𝐗𝐏𝐀(Y),E\mbox{${}\perp\mkern-11.0mu\perp{}$}Y\,|\,\mathbf{X}_{{\mathbf{PA}(Y)}}, (3)

where the variable EE represents the environment. In practice, we often work with model classes such as linear or logistic regression for modelling the conditional distribution Y|𝐗𝐏𝐀(Y)Y\,|\,\mathbf{X}_{{\mathbf{PA}(Y)}}. For such model classes, the above statements simplify. In case of linear models, for example, Equations (2) and (3) translate to regression coefficients and error variances being equal across different environments.

For an example, consider a system that, for environment E=1E=1, is governed by the following structural assignments

SCM for E=1E=1:

X1\displaystyle X_{1} :=N1\displaystyle:=N_{1}
X2\displaystyle X_{2} :=1X1+N2\displaystyle:=1\cdot X_{1}+N_{2}
X3\displaystyle X_{3} :=N3\displaystyle:=N_{3}
Y\displaystyle Y :=X1+X2+X3+NY\displaystyle:=X_{1}+X_{2}+X_{3}+N_{Y}
X4\displaystyle X_{4} :=Y+N2\displaystyle:=Y+N_{2}
X1X_{1}X4X_{4}X3X_{3}X2X_{2}YY
[Uncaptioned image]

with N1,N2,N3,N4,NYN_{1},N_{2},N_{3},N_{4},N_{Y} mutually independent and standard-normal, and where environment E=1E=-1 corresponds to an intervention changing the weight of X1X_{1} in the assignment for X2X_{2} to 1-1. Here, for example, {X1,X2,X3}\{X_{1},X_{2},X_{3}\} and {X1,X2,X4}\{X_{1},X_{2},X_{4}\} are so-called invariant sets: the conditionals Y|X1,X2,X3Y|X_{1},X_{2},X_{3} and Y|X1,X2,X4Y|X_{1},X_{2},X_{4} are the same in both environments. The invariant models Y|X1,X2,X3Y|X_{1},X_{2},X_{3} and Y|X1,X2,X4Y|X_{1},X_{2},X_{4} generalise to a new environment E=2E=-2, which changes the same weight to 2-2, in that they would still predict well. Note that Y|X1,X2,X4Y|X_{1},X_{2},X_{4} is a non-causal model. The lack of invariance of Y|X2Y|X_{2} is illustrated by the different regression lines in the scatter plot on the right.

The validity of (2) and (3) follows from the fact that the interventions do not act on YY directly and can be proved using the equivalence of Markov conditions [[]Section 6.6]Lauritzen1996,Peters2016jrssb. Here, we try to argue that it also makes sense intuitively. Suppose that someone proposes to have found a complete causal model for a target variable YY, using certain covariates 𝐗S\mathbf{X}_{S} (for YY, we may again think of the reaction time in Example A). Suppose that fitting that model for different subjects yields significantly different model fits – maybe even with different signs for the causal effects from variables in 𝐗S\mathbf{X}_{S} to YY such that EY𝐗SE\mbox{${}\perp\mkern-11.0mu\perp{}$}Y\mid\mathbf{X}_{S} is violated. In this case, we would become sceptical about whether the proposed model is indeed a complete causal model. Instead, we might suspect that the model is missing an important variable describing how reaction time depends on brain activity.

In practice, environments can represent different sources of heterogeneity. In a cognitive neuroscience setting, environments may be thought of as different subjects who react differently, yet not arbitrarily so (cf. Section 2.1), to varying levels of alcohol consumption. Likewise, different experiments that are thought to involve the same cognitive processes may be thought of as environments; for example, the relationship ‘neuronal activity \to reaction time’ (cf. Example A, Section 1.1) may be expected to translate from an experiment that compares reaction times after consumption of alcoholic versus non-alcoholic beers to another experiment where subjects are exposed to Burgundy wine versus grape juice. The key assumption is that the environments do not alter the mechanism of YY—that is, fY(𝐏𝐀(Y),NY)f_{Y}({\mathbf{PA}(Y)},N_{Y})—directly or, more formally, there are no interventions on YY. To test whether a set of covariates is invariant, as described in (2) and (3), no causal background knowledge is required.

The above invariance principle is also known as ‘modularity’ or ‘autonomy’. It has been discussed not only in the field of econometrics [37, 1, 43], but also in philosophy of science. [121] discusses how the invariance idea rejects that ‘either a generalisation is a law or else is purely accidental’. In our notion, the criteria (2) and (3) depend on the environments EE. In particular, a model may be invariant with respect to some changes, but not with respect to others. In this sense, robustness and invariance should always be thought with respect to a certain set of changes. [121] introduces the possibility to talk about various degrees of invariance, beyond the mere existence or absence of invariance, while acknowledging that mechanisms that are sensitive even to mild changes in the background conditions are usually considered as not scientifically interesting. [13] analyses the relationship between invariant and causal relations using linear deterministic systems and draws conclusions analogous to the ones discussed above. In the context of the famous Lucas critique [59], it is debated to which extent invariance can be used for predicting the effect of changes in economic policy [14]: Economy consists of many individual players who are capable of adapting their behaviour to a change in policy. In cognitive neuroscience, we believe that the situation is different. Cognitive mechanisms do change and adapt, but not necessarily arbitrarily quickly. Some cognitive mechanism of an individual at the same day can be assumed to be invariant with respect to changes in the visual input, say. Depending on the precise setup, however, we may expect moderate changes of the mechanisms, say, for example, the development of cognitive function in children or learning effects. In other settings, where mechanisms may be subject to arbitrary large changes, scientific insight seems impossible (see Section 2.1).

Recently, the principle of invariance has also received increasing attention in the statistics and machine learning community [93, 74, 60]. It can also be applied to models that do not have the form of an SCM. Examples include dynamical models that are governed by differential equations [78].

4.2 Distributional robustness and scarcity of interventional data

The idea of distributional robustness across changing background conditions may help us to falsify causal hypotheses, even when interventional data is difficult to obtain, and in this sense may guide us towards models that are closer to the causal ground truth. For this, suppose that the data are obtained in different environments and that we expect a causal model for YY to yield robust performance across these environments (see Section 4.1). Even if we lack targeted interventional data in cognitive neuroscience and thus cannot test a causal hypothesis directly, we can test the above implication. We can test the invariance, for example, using conditional independence tests or specialised tests for linear models [19]. We can, as a surrogate, hold out one environment, train our model on the remaining environments, and evaluate how well that model performs on the held-out data (cf. Figure 2); the reasoning is that a non-invariant model may not exhibit robust predictive performance and instead yield a bad predictive performance for one or more of the folds. If a model fails the above then either (1) we included the wrong variables, (2) we have not observed important variables, or (3) the environment directly affects YY. Tackling (1), we can try to refine our model and search for different variable representations and variable sets that render our model invariant and robust in the post-analysis. In general, there is no way to recover from (2) and (3), however.

While a model that is not invariant across environments cannot be the complete causal model (assuming the environments do not act directly on the target variable), it may still have non-trivial prediction performance and predict better than a simple baseline method in a new, unseen environment. The usefulness of a model is questionable, however, if its predictive performance on held-out environments is not significantly better than a simple baseline. Conversely, if our model shows robust performance on the held-out data and is invariant across environments, it has the potential of being a causal model (while it need not be; see Section 4.1 for an example). Furthermore, a model that satisfies the invariance property is interesting in itself as it may enable predictions in new, unseen environments. For this line of argument, it does not suffice to employ a cross-validation scheme that ignores the environment structure and only assesses predictability of the model on data pooled across environments. Instead, we need to respect the environment structure and assess the distributional robustness of the model across these environments.

11n1n_{1}22n2n_{2}\vdotsnkn_{k}KKnKn_{K}Available dataKK environmentsEach environment kkholds nkn_{k} observations22n2n_{2}\ldotsnkn_{k}KKnKn_{K}Fit model for Y|𝐗Y|\mathbf{X}targetcovariates…repeat for KK splits…11n1n_{1}Predictive onheld-out fold?Leave-one-environment-out cross-validationInvariance-test across environments11n1n_{1}22n2n_{2}\ldotsnkn_{k}KKnKn_{K}Y1|𝐗1Y^{1}|\mathbf{X}^{1}Y2|𝐗2Y^{2}|\mathbf{X}^{2}Yk|𝐗kY^{k}|\mathbf{X}^{k}YK|𝐗KY^{K}|\mathbf{X}^{K}Invariant across folds?
Figure 2: Illustration of a cross-validation scheme across KK environments (cf. Section 4.2). Environments can correspond to recordings on different days, during different tasks, or on different subjects, and define how the data is split into folds for the cross-validation scheme. We propose to assess a model by (a) leave-one-environment-out cross-validation testing for robust predictive performance on the held-out fold and (b) an invariance-test across environments assessing whether the model is invariant across folds. The cross-validation scheme (a) is repeated KK times, so that each environment acts as a held-out fold once. Models whose predictive performance does not generalise to held-out data or that are not invariant across environments can be refuted as non-causal. For linear models, for example, invariance across environments can be assessed by evaluating to which extent regression coefficients and error variances differ across folds (cf. Section 4.2).

For an illustration of the interplay between invariance and predictive performance, consider a scenario in which X1YHX2X_{1}\to Y\to H\to X_{2}, where HH is unobserved. Here, we regard different subjects as different environments and suppose that (unknown to us) the environment acts on HH: One may think of a variable EE pointing into HH. Let us assume that our study contains two subjects, one that we use for training and another one that we use as held-out fold. We compare a model of the form Y^=f(𝐗𝐏𝐀(Y))=f(X1)\widehat{Y}=f\left(\mathbf{X}_{{\mathbf{PA}(Y)}}\right)=f(X_{1}) with a model of the form Y~=g(𝐗)=g(X1,X2)\widetilde{Y}=g\left(\mathbf{X}\right)=g(X_{1},X_{2}). On a single subject, the latter model including all observed variables has more predictive power than the former model that only includes the causes of YY. The reason is that X2X_{2} carries information about HH, which can be leveraged to predict YY. As a result, g(X1,X2)g(X_{1},X_{2}) may predict YY well (and even better than f(X1)f(X_{1})) on the held-out subject if it is similar to the training subject in that the distribution of HH does not change between the subjects. If, however, HH was considerably shifted for the held-out subject, then the performance of predicting YY by g(X1,X2)g(X_{1},X_{2}) may be considerably impaired. Indeed, the invariance is violated and we have E\centernotY|X1,X2E\centernot\mbox{${}\perp\mkern-11.0mu\perp{}$}Y|X_{1},X_{2}. In contrast, the causal parent model f(X1)f(X_{1}) may have worse accuracy on the training subject but satisfies invariance: Even if the distribution of HH is different for held-out subjects compared to the training subject, the predictive performance of the model f(X1)f(X_{1}) does not change. We have EY|X1E\mbox{${}\perp\mkern-11.0mu\perp{}$}Y|X_{1}.

In practice, we often consider more than two environments. We hence have access to several environments when training our model, even if we leave out one of the environments to test on. In principle, we can thus already during training distinguish between invariant and non-invariant models. While some methods have been proposed that explicitly make use of these different environments during training time (cf. Section 4.4), we regard this as a mainly unexplored but promising area of research. In Section 4.2.1, we present a short analysis of classifying motor imagery conditions on EEG data that demonstrates how leveraging robustness may yield models that generalise better to unseen subjects.

In summary, employing distributional robustness as guiding principle prompts us to reject models as non-causal if they are not invariant or if they do not generalise better than a simple baseline to unseen environments, such as sessions, days, neuroimaging modalities, subjects, or other slight variations to the experimental setup. Models that are distributionally robust and do generalise to unseen environments are not necessarily causal but satisfy the prerequisites for being interesting candidate models when it comes to capturing the underlying causal mechanisms.

4.2.1 Exemplary proof-of-concept EEG analysis: leave-one-environment-out cross-validation

Refer to caption
Figure 3: We compare 120120 models for the prediction of 44 motor imagery tasks that leverage different EEG components, bandpower features in different frequency bands, and different classifiers. The left and right panel consider classification accuracy or AUC averaged over four held-out subjects as performance measure, respectively. The leave-one-subject-out (loso) cross-validation accuracy on 55 training subjects captures how robust a model is across training subjects. This leave-one-environment-out cross-validation scheme (see Figure 2) seems indeed able to identify models that generalise slightly better to new unseen environments (here the 44 held-out subjects) than a comparable 77-fold leave-one-block-out (lobo)-cv or the leave-one-observation-out (looo) scheme. This is reflected in the Kendall’s τ\tau rank correlation and the scores of the top-ranked models. All top-ranked models outperform random guessing on the held-out subjects (which corresponds to 25%25\% and 50%50\% in the left and right figure, respectively). The displacement along the xx-axis of the lobo- and loso-cv scores indicates the previously reported overestimation of held-out performance when using those cross-validation schemes.

Here, we illustrate the proposed cross-validation scheme presented in Figure 2 on motor imagery EEG data due to [105]. The data consist of EEG recordings of 99 subjects performing multiple trials of 44 different motor imagery tasks. For each subject 2222-channel EEG recordings at 250250 Hz sampling frequency are available for 22 days with 66 runs of 4848 trials each. We analysed the publicly available data that is bandpass filtered between 0.50.5 and 100100 Hz and 5050 Hz notch filtered. The data was further preprocessed by re-referencing to common average reference (car) and projecting onto the orthogonal complement of the null component. Arguably, the full causal structure in this problem is unkown. Instead of assessing the causal nature of a model directly, we therefore evaluate whether distributional robustness of a model across training subjects may help to find models that generalise better to new unseen subjects.

120120 models were derived on the training data comprising recordings of the first 55 subjects. Models relied on 66 different sets of extracted timeseries components: the re-referenced EEG channels, 33 different sets of 55 PCA components with varying variance-explained ratios, and 22 sets of 55 coroICA components using neighbouring covariance pairs and a partition size of 15 seconds (cf. [80] for more details on coroICA). Signals were bandpass filtered in 44 different frequency bands (8308-30, 8208-20, 203020-30, and 588058-80). For each trial and feature set, bandpower features for classification were obtained as the log-variance of the bandpass-filtered signal during seconds 363-6 of each trial. For each of the 64=246\cdot 4=24 configurations of trial features, we fitted 55 different linear discriminant analysis classifiers without shrinkage, with automatic shrinkage based on the Ledoit-Wolf lemma, and shrinkage parameter settings 0.20.2, 0.50.5, and 0.80.8. These 120120 pipelines were fitted once on the entire training data and classification accuracies and areas under the receiver operating curve scores obtained on 44 held-out subjects (yy-axes in Figure 3). Classifier performance was cross-validated on the training data following the following three different cross-validation schemes (cross-validation scores are shown on the xx-axes in Figure 3):

loso-cv

Leave-one-subject-out cross-validation is the proposed cross-validation scheme. We hold out data corresponding to each training subject once, fit an LDA classifier on the remaining training data, and assess the models accuracy on the held-out training subject. The average of those cross-validation scores reflects how robustly each of the 120120 classifier models performs across environments (here subjects).

lobo-cv

Leave-one-block-out cross-validation is a 77-fold cross-validation scheme that is similar to the above loso-cv scheme, where the training data is split into random 77 blocks of roughly equal size. Not respecting the environment structure within the training data, this cross-validation scheme does not capture a models robustness across environments.

looo-cv

Leave-one-observation-out cross-validation leaves out a single observation and is equivalent to lobo-cv with a block size of one.

In Figure 3 we display the results of the different cross-validation schemes and the Kendall’s τ\tau rank correlation between the different cv-scores derived on the training data and a model’s classification performance on the four held-out subjects. The loso-cv scores correlate more strongly with held-out model performance and thereby slightly better resolve the relative model performance. Considering the held-out performance for the models with top cv-scores, we observe that selecting models based on the loso-cv score may indeed select models that tend to perform slightly better on new unseen subjects. Furthermore, comparing the displacement of the model scores from the diagonal shows that the loso-cv scheme’s estimates are less-biased than the lobo and looo cross-validation scores, when used as an estimate for the performance on held-out subjects; this is in line with [111].

4.3 Robustness and variable constructions

Whether distributional robustness holds can depend on whether we consider the right variables. This is shown by the following example. Assume that the target YY is caused by the two brain signals X1X_{1} and X2X_{2} via

Y:=αX1+βX2+NY,Y:=\alpha X_{1}+\beta X_{2}+N_{Y}\text{,}

for some α0\alpha\neq 0, β0\beta\neq 0, and noise variable NYN_{Y}. Assume further that the environment influences the covariates X1X_{1} and X2X_{2} via X1:=X2+E+N1X_{1}:=X_{2}+E+N_{1} and X2:=E+N2X_{2}:=E+N_{2}, but does not influence YY directly. Here, X1X_{1} and X2X_{2} may represent neuronal activity in two brain regions that are causal for reaction times while EE may indicate the time of day or respiratory activity. We then have the invariance property

EY|X1,X2.E\mbox{${}\perp\mkern-11.0mu\perp{}$}Y\,|\,X_{1},X_{2}.

If, however, we were to construct or—due to limited measurement ability—only be able to observe X~:=X1+X2\widetilde{X}:=X_{1}+X_{2}, then whenever αβ\alpha\neq\beta we would find that

E\centernotY|X~.E\centernot\mbox{${}\perp\mkern-11.0mu\perp{}$}Y\,|\,\widetilde{X}.

This conditional dependence is due to many value pairs for (X1,X2)(X_{1},X_{2}) leading to the same value of X~\widetilde{X}: Given X~=x~\widetilde{X}=\widetilde{x}, the value of EE holds information about whether say (X1,X2)=(x~,0)(X_{1},X_{2})=(\widetilde{x},0) or (X1,X2)=(0,x~)(X_{1},X_{2})=(0,\widetilde{x}) is more probable and thus–since X1X_{1} and X2X_{2} enter YY with different weights–holds information about YY; EE and YY are conditionally dependent given X~\widetilde{X}. Thus, the invariance may break down when aggregating variables in an ill-suited way. This example is generic in that the same conclusions hold for all assignments X1:=f1(X2,E,N1)X_{1}:=f_{1}(X_{2},E,N_{1}) and X2:=f2(E,N2)X_{2}:=f_{2}(E,N_{2}), as long as causal minimality, a weak form of faithfulness, is satisfied [100].

Rather than taking the lack of robustness as a deficiency, we believe that this observation has the potential to help us finding the right variables and granularity to model our system of interest. If we are given several environments, the guiding principle of distributional robustness can nudge our variable definition and ROI definition towards the construction of variables that are more suitable for causally modelling some cognitive function. If some ROI activity or some EEG bandpower feature does not satisfy any invariance across environments then we may conclude that our variable representation is misaligned with the underlying causal mechanisms or that important variables have not been observed (assuming that the environments do not act on YY directly).

This idea can be illustrated by a thought experiment that is a variation of the LDL-HDL example in Section 3.2: Assume we wish to aggregate multiple voxel-activities and represent them by the activity of a ROI defined by those voxels. For example, let us consider the reaction time scenario (Example A, Section 1.1) and voxels X1,,XdX_{1},\dots,X_{d}. Then we may aggregate the voxels X¯=i=1dXi\overline{X}=\sum_{i=1}^{d}X_{i} to obtain a macro-level model in which we can still sensibly reason about the effect of an intervention on the treatment variable TT onto the distribution of X¯\overline{X}, the ROIs average activity. Yet, the model is in general not causally consistent with the original model. First, our ROI may be chosen too coarse such that for 𝐱^𝐱~d\mathbf{\widehat{x}}\neq\mathbf{\widetilde{x}}\in\mathbb{R}^{d} with i=1d𝐱^i=i=1d𝐱~i=x¯\sum_{i=1}^{d}\mathbf{\widehat{x}}_{i}=\sum_{i=1}^{d}\mathbf{\widetilde{x}}_{i}=\overline{x} the interventional distributions induced by the micro-level model corresponding to setting all Xi:=𝐱~iX_{i}:=\mathbf{\widetilde{x}}_{i} or alternatively to Xi:=𝐱^iX_{i}:=\mathbf{\widehat{x}}_{i} do not coincide—for example, a ROI that effectively captures the global average voxel-activity cannot resolve whether a higher activity is due to increased reaction-time-driving neuronal entities or due to some upregulation of other neuronal processes unrelated to the reaction time, such as respiratory activity. This ROI would be ill-suited for causal reasoning and non-robust as there are two micro-level interventions that imply different distributions on the reaction times while corresponding to the same intervention setting X¯:=x¯\overline{X}:=\overline{x} with x¯=i=1d𝐱^i=i=1d𝐱~i\overline{x}=\sum_{i=1}^{d}\mathbf{\widehat{x}}_{i}=\sum_{i=1}^{d}\mathbf{\widetilde{x}}_{i} in the macro-level model. Second, our ROI may be defined too fine grained such that, for example, the variable representation does only pick up on the left-hemisphere hub of a distributed neuronal process relevant for reaction times. If the neuronal process has different laterality in different subjects, then predicting the effects of interventions on only the left-hemispherical neuronal activity cannot be expected to translate to all subjects. Here, a macro-variable that averages more voxels, say symmetric of both hemispheres, may be more robust to reason about the causes of reaction times than the more fine grained unilateral ROI. In this sense, seeking for variable constructions that enable distributionally robust models across subjects, may nudge us to meaningful causal entities. The spatially refined and finer resolved cognitive atlas obtained by [112], whose map definition procedure was geared towards an atlas that would be robustness across multiple studies and 196196 different experimental conditions, may be seen as an indicative manifestation of the above reasoning.

4.4 Existing methods exploiting robustness

We now present some existing methods that explicitly consider the invariance of a model. While many of these methods are still in their infancy when considering real world applications, we believe that further development in that area could play a vital role when tackling causal questions in cognitive neuroscience.

Robust Independent Component Analysis.

Independent component analysis (ICA) is commonly performed in the analysis of magneto- and electro-electroencephalography (MEG and EEG) data in order to invert the inevitable measurement transformation that leaves us with observations of a linear superposition of underlying cortical (and non-cortical) activity. The basic ICA model assumes our vector of observed variables XX is being generated as X=ASX=AS where AA is a mixing matrix and S=[S1,,Sd]S=[S_{1},\dots,S_{d}]^{\top} is a vector of unobserved mutually independent source signals. The aim is to find the unmixing matrix V=A1V=A^{-1}. If we perform ICA on individual subjects’ data separately, the resulting unmixing matrices will often differ between subjects. This not only hampers the interpretation of the resulting sources as some cortical activity that we can identify across subjects, it also hints—in light of the above discussion—at some unexplained variation that is due to shifts in background conditions between subjects such as different cap positioning or neuroanatomical variation. Instead of simply pooling data across subjects, [80] propose a methodology that explicitly exploits the existence of environments, that is, the fact that EEG samples can be grouped by subjects they were recorded from. This way, the proposed confounding-robust ICA (coroICA) procedure identifies an unmixing of the signals that generalises to new subjects. The additional robustness resulted, for their considered example, in improved classification accuracies on held-out subjects and can be viewed as a first-order adjustment for subject specific differences. The application of ICA procedures to pooled data will generally result in components that do not robustly transfer to new subjects and are thus necessarily variables that do not lend themselves for a causal interpretation. The coroICA procedure aims to exploit the environments to identify unmixing matrices that generalise across subjects.

Causal discovery with exogenous variation.

Invariant causal prediction, proposed by [74], aims at identifying the parents of YY within a set of covariates X1,,XdX_{1},\ldots,X_{d}. We have argued that the true causal model is invariant across environments, see Equation (2), if the data are obtained in different environments and the environment does not directly influence YY. That is, when enumerating all invariant models by searching through subsets of X1,,XdX_{1},\ldots,X_{d}, one of these subsets must be the set of causal parents of YY. As a result, the intersection S^=S:SinvariantS\widehat{S}=\cap_{S:S\text{invariant}}S of all sets of covariates that yield invariant models is guaranteed to be a subset of the causes 𝐏𝐀(Y){\mathbf{PA}(Y)} of YY. (Here, we define the intersection over the empty index set as the empty set.) Testing invariance with a hypothesis test to the level α\alpha, say α=0.05\alpha=0.05, one obtains that S^\widehat{S} is contained in the set of parents of YY with high probability

𝐏(S^𝐏𝐀(Y))1α.\mathbf{P}\left(\widehat{S}\subseteq{\mathbf{PA}(Y)}\right)\geq 1-\alpha.

Under faithfulness, the method can be shown to be robust against violation of the above assumptions. If the environment acts on YY directly, for example, there is no invariant set and in the presence of hidden variables, the intersection S^\widehat{S} of invariant models can still be shown to be a subset of the ancestors of YY with large probability.

It is further possible to model the environment as a random variable (using an indicator variable, for example), that is often called a context variable. One can then exploit the background knowledge of its exogeneity to identify the full causal structure instead of focussing on identifying the causes of a target variable. Several approaches have been suggested [[, for example,]]Spirtes2000, Eaton2007, Zhang-ijcai2017, Mooij++_JMLR_20. Often, these methods first identify the target of the intervention and then exploit known techniques of constraint- or score-based methods. Some of the above methods also make use of time as a context variable or environment [125, 79].

Anchor regression.

We argued above that focusing on invariance has an advantage when inferring causal structure from data. If we are looking for generalisability across environments, however, focusing solely on invariance may be too restrictive. Instead, we may select the most predictive model among all invariant models. The idea of anchor regression is to explicitly trade off invariance and predictability [87]. For a target variable YY, predictor variables 𝐗\mathbf{X}, and so-called anchor variables 𝐀=[A1,,Aq]\mathbf{A}=[A_{1},\dots,A_{q}]^{\top} that represent the different environments and are normalised to have unit variance, the anchor regression coefficients are obtained as solutions to the following minimisation problem

b^γ:=argminbd𝔼[(Yb𝐗)2]prediction+γ𝔼[𝐀(Yb𝐗)22]invariance.\widehat{b}_{\gamma}:=\arg\min_{{b}\in\mathbb{R}^{d}}\ \underbrace{\mathbb{E}\left[(Y-b^{\top}\mathbf{X})^{2}\right]}_{\text{prediction}}+\gamma\underbrace{\mathbb{E}\left[\|\mathbf{A}(Y-b^{\top}\mathbf{X})\|^{2}_{2}\right]}_{\text{invariance}}.

Higher parameters γ\gamma steer the regression towards more invariant predictions (converging against the two stage least squares solutions in identifiable instrumental variable settings). For γ=0\gamma=0 we recover the ordinary least square solution. The solution b^γ\widehat{b}_{\gamma} can be shown to have the best predictive power under shift interventions up to a certain strength that depends on γ\gamma. As before, the anchor variables can code time, environments, subjects, or other factors, and we thus obtain a regression that is distributionally robust against shifts in those factors.

4.5 Time series data

So far, we have mainly focused on the setting of i.i.d. data. Most of the causal inference literature dealing with time dependency considers discrete-time models. This comes with additional complications for causal inference. For example, there are ongoing efforts to adapt causal inference algorithms and account for sub- or sup-sampling and temporal aggregation [23, 46]. Problems of temporal aggregation relate to Challenge 2 of finding the right variables, which is a conceptual problem in time series models that requires us to clarify our notion of intervention for time-evolving systems [88]. When we observe time series with non-stationarities we may consider these as resulting from some unknown shift interventions. That is, non-stationarities over time may be due to shifts in the background conditions and as such can be understood as shifts in environments analogous to the i.i.d. setting. This way, we may again leverage the idea of distributional robustness for inference on time-evolving systems for which targeted interventional data is scarce. Extensions of invariant causal prediction to time series data that aim to leverage such variation have been proposed by [20, 79] and the ICA procedure described in Section 4.4 also exploits non-stationarity over time. SCMs extend to continuous-time models [72], where the idea to trade off prediction and invariance has been applied to the problem of inferring chemical reaction networks [78].

A remark is in order if we wish to describe time-evolving systems by one causal summary graph where each time series component is collapsed into one node: For this to be reasonable, we need to assume a time-homogeneous causal structure. Furthermore, it requires us to carefully clarify its causal semantics: While summary graphs can capture the existence of cause-effect relationships, they do in general not correspond to a structural causal model that admits a causal semantics nor enables interventional predictions that are consistent with the underlying time-resolved structural causal model [88, 49]. That is, the wrong choice of time-agnostic variables and corresponding interventions may be ill-suited to represent the cause-effect relationships of a time-evolving system [[, cf. Challenge 2 and]]rubenstein2017causal,janzing2018structural.

5 Conclusion and future work

Causal inference in cognitive neuroscience is ambitious. It is important to continue the open discourse about the many challenges, some of which are mentioned above. Thanks to the open and critical discourse there is great awareness and caution when interpreting neural correlates [84]. Yet, “FC [functional connectivity] researchers already work within a causal inference framework, whether they realise it or not” [86].

In this article we have provided our view on the numerous obstacles to a causal understanding of cognitive function. If we, explicitly or often implicitly, ask causal questions, we need to employ causal assumptions and methodology. We propose to exploit that causal models using the right variables are distributionally robust. In particular, we advocate distributional robustness as a guiding principle for causality in cognitive neuroscience. While causal inference in general and in cognitive neuroscience in particular is a challenging task, we can at least exploit the rational to refute models and variables as non-causal that are frail to shifts in the environment. This guiding principle does not necessarily identify causal variables nor causal models, but it nudges our search into the right direction away from frail models and non-causal variables. While we presented first attempts that aim to leverage observations obtained in different environments (cf. Section 4.4), this article poses more questions for future research than it answers.

We believe that procedures that exploit environments during training are a promising avenue for future research. While we saw mildly positive results in our case study, further research needs to show whether this trend persists in studies with many subjects. It may be possible to obtain improvements when combining predictive scores on held-out-subjects with other measures of invariance and robustness. The development of such methods may be spurred and guided by field-specific benchmarks (or competitions) that assess models’ distributional robustness across a wide range of scenarios, environments, cognitive tasks, and subjects.

When considering robustness or invariance across trainings environments, the question arises how the ability to infer causal structure and the generalisation performance to unseen environments depend on the number of training environments. While first attempts have been made to theoretically understand that relation [87, 21], most of the underlying questions are still open. We believe that an answer would depend on the strength of the interventions, the sample size, the complexity of the model class and, possibly, properties of the (meta-)distribution of environments.

We believe that advancements regarding the errors-in-variable problem may have important implications for cognitive neuroscience. Nowadays, we can obtain neuroimaging measurements at various spatial and temporal resolutions using, among others, fMRI, MEG and EEG, positron emission tomography, or near-infrared spectroscopy [31, 81]. Yet, all measurement modalities are imperfect and come with different complications. One general problem is that the observations are corrupted by measurement noise. The errors-in-variables problem complicates even classical regression techniques where we wish to model YβX+ϵY\approx\beta X^{*}+\epsilon but only have access to observations of a noise-corrupted X=X+ηX=X^{*}+\eta [92]. This inevitably carries over and hurdles causal inference as the measurement noise spoils conditional independence testing, biases any involved regression steps, and troubles additive noise approaches that aim to exploit noise properties for directing causal edges and methods testing for invariance. First steps addressing these issues in the context of causal discovery have been proposed by [124, 8, 91].

Summarising, we believe that there is a need for causal models if we aim to understand the neuronal underpinnings of cognitive function. Only causal models equip us with concepts that allow us to explain, describe, predict, manipulate, deal and interact with, and reason about a system and that allow us to generalise to new, unseen environments. A merely associational model suffices to predict naturally unfolding disease progression, for example. We need to obtain understanding in form of a causal model if our goal is to guide rehabilitation after cognitive impairment or to inform the development of personalised drugs that target specific neuronal populations. Distributional robustness and generalisability to unseen environments is an ambitious goal, in particular in biological systems and even more so in complex systems such as the human brain. Yet, it may be the only and most promising way forward.

Acknowledgments

The authors thank the anonymous reviewers for their constructive and helpful feedback on an earlier version of this manuscript. SW was supported by the Carlsberg Foundation.

References

  • [1] J. Aldrich “Autonomy” In Oxford Economic Papers 41, 1989, pp. 15–34
  • [2] Andrea Antal and Christoph S Herrmann “Transcranial alternating current and random noise stimulation: possible mechanisms” In Neural Plasticity 2016 Hindawi, 2016
  • [3] Dominik R Bach, Mkael Symmonds, Gareth Barnes and Raymond J Dolan “Whole-Brain Neural Dynamics of Probabilistic Reward Prediction” In Journal of Neuroscience 37.14, 2017, pp. 3789–3798
  • [4] Sander Beckers, Frederick Eberhardt and Joseph Y Halpern “Approximate Causal Abstraction” In Proceedings of the Thirty-Fifth Conference on Uncertainty in Artificial Intelligence (UAI) AUAI Press, 2019
  • [5] Sander Beckers and Joseph Y Halpern “Abstracting Causal Models” In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence AAAI Press, 2019
  • [6] Til Ole Bergmann and Gesa Hartwigsen “Inferring Causality from Noninvasive Brain Stimulation in Cognitive Neuroscience” Forthcoming In Journal of Cognitive Neuroscience, 2020, pp. 1–29
  • [7] Sven Bestmann and Vincent Walsh “Transcranial electrical stimulation” In Current Biology 27.23 Elsevier, 2017, pp. R1258–R1262
  • [8] T. Blom, A. Klimovskaia, S. Magliacane and J.. Mooij “An Upper Bound for Random Measurement Error in Causal Discovery” In Proceedings of the 34th Annual Conference on Uncertainty in Artificial Intelligence (UAI), 2018
  • [9] Kenneth A Bollen “Structural Equations with Latent Variables” John Wiley & Sons, 1989
  • [10] Stephan Bongers, Jonas Peters, Bernhard Schölkopf and Joris M Mooij “Theoretical Aspects of Cyclic Structural Causal Models” In arXiv preprint arXiv:1611.06221, 2018
  • [11] Michael Breakspear “Dynamic and stochastic models of neuroimaging data: A comment on Lohmann et al.” In NeuroImage 75, 2013, pp. 270–274
  • [12] P. Bühlmann, J. Peters and J. Ernest “CAM: Causal Additive Models, high-dimensional order search and penalized regression” In Annals of Statistics 42, 2014, pp. 2526–2556
  • [13] N. Cartwright “Two Theorems on Invariance and Causality” In Philosophy of Science 70.1, 2003, pp. 203–224
  • [14] N. Cartwright “Causality, invariance and policy” In The Oxford handbook of philosophy of economics New York, NY: Oxford University Press, 2009, pp. 410–423
  • [15] Krzysztof Chalupka, Frederick Eberhardt and Pietro Perona “Multi-Level Cause-Effect Systems” In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics 51, Proceedings of Machine Learning Research PMLR, 2016, pp. 361–369
  • [16] Krzysztof Chalupka, Pietro Perona and Frederick Eberhardt “Visual Causal Feature Learning” In Proceedings of the Thirty-First Conference on Uncertainty in Artificial Intelligence (UAI) AUAI Press, 2015
  • [17] Daniel Chicharro and Stefano Panzeri “Algorithms of causal inference for the analysis of effective connectivity among brain regions” In Frontiers in Neuroinformatics 8, 2014, pp. 64
  • [18] D.. Chickering “Optimal structure identification with greedy search” In Journal of Machine Learning Research 3, 2002, pp. 507–554
  • [19] G.. Chow “Tests of equality between sets of coefficients in two linear regressions” In Econometrica 28, 1960, pp. 591–605
  • [20] R. Christiansen and J. Peters “Invariant Causal Prediction in the Presence of Latent Variables” In arXiv preprint arXiv:1808.05541, 2018
  • [21] R. Christiansen et al. “The Difficult Task of Distribution Generalization in Nonlinear Models” In arXiv preprint arXiv:2006.07433, 2020
  • [22] Denis Conniffe “R.A. Fisher and the development of statistics - a view in his centerary year” In Journal of the Statistical and Social Inquiry Society of Ireland 26.3, 1991, pp. 55–108
  • [23] David Danks and Sergey Plis “Learning causal structure from undersampled time series” Carnegie Mellon University, 2013
  • [24] G. Darmois “Analyse générale des liaisons stochastiques” In Revue de l’Institut International de Statistique 21, 1953, pp. 2–8
  • [25] Tyler Davis et al. “What do differences between multi-voxel and univariate analysis mean? How subject-, voxel-, and trial-level variance impact fMRI analysis” In NeuroImage 97, 2014, pp. 271–283
  • [26] Julien Dubois et al. “Causal mapping of emotion networks in the human brain: Framework and initial findings” In Neuropsychologia, 2017
  • [27] D. Eaton and K.. Murphy “Exact Bayesian structure learning from uncertain interventions” In Proceedings of the 11th International Conference on Artificial Intelligence and Statistics (AISTATS), 2007, pp. 107–114
  • [28] Frederick Eberhardt “Experimental indistinguishability of causal structures” In Philosophy of Science 80.5 University of Chicago Press, 2013, pp. 684–696
  • [29] Frederick Eberhardt “Green and grue causal variables” In Synthese 193.4, 2016, pp. 1029–1046
  • [30] Gaute T Einevoll, Christoph Kayser, Nikos K Logothetis and Stefano Panzeri “Modelling and analysis of local field potentials for studying the function of cortical circuits” In Nature Reviews Neuroscience 14.11 Nature Publishing Group, 2013, pp. 770–785
  • [31] Aaron Filler “The history, development and impact of computed imaging in neurological diagnosis and neurosurgery: CT, MRI, and DTI” In Nature Precedings Nature Publishing Group, 2009
  • [32] K.. Friston, L. Harrison and W. Penny “Dynamic causal modelling” In NeuroImage 19.4, 2003, pp. 1273–1302
  • [33] Karl Friston, Jean Daunizeau and Klaas Enno Stephan “Model selection and gobbledygook: Response to Lohmann et al.” In NeuroImage 75, 2013, pp. 275–278
  • [34] C… Granger “Investigating Causal Relations by Econometric Models and Cross-spectral Methods” In Econometrica 37.3, 1969, pp. 424–438
  • [35] Moritz Grosse-Wentrup, Dominik Janzing, Markus Siegel and Bernhard Schölkopf “Identification of causal relations in neuroimaging data with latent confounders: An instrumental variable approach” In NeuroImage 125, 2016, pp. 825–833
  • [36] L. Györfi, M. Kohler, A. Krzyżak and H. Walk “A Distribution-Free Theory of Nonparametric Regression” Springer, 2002
  • [37] T. Haavelmo “The Probability Approach in Econometrics” In Econometrica 12, 1944, pp. S1–S115 (supplement)
  • [38] S. Haufe et al. “On the interpretation of weight vectors of linear models in multivariate neuroimaging” In NeuroImage 87, 2014, pp. 96–110
  • [39] Suzana Herculano-Houzel “The remarkable, yet not extraordinary, human brain as a scaled-up primate brain and its associated cost” In Proceedings of the National Academy of Sciences 109.Supplement 1 National Academy of Sciences, 2012, pp. 10661–10668
  • [40] Christoph S Herrmann, Stefan Rach, Toralf Neuling and Daniel Strüber “Transcranial alternating current stimulation: a review of the underlying mechanisms and modulation of cognitive processes” In Frontiers in human neuroscience 7 Frontiers, 2013, pp. 279
  • [41] Erik P. Hoel “When the Map Is Better Than the Territory” In Entropy 19.5, 2017
  • [42] Erik P Hoel, Larissa Albantakis and Giulio Tononi “Quantifying causal emergence shows that macro can beat micro” In Proceedings of the National Academy of Sciences 110.49, 2013, pp. 19790–19795
  • [43] K.. Hoover “Causality in Economics and Econometrics” In The New Palgrave Dictionary of Economics Basingstoke, UK: Palgrave Macmillan, 2008
  • [44] Patrik O Hoyer et al. “Nonlinear causal discovery with additive noise models” In Advances in Neural Information Processing Systems 21 (NeurIPS) NeurIPS Foundation, 2008, pp. 689–696
  • [45] Alexander G Huth et al. “Decoding the Semantic Content of Natural Movies from Human Brain Activity” In Frontiers in Systems Neuroscience 10, 2016, pp. 81
  • [46] Antti Hyttinen et al. “Causal discovery from subsampled time series data by constraint optimization” In Proceedings of the Eighth International Conference on Probabilistic Graphical Models 52, Proceedings of Machine Learning Research PMLR, 2016, pp. 216–227
  • [47] G.. Imbens and D.. Rubin “Causal Inference for Statistics, Social, and Biomedical Sciences” Cambridge University Press, 2015
  • [48] Fattaneh Jabbari, Joseph Ramsey, Peter Spirtes and Gregory Cooper “Discovery of causal models that contain latent variables through Bayesian scoring of independence constraints” In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2017, pp. 142–157 Springer
  • [49] Dominik Janzing, Paul Rubenstein and Bernhard Schölkopf “Structural causal models for macro-variables in time-series” In arXiv preprint arXiv:1804.03911, 2018
  • [50] Kohitij Kar, Takuya Ito, Michael W. Cole and Bart Krekelberg “Transcranial alternating current stimulation attenuates BOLD adaptation and increases functional connectivity” In Journal of Neurophysiology 123.1, 2020, pp. 428–438
  • [51] S. Lauritzen “Graphical Models” Oxford University Press, 1996
  • [52] S.. Lauritzen, A.. Dawid, B.. Larsen and H.-G. Leimer “Independence properties of directed markov fields” In Networks 20, 1990, pp. 491–505
  • [53] Jin Hyung Lee et al. “Global and local fMRI signals driven by neurons defined optogenetically by type and wiring” In Nature 465.7299 Nature Publishing Group, 2010, pp. 788–792
  • [54] Zhifeng Liang et al. “Mapping the functional network of medial prefrontal cortex by combining optogenetics and fMRI in awake rats” In NeuroImage 117, 2015, pp. 114–123
  • [55] Nikos K Logothetis “What we can do and what we cannot do with fMRI” In Nature 453.7197 Nature Publishing Group, 2008, pp. 869–878
  • [56] Gabriele Lohmann, Kerstin Erfurth, Karsten Müller and Robert Turner “Critical comments on dynamic causal modelling” In NeuroImage 59.3, 2012, pp. 2322–2329
  • [57] Gabriele Lohmann, Karsten Müller and Robert Turner “Response to commentaries on our paper: Critical comments on dynamic causal modelling” In NeuroImage 75, 2013, pp. 279–281
  • [58] Virginia López-Alonso, Binith Cheeran, Dan Rio-Rodriguez and Miguel Fernández-del-Olmo “Inter-individual variability in response to non-invasive brain stimulation paradigms” In Brain stimulation 7.3 Elsevier, 2014, pp. 372–380
  • [59] R. Lucas “Econometric policy evaluation: A critique” In Carnegie-Rochester Conference Series on Public Policy 1.1, 1976, pp. 19–46
  • [60] I. M. and D. Lopez-Paz “Invariant Risk Minimization” In arXiv preprint arXiv:1907.02893, 2019
  • [61] D. Marinazzo, M. Pellicoro and S. Stramaglia “Kernel-Granger causality and the analysis of dynamical networks” In Physical Review E 77.5 American Physical Society, 2008, pp. 056215
  • [62] Daniele Marinazzo, Wei Liao, Huafu Chen and Sebastiano Stramaglia “Nonlinear connectivity by Granger causality” In NeuroImage 58.2, 2011, pp. 330–338
  • [63] A Mastakouri, B Schölkopf and D Janzing “Selecting causal brain features with a single conditional independence test per feature” In Advances in Neural Information Processing Systems 32 (NeurIPS) NeurIPS Foundation, 2019, pp. 12532–12543
  • [64] Helen S Mayberg et al. “Deep Brain Stimulation for Treatment-Resistant Depression” In Neuron 45.5 Elsevier, 2005, pp. 651–660
  • [65] David Marc Anton Mehler and Konrad Paul Kording “The lure of causal statements: Rampant mis-inference of causality in estimated connectivity” In arXiv preprint arXiv:1812.03363, 2018
  • [66] Ravi D Mill, Takuya Ito and Michael W Cole “From connectome to cognition: The search for mechanism in human functional brain networks” In NeuroImage 160 Elsevier, 2017, pp. 124–139
  • [67] J.. Mooij, S. Magliacane and T. Claassen “Joint Causal Inference from Multiple Contexts” In Journal of Machine Learning Research 21.99, 2020, pp. 1–108
  • [68] Jeanette A Mumford and Joseph D Ramsey “Bayesian networks for fMRI: A primer” In NeuroImage 86, 2014, pp. 573–582
  • [69] Michael A. Nitsche et al. “Transcranial direct current stimulation: State of the art 2008” In Brain Stimulation 1.3, 2008, pp. 206–223
  • [70] P.. Nunez and R. Srinivasan “Electric Fields of the Brain” Oxford University Press, 2006
  • [71] Judea Pearl “Causality” Cambridge University Press, 2009
  • [72] J. Peters, S. Bauer and N. Pfister “Causal models for dynamical systems” In arXiv preprint arXiv:2001.06208, 2020
  • [73] J. Peters and P. Bühlmann “Identifiability of Gaussian Structural Equation Models with Equal Error Variances” In Biometrika 101.1, 2014, pp. 219–228
  • [74] J. Peters, P. Bühlmann and N. Meinshausen “Causal inference using invariant prediction: identification and confidence intervals” In Journal of the Royal Statistical Society: Series B (with discussion) 78.5, 2016, pp. 947–1012
  • [75] J. Peters, D. Janzing and B. Schölkopf “Causal Inference on Time Series using Structural Equation Models” In Advances in Neural Information Processing Systems 26 (NeurIPS) NeurIPS Foundation, 2013, pp. 585–592
  • [76] J. Peters, D. Janzing and B. Schölkopf “Elements of Causal Inference” MIT Press, 2017
  • [77] J. Peters, J.. Mooij, D. Janzing and B. Schölkopf “Causal Discovery with Continuous Additive Noise Models” In Journal of Machine Learning Research 15, 2014, pp. 2009–2053
  • [78] N. Pfister, S. Bauer and J. Peters “Learning stable and predictive structures in kinetic systems” In Proceedings of the National Academy of Sciences 116.51 National Academy of Sciences, 2019, pp. 25405–25411
  • [79] N. Pfister, P. Bühlmann and J. Peters “Invariant Causal Prediction for Sequential Data” In Journal of the American Statistical Association 114.527, 2018, pp. 1264–1276
  • [80] Niklas Pfister, Sebastian Weichwald, Peter Bühlmann and Bernhard Schölkopf “Robustifying Independent Component Analysis by Adjusting for Group-Wise Stationary Noise” In Journal of Machine Learning Research 20.147, 2019, pp. 1–50
  • [81] Russell A Poldrack “The new mind readers: What neuroimaging can and cannot reveal about our thoughts” Princeton University Press, 2018
  • [82] Jonathan D Power et al. “Functional network organization of the human brain” In Neuron 72.4 Elsevier, 2011, pp. 665–678
  • [83] J D Ramsey et al. “Six problems for causal inference from fMRI” In NeuroImage 49.2, 2010, pp. 1545–1558
  • [84] Geraint Rees, Gabriel Kreiman and Christof Koch “Neural correlates of consciousness in humans” In Nature Reviews Neuroscience 3, 2002, pp. 261–270
  • [85] H. Reichenbach “The Direction of Time” University of California Press, 1956
  • [86] Andrew T Reid et al. “Advancing functional connectivity research from association to causation” In Nature neuroscience 1.10 Nature Publishing Group, 2019
  • [87] Dominik Rothenhäusler, Nicolai Meinshausen, Peter Bühlmann and Jonas Peters “Anchor regression: heterogeneous data meets causality” In arXiv preprint arXiv:1801.06229, 2018
  • [88] Paul K Rubenstein et al. “Causal Consistency of Structural Equation Models” In Proceedings of the Thirty-Third Conference on Uncertainty in Artificial Intelligence (UAI), 2017
  • [89] Ruben Sanchez-Romero and Michael W Cole “Combining Multiple Functional Connectivity Methods to Improve Causal Inferences” Forthcoming In Journal of Cognitive Neuroscience, 2020, pp. 1–15
  • [90] Ruben Sanchez-Romero et al. “Estimating feedforward and feedback effective connections from fMRI time series: Assessments of statistical methods” In Network Neuroscience 3.2, 2019, pp. 274–306
  • [91] R. Scheines and J. Ramsey “Measurement error and causal discovery” In CEUR Workshop Proceedings 1792, 2016, pp. 1–7
  • [92] Susanne M Schennach “Recent advances in the measurement error literature” In Annual Review of Economics 8 Annual Reviews, 2016, pp. 341–377
  • [93] B. Schölkopf et al. “On causal and anticausal learning” In Proceedings of the 29th International Conference on Machine Learning (ICML) Omnipress, 2012, pp. 1255–1262
  • [94] R. Shah and J. Peters “The Hardness of Conditional Independence Testing and the Generalised Covariance Measure” In Annals of Statistics (accepted), ArXiv e-prints (1804.07203), 2020
  • [95] S. Shimizu, P.. Hoyer, A. Hyvärinen and A. Kerminen “A Linear Non-Gaussian Acyclic model for Causal Discovery” In Journal of Machine Learning Research 7.10, 2006, pp. 2003–2030
  • [96] Ricardo Silva, Richard Scheine, Clark Glymour and Peter Spirtes “Learning the structure of linear latent variable models” In Journal of Machine Learning Research 7, 2006, pp. 191–246
  • [97] V.. Skitovič “Linear combinations of independent random variables and the normal distribution law” In Selected Translations in Mathematical Statistics and Probability 2, 1962, pp. 211–228
  • [98] Stephen M Smith “The future of FMRI connectivity” In Neuroimage 62.2 Elsevier, 2012, pp. 1257–1266
  • [99] Stephen M Smith et al. “Network modelling methods for FMRI” In Neuroimage 54.2 Elsevier, 2011, pp. 875–891
  • [100] Peter Spirtes, Clark N Glymour and Richard Scheines “Causation, Prediction, and Search” MIT Press, 2001
  • [101] Peter Spirtes and Richard Scheines “Causal Inference of Ambiguous Manipulations” In Philosophy of Science 71.5, 2004, pp. 833–845
  • [102] Daniel Steinberg “The Cholesterol Wars” Academic Press, 2007
  • [103] S. Stramaglia, Guo-Rong Wu, M. Pellicoro and D. Marinazzo “Expanding the transfer entropy to identify information circuits in complex systems” In Physical Review E 86.6 American Physical Society, 2012, pp. 066211
  • [104] Sebastiano Stramaglia, Jesus M Cortes and Daniele Marinazzo “Synergy and redundancy in the Granger causal analysis of dynamical networks” In New Journal of Physics 16.10 IOP Publishing, 2014, pp. 105003
  • [105] M. Tangermann et al. “Review of the BCI Competition IV” In Frontiers in Neuroscience 6, 2012, pp. 55
  • [106] Michael T Todd, Leigh E Nystrom and Jonathan D Cohen “Confounds in multivariate pattern analysis: Theory and rule representation case study” In NeuroImage 77, 2013, pp. 157–165
  • [107] A Stewart Truswell “Cholesterol and Beyond” Springer, 2010
  • [108] Lucina Q. Uddin, B.. Yeo and R. Spreng “Towards a Universal Taxonomy of Macro-scale Functional Human Brain Networks” In Brain Topography 32.6, 2019, pp. 926–942
  • [109] C. Uhler, G. Raskutti, P. Bühlmann and B. Yu “Geometry of the faithfulness assumption in causal inference” In Annals of Statistics 41.2, 2013, pp. 436–463
  • [110] Pedro A Valdes-Sosa, Alard Roebroeck, Jean Daunizeau and Karl Friston “Effective connectivity: influence, causality and biophysical modeling” In Neuroimage 58.2 Elsevier, 2011, pp. 339–361
  • [111] Gaël Varoquaux et al. “Assessing and tuning brain decoders: Cross-validation, caveats, and guidelines” Individual Subject Prediction In NeuroImage 145, 2017, pp. 166–179
  • [112] Gaël Varoquaux et al. “Atlases of cognition with large-scale human brain mapping” In PLOS Computational Biology 14.11, 2018, pp. 1–18
  • [113] Thomas Verma and Judea Pearl “Equivalence and Synthesis of Causal Models” In Proceedings of the Sixth Annual Conference on Uncertainty in Artificial Intelligence (UAI), 1990
  • [114] Johannes Vosskuhl, Daniel Strüber and Christoph S Herrmann “Non-invasive brain stimulation: a paradigm shift in understanding brain oscillations” In Frontiers in human neuroscience 12 Frontiers, 2018, pp. 211
  • [115] Lourens Waldorp, Ingrid Christoffels and Vincent Ven “Effective connectivity of fMRI data using ancestral graph theory: Dealing with missing regions” In NeuroImage 54.4, 2011, pp. 2695–2705
  • [116] Sebastian Weichwald “Pragmatism and Variable Transformations in Causal Modelling”, 2019
  • [117] Sebastian Weichwald, Arthur Gretton, Bernhard Schölkopf and Moritz Grosse-Wentrup “Recovery of non-linear cause-effect relationships from linearly mixed neuroimaging data” In Pattern Recognition in Neuroimaging (PRNI), 2016 International Workshop on, 2016 IEEE
  • [118] Sebastian Weichwald, Moritz Grosse-Wentrup and Arthur Gretton “MERLiN: Mixture Effect Recovery in Linear Networks” In IEEE Journal of Selected Topics in Signal Processing 10.7, 2016, pp. 1254–1266
  • [119] Sebastian Weichwald et al. “Causal interpretation rules for encoding and decoding models in neuroimaging” In NeuroImage 110, 2015, pp. 48–59
  • [120] Sebastian Weichwald, Bernhard Schölkopf, Tonio Ball and Moritz Grosse-Wentrup “Causal and anti-causal learning in pattern recognition for neuroimaging” In Pattern Recognition in Neuroimaging (PRNI), 2014 International Workshop on, 2014 IEEE
  • [121] James Woodward “Making Things Happen: A Theory of Causal Explanation” Oxford University Press, 2005
  • [122] Alex Woolgar, Polina Golland and Stefan Bode “Coping with confounds in multivoxel pattern analysis: What should we do about reaction time differences? A comment on Todd, Nystrom & Cohen 2013” In NeuroImage 98, 2014, pp. 506–512
  • [123] B.. Yeo et al. “The organization of the human cerebral cortex estimated by intrinsic functional connectivity” In Journal of Neurophysiology 106.3, 2011, pp. 1125–1165
  • [124] K. Zhang et al. “Causal Discovery with Linear Non-Gaussian Models underMeasurement Error: Structural Identifiability Results” In Proceedings of the 34th Annual Conference on Uncertainty in Artificial Intelligence (UAI) AUAI Press, 2018
  • [125] K. Zhang et al. “Causal Discovery from Nonstationary/Heterogeneous Data: Skeleton Estimation and Orientation Determination” In Proceedings of the 26th International Joint Conference on Artificial Intelligence (IJCAI), 2017, pp. 1347–1353
  • [126] K. Zhang and A. Hyvärinen “On the Identifiability of the Post-Nonlinear Causal Model” In Proceedings of the Twenty-Fifth Annual Conference on Uncertainty in Artificial Intelligence (UAI) AUAI Press, 2009
  • [127] Xun Zheng et al. “Learning sparse nonparametric DAGs” (to appear) In Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020