Skip to main content icon/video/no-internet

From their inception in the early 20th century, causal systems models (more commonly known as structural-equations models) were accompanied by graphical representations or path diagrams that provided compact summaries of qualitative assumptions made by the models. Figure 1 provides a graph that would correspond to any system of five equations encoding these assumptions:

  • Independence of A and B
  • Direct dependence of C on A and B
  • Direct dependence of E on A and C
  • Direct dependence of F on C
  • Direct dependence of D on B, C, and E

The interpretation of ‘direct dependence’ was kept rather informal and usually conveyed by causal intuition, for example, that the entire influence of A on F is ‘mediated’ by C.

By the 1980s, it was recognized that these diagrams could be reinterpreted formally as probability models, which opened the visual power of graph theory for use in probabilistic inference and allowed easy deduction of other independence conditions implied by the assumptions. By the 1990s, it was further recognized that these diagrams could also be used as a formal tool for causal inference, such as predicting the effects of external interventions. Given that the graph is correct, one can see whether the causal effects of interest (target effects, or causal estimands) can be estimated from available data, or what additional observations are needed to validly estimate those effects. One can also see how to represent the effects as familiar standardized effect measures.

Figure 1 Example of a Directed Acrylic Graph

None

This entry gives an overview of (1) components of causal graph theory, (2) probability interpretations of graphical models, and (3) the methodological implications of the causal and probability structures encoded in the graph.

Basics of Graph Theory

As befitting a well-developed mathematical topic, graph theory has an extensive terminology that, once mastered, provides access to a number of elegant results that may be used to model any system of relations. The term dependence in a graph, usually represented by connectivity, may refer to mathematical, causal, or statistical dependencies. The connectives joining variables in the graph are called arcs, edge, or links, and the variables are also called nodes or vertices. Two variables connected by an arc are adjacent or neighbors, and arcs that meet at a variable are also adjacent. If the arc is an arrow, the tail (starting) variable is the parent and the head (ending) variable is the child. In causal diagrams, an arrow represents a ‘direct effect’ of the parent on the child, although this effect is direct only relative to a certain level of abstraction, in that the graph omits any variables that might mediate the effect.

A variable that has no parent (such as A and B in Figure 1) is exogenous or external, or a root or source node, and is determined only by forces outside the graph; otherwise it is endogenous or internal. A variable with no children (such as D in Figure 1) is a sink or terminal node. The set of all parents of a variable X (all variables at the tail of an arrow pointing into X) is denoted by pa[Xi]; in Figure 1, pa[D] = {B, C, E}.

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading