Skip to main content icon/video/no-internet

Modeling Causal Learning

Humans display remarkable ability to acquire causal knowledge. Hume's philosophical analysis of causation set the agenda for discovering how causal relations can be inferred from observable data, including temporal order and covariations among events. Recent computational modeling work on causal learning has made extensive use of formalisms based on directed causal graphs (Figure 1). Within a causal graph, each arrow connects a node representing a cause to an effect node, reflecting core assumptions that a cause precedes its effect and has some power to generate or prevent it. The computational goal is to infer from the observable data the unobservable causal structure conveyed by the graph and the magnitude of the power of each cause to influence its effect.

Figure 1 A simple example of a causal graph

None

This entry covers some key issues that arise in modeling human causal learning. Alternative models of causal learning vary depending on the assumptions adopted in the computation, the goal of the computation, and the presentation format of the input data. Understanding models from these perspectives can clarify their commonalities and differences, guide the design of psychological experiments to test the validity of key assumptions, and assess whether models can potentially be extended to real-life problems, such as medical diagnosis and scientific discovery.

Alternative Causal Assumptions

When the causal graph includes multiple potential causes of a single effect, two leading classes of models make different assumptions about the integration rule used to combine causal influences. One class (including the classic delta-P model and the associative Rescorla-Wagner model) assumes a linear integration rule: Each candidate cause changes the probability of the effect by a constant amount regardless of the presence or absence of other causes. A second class is represented by the power PC theory, a theory of causal judgments postulating that learners assume that unobservable causal influences operate independently to produce the effect. Guided by this assumption, causal integration is based on probabilistic versions of various logical operators, such as OR and AND-NOT, chosen to reflect the polarity of causal influence (i.e., whether a cause produces or prevents the occurrence of an effect).

When the causal graph includes multiple effects of a single cause (e.g., flu causes headache and chest pain, as shown in Figure 1), the causal Markov assumption states that the probability of one effect occurring is independent of the probability of other effects occurring, given that its own direct causes are present. Statistical models that examine causal relationships adopt the Markov assumption, which guides exploration of conditional independencies that hold in a body of data, thereby constraining the search space by eliminating highly unlikely cause-effect relations. The extent to which humans employ the causal Markov assumption remains controversial.

When observations are limited, human causal learning relies heavily on some type of prior knowledge. Prior knowledge can be specific to a domain (based on known categories), but it can also include abstract assumptions about properties of a system of cause-effect relations (e.g., preference for causal networks that exhibit various types of simplicity). Use of appropriate prior knowledge can explain the rapid acquisition of causal relations often exhibited by humans.

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading