Skip to main content icon/video/no-internet

A Bayesian network is a graphical representation of a multivariate probability distribution on a set of discrete random variables. Representational efficiency is achieved by explicit separation of information about conditional independence relations between the variables (coded in the network structure) and information about the probabilities involved (coded as a set of numeric parameters or functions). The network structure is expressed as a directed acyclic graph (DAG) that makes the representation amenable to an intuitively appealing, causal interpretation. Algorithms exist for learning both network structure and parameters from data. Furthermore, Bayesian networks allow for computing any marginal or conditional probability regarding the variables involved, thus offering a powerful framework for reasoning with uncertainty. Bayesian networks are also called belief networks and causal probabilistic networks.

Figure 2 A Mixed Treatment Comparison network involving six thrombolytic treatments following acute myocardial infarction and one surgical treatment, percutaneous transluminal angioplasty (PCTA)

None
Note: Each edge indicates that the treatments have been compared in at least one randomized, controlled trial.

Bayesian networks are suited to model the uncertainty that inheres in many biomedical domains and are, therefore, frequently used in applications of computer-assisted decision making in biomedicine. Furthermore, extensions of Bayesian networks (called influence diagrams) can be used to perform decision analyses.

This entry first sketches the historical background of Bayesian networks. Subsequently, it elaborates on model structure, approaches for network construction, inference methods, medical applications, and software.

Historical Background

Bayesian networks originated in the mid-1980s from the quest for mathematically sound and computationally tractable methods for reasoning with uncertainty in artificial intelligence. In the preceding decade, the first applications of computer-assisted decision making had found their way to the medical field, mostly focusing on the diagnostic process. This had required the development of methods for reasoning with uncertain and incomplete diagnostic information.

One popular method was the naive Bayesian approach that required specification of positive and negative predictive values for each of a set of predefined diagnostic tests and a prior (i.e., marginal) probability distribution over possible diagnostic hypotheses. The approach assumed that all test results were mutually independent markers of disease and used Bayes's theorem to compute posterior (i.e., conditional) probabilities on the hypotheses of interest. The approach is simple and fast and requires a relatively small number of marginal and conditional probabilities to be specified. However, the assumption of independence is mostly wrong and leads to overly extreme posterior probabilities.

Another approach arose in the field of expert systems, where algorithms had been devised to reason with so-called certainty factors, parameters expressing the strength of association in if-then rules. The underlying reasoning principles were mostly ad hoc and not rooted in probability theory, but large sets of if-then rules allowed for a domain representation that was structurally richer and more complex than naive Bayesian models. Bayesian networks bring together the best of both approaches by combining representational expressiveness with mathematical rigor.

Model Structure

Bayesian networks belong to the family of probabilistic graphical models (PGMs), graphs in which nodes represent random variables, and the (lack of) arcs represent conditional independence assumptions. Let G = (V(G), A(G)) be a directed acyclic graph, where the nodes V(G) = {V1, …, Vn} represent discrete random variables with a finite value domain. For each node ViV(G), let πi denote the set of parent nodes of Vi in graph G. A Bayesian network now is a pair B = (G, Θ), where Θ={θi|ViV(G)} is a set of parametrization functions. The function θi describes a local model for node ViV(G) by specifying a conditional probability θi(v|s) for each possible value v of variable Vi and all possible value assignments s to its parents πi. The Bayesian network B defines a unique multivariate probability distribution Pr on V1, …, Vn using the

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading