A probabilistic graphical model encoding conditional dependencies among variables via directed acyclic graphs.
A Bayesian Network is a probabilistic graphical model that represents a set of variables and their conditional dependencies using a directed acyclic graph (DAG). Each node in the graph corresponds to a random variable — which may be observable, latent, or hypothetical — while each directed edge encodes a conditional dependency between a parent and child node. The strength of these relationships is quantified through conditional probability tables (CPTs) attached to each node, specifying the probability of each variable's states given the states of its parents. Together, the graph structure and CPTs define a compact factorization of the full joint probability distribution over all variables.
Inference in Bayesian Networks involves computing the posterior probability of unobserved variables given evidence about observed ones. Exact inference algorithms such as variable elimination and belief propagation exploit the graph's conditional independence structure to perform these computations efficiently. For large or densely connected networks where exact inference becomes intractable, approximate methods like Markov Chain Monte Carlo (MCMC) sampling or variational inference are employed. Learning a Bayesian Network from data involves two tasks: structure learning, which identifies the DAG topology, and parameter learning, which estimates the CPTs — both of which can be approached from frequentist or Bayesian perspectives.
Bayesian Networks are particularly powerful because they make conditional independence assumptions explicit and interpretable, allowing domain experts to incorporate prior knowledge directly into the model structure. This transparency distinguishes them from many black-box machine learning approaches and makes them well-suited for applications where explainability matters, such as medical diagnosis, fault detection, and risk assessment. They also handle missing data naturally through marginalization, a significant practical advantage.
The framework became foundational to modern probabilistic machine learning and influenced the development of more expressive models, including dynamic Bayesian Networks for temporal data and hierarchical Bayesian models. Concepts central to Bayesian Networks — such as d-separation, Markov blankets, and belief propagation — remain core tools in probabilistic reasoning and underpin many contemporary approaches in causal inference and generative modeling.