Understand the foundation of probabilistic graphical models: graph structures, directed vs undirected models, and generative vs discriminative modeling paradigms.
Probabilistic Graphical Models (PGMs) are a class of probability models that use graph structures as core tools to represent and reason about probabilistic relationships between random variables.
Through graph structure, simplify probability computation and achieve inference:
Probabilistic graphical models are classified into two main categories based on the type of graph structure used:
Use Directed Acyclic Graphs (DAG) to represent one-way dependency relationships (e.g., "parent node → child node" causal dependencies).
Key Characteristics:
Typical Examples:
Use undirected graphs to represent bidirectional correlation relationships between variables.
Key Characteristics:
Typical Examples:
Probabilistic graphical models can also be classified by their modeling approach:
Model the joint probability distribution , where:
Derive conditional distribution from the joint distribution.
Examples:
Directly model the conditional probability distribution , focusing on the direct relationship between target variables and observable variables.
Key Advantage:
More flexible and often more accurate for classification tasks, as they don't need to model the full joint distribution of all variables.
Examples:
Generative Models:
Discriminative Models:
Consider a medical diagnosis system using a directed probabilistic graphical model (Bayesian network) to diagnose diseases based on patient symptoms and test results.
| Variable | Type | Possible Values | Parent Nodes |
|---|---|---|---|
| Age | Observable | Young, Middle, Old | None |
| Disease | Hidden | Healthy, Flu, Pneumonia | Age |
| Fever | Observable | No, Yes | Disease |
| Cough | Observable | No, Yes | Disease |
| Test Result | Observable | Negative, Positive | Disease |
Network structure: Age → Disease → {Fever, Cough, Test Result}. This directed graph represents causal relationships: age affects disease probability, and disease causes symptoms and test results.
The joint probability distribution factorizes according to the graph structure:
This factorization simplifies the joint distribution from 5 variables to products of simpler conditional probabilities.
Inference Example:
Given observable variables (Age=Middle, Fever=Yes, Cough=Yes, Test=Positive), we can compute using Bayes' theorem and the graph structure, without enumerating all possible combinations.
A: Directed graphs use arrows to represent causal or conditional dependencies (parent → child), while undirected graphs use lines to represent symmetric correlations. Directed graphs are better for causal modeling, while undirected graphs are better for spatial or symmetric relationships.
A: Use generative models when you need to generate new samples or model the full data distribution. Use discriminative models when you only care about classification accuracy and don't need to generate samples. Discriminative models are often more accurate for classification tasks.
A: By factorizing the joint distribution based on the graph structure, we can break down complex high-dimensional probability calculations into simpler local computations. Instead of computing directly, we compute products of conditional probabilities, which is much more efficient.
A: Yes! There are hybrid models like chain graphs that combine both directed and undirected edges. However, most practical models use either purely directed (Bayesian networks) or purely undirected (Markov networks) structures for simplicity and computational efficiency.