ml:graphical_models
Table of Contents
Graphical Models
Graphical models (or probabilistic graphical models, PGMs) are sub-area of machine learning and statistics. PGMs are a framework for representing independence assumptions of random variables in probability distributions. Broadly, the study of PGMs includes the study of algorithms for learning and inference for these complex probability distributions. PGMs have applications in machine learning, statistics, natural language processing, speech recognition, computer vision, robotics, and other areas. Topics include Bayesian Networks, Hidden Markov Models (HMMs), Conditional Random Fields (CRFs), Markov Random Fields (MRFs), Variational Inference and Bayesian nonparametrics.
Overviews
- Graphical Models
- Koller et al 2007 - Graphical Models in a Nutshell (book chapter)
- Deep Latent Variable Models
Models
- Bayesian Networks
- Markov Random Fields
- Factor Graphs
Inference
- Belief Propagation
- Bert Huang's Video Talks about relation of BP and Lagrangian relaxation at the end.
- Markov Chain Monte-Carlo (MCMC)
- Variational Inference
- Great description here: (see section 2.2) Kucukelbir 2016
- Great video: Blei - Variational Inference: Foundations and Innovations (nice overview at ~10:00)
Old Papers
- Kschischang et al 1998 - Factor Graphs and The Sum-Product Algorithm The paper that introduced factor graphs
Recent Papers
- Chiang & Riley 2020 - Factor Graph Grammars Introduces a new kind of graphical model (factor graph grammars) that are more expressive than plate notation or dynamic graphical models. It is expressive enough to represent CFG parsing a graphical model. Very cool.
Interesting NLP Deep Learning + PGM Papers
See also recent advances in HMMs and CRFs.
- Wang et al 2019 - Second-Order Semantic Dependency Parsing with End-to-End Neural Networks Uses loopy BP and variational inference
- Wang & Cho 2019 - BERT has a Mouth, and It Must Speak:BERT as a Markov Random Field Language Model WARNING: Mistake in this paper, it's not an MRF
- Gao & Gormley 2020 - Training for Gibbs Sampling on Conditional Random Fields with Neural Scoring Factors (basically adapted Finkel et al 2005 to the neural era)
- Lee et al 2020 - On the Discrepancy between Density Estimation and Sequence Generation Uses latent variables for fast non-autoregressive generation
- Yang et al 2021 - Neural Bi-Lexicalized PCFG Induction Uses a Bayesian network to describe their model
Recent NLP Papers that Use PGMs
- Belief Propagation
- MCMC and Sampling
- Gao & Gormley 2020 - Training for Gibbs Sampling on Conditional Random Fields with Neural Scoring Factors (basically adapted Finkel et al 2005 to the neural era)
- Variational Inference
- Other Papers
Courses, Tutorials, and Overview Papers
- Sometimes PGMS are covered in the UCSC course CSE 290C (when Lisa Geetoor teaches it)
- Course at CMU: Probabilistic Graphical Models Spring 2020 Lectures with videos 2014 (with videos and scribe notes)
- Stanford course: CS 228 - Probabilistic Graphical Models
- Matt Gormley's course at CMU: 10418 (with videos)
- Best overview tutorial: CS228 Lecture Notes
Related Pages
- Bayesian Methods Bayesian methods often use techniques from graphical models, such as MCMC and variational inference, as well as representing likelihood and prior as a graphical model
ml/graphical_models.txt · Last modified: 2025/05/03 01:43 by jmflanig