Probabilistic Graphical Models and Structured Prediction
CS6424/ECE6424 Spring 2016

Announcements

Description

In this course, we will study the wide variety of research on probabilistic graphical models and structured prediction. These areas of machine learning research cover algorithmic approaches to learning and prediction in settings with dependencies in data, variables, and outputs. We will examine the currently understood methods for efficiently managing the complexities that arise when we consider these dependencies in models, study algorithms for reasoning and predicting using these models, and explore how other algorithms can learn parameters and structure from data.

The class will be a shared learning experience, where the students will lead discussions as we work through some seminal texts and recent papers.

Class meets Tuesday and Thursday from 3:30 PM to 4:45 PM in McBryde 232.

The course homepage is http://people.cs.vt.edu/~bhuang/courses/pgmsp16/ or equivalently http://berthuang.com/courses/pgmsp16/.

Topics and Goals

Topics include directed models (Bayes nets), undirected models (Markov/conditional random fields), exact inference (junction tree), approximate inference (belief propagation, dual decomposition), parameter learning (MLE, MAP, EM, max-margin), and structure learning.

Having successfully completed this course, the student will be able to:

Prerequisites

The listed prerequisite courses cover relevant material that includes introductory-level knowledge of machine learning and statistical learning (likelihood, maximum likelihood, bias and variance, underfitting, overfitting, regularization, cross-validation), probability and statistics, and algorithms.

Please speak with the instructor if you are concerned about your background. Note: If any student needs special accommodations because of any disabilities, please contact the instructor during the first week of classes.

Reading and Materials

This class will be taught using a combination of materials. The main two texts we will use are seminal texts in the field of machine learning. One is a textbook, and the other is a long journal paper, which is available for free.

Grading Breakdown

Format and Attendance

After the first two weeks of the semester, the class will center around discussion. We will work through the reading materials together, and you will lead discussion and participate in panels during class sessions. In each class session, you will have one of three roles: panel leader, panelist, or audience.

The goal of each class is to understand the ideas presented in the assigned reading as much as is possible. We will work through examples, work together to clarify points of confusion, and discuss the wider implications of the topics from the reading. To make sure this experience works for everyone, class attendance is mandatory. I will take attendance and penalize unexused absences at my discretion.

In the first week of class, you will rank dates for when to serve as panel leaders or panelists, and I will work out a schedule that satisfies your preferences as much as possible. Keep in mind that since every student must do the reading, serving as a panelist or leader is not much more work on top of the reading.

Summaries and Questions

For each assigned reading, you will submit a writeup in your own words that contains two major components: a summary and at least two questions.

The summary should be 1-2 paragraphs describing what main ideas were presented in the text. You should cover the takeaway points, not the fine details.

You will then write (at least) two questions. The first will be a clarification question. What idea in the text was unclear to you? What do you want to understand better? For example, you might write, "Why is Lemma 3, which is presented without proof, guaranteed to be true?" In theory, clarification questions should have answers that we can attempt to find together as a class.

The second question you will include is a discussion question. You might ask about research ideas, how the technique we studied can be applied, or why a mathematical concept is important. In theory, a discussion question may not have a true answer, but we can speculate and brainstorm about them during our discussion.

You should submit your summary and questions on Canvas as posts in the discussion thread for the class session. These threads are set so you cannot see other posts until you post your own summary and questions.

Project

For the class project, you will conduct research with the goal of producing findings worthy of publication at a conference. The project should be done in groups of 1–3 students, and should feature a novel algorithmic contribution to some aspect of learning, inference, prediction, analysis, or application using probabilistic graphical models or structured predictors. You are strongly encouraged to incorporate this project with any other research you are working on, and you are welcome to include collaborators outside the class as long as you are doing a substantial proportion of the research yourself.

You will first write a 1-page proposal on your project before Spring Break. You are highly encouraged to begin brainstorming about topics as soon as possible. You will write a 6–10 page paper on your findings due at the end of the semester, reporting your contribution, background material, evaluation (experiments and/or analysis), and conclusions. Details on the proposal and paper details are available at ./project_deliverables.html.

Here is a list of brainstormed project ideas to help you design your own project. See additional project ideas from similar courses, such as from CMU, UPenn, and the projects from the previous PGM course at VT for more ideas.

External Resources

Various related courses have valuable materials that can help supplement our official reading. Some of these are linked below.

Academic Integrity

The tenets of the Virginia Tech Graduate Honor Code will be strictly enforced in this course, and all assignments shall be subject to the stipulations of the Graduate Honor Code. For more information on the Graduate Honor Code, please refer to the GHS Constitution at http://ghs.graduateschool.vt.edu.

This course will have a zero-tolerance policy regarding plagiarism or other forms of cheating. Your homework assignments must be your own work, and any external source of code, ideas, or language must be cited to give credit to the original source. I will not hesitate to report incidents of academic dishonesty to the graduate school.

Schedule

Session Date Topic Reading Panelists (Leaders in bold)
1 Tue 1/19 Overview of Graphical Models and Structured Prediction KF 1, 2
Why ML Needs Statistics by Welling (link)
None
2 Thu 1/21 Bayesian Networks KF 3-3.2 None
3 Tue 1/26 Bayesian Networks KF 3.3-3.5 None
4 Thu 1/28 Undirected Graphical Models KF 4-4.4 None
5 Tue 2/2 Undirected Graphical Models KF 4.5-4.7 Tianyi Li, Jason Louis Granstedt, William Ryan Doan
6 Thu 2/4 Exact Inference: Variable Elimination KF 9-9.4, 9.7 Sirui Yao, Abhilash Chowdhary, Aishwarya Agrawal, Arjun Chandrasekaran
7 Tue 2/9 Exact Inference: Clique Trees KF 10 Arijit Ray, Aishwarya Agrawal, Arjun Chandrasekaran
8 Thu 2/11 Inference as Optimization KF 11-11.3 Bert Huang, Aishwarya Agrawal, Viresh Ranjan, Xuan Zhang
9 Tue 2/16 Inference as Optimization KF 11.4-11.6 Bert Huang
10 Thu 2/18 MAP Inference KF 13-13.5 Yi Tian
11 Tue 2/23 Parameter Learning in Bayesian Networks KF 17 Abhilash Chowdhary
12 Thu 2/25 Structure Learning in Bayesian Networks KF 18 Bijaya Adhikari
13 Tue 3/1 Partially Observed Data KF 19.1-19.2 Jiali Lin
14 Thu 3/3 Learning Undirected Models KF 20 Bert Huang
Spring Break
15 Tue 3/15 Graphical Models as Exponential Families WJ 3 Elaheh Raisi
16 Thu 3/17 Sum-Product, Bethe-Kikuchi, Expectation-Propagation WJ 4 Aroma Mahendru
17 Tue 3/22 Variational Methods in Parameter Estimation WJ 6 Viresh Ranjan
18 Thu 3/24 Convex Relaxations and Upper Bounds WJ 7 Bert Huang
19 Tue 3/29 Dual Decomposition for Inference Introduction to Dual Decomposition for Inference by Sontag et al. (link) Arjun Chandrasekaran
20 Thu 3/31 Structured SVM Cutting Plane Training of Structural SVMs by Joachims et al. (link) Jinwoo Choi
21 Tue 4/5 Max-Margin Markov Networks and Inner-Dual Learning Max-Margin Markov Networks by Taskar et al. (link)
Learning Structured Prediction Models: A Large Margin Approach by Taskar et al. (link)
Shuangfei Fan
22 Thu 4/7 Inner-dual Learning with Dual Decomposition Learning Efficiently with Approximate Inference via Dual Losses by Meshi et al. (link)
A Primal-Dual Message-Passing Algorithm for Approximated Large Scale Structured Prediction by Hazan and Urtasun (link)
Bert Huang
23 Tue 4/12 Markov Logic Networks Markov Logic Networks by Richardson and Domingos (link) Chris Dusold
24 Thu 4/14 Hinge-Loss MRFs and Probabilistic Soft Logic Hinge-Loss Markov Random Fields and Probabilistic Soft Logic by Bach et al. (link). Read Sections 1, 2, 3, 5, 6-6.3. Optionally skim the rest. Walid Chaabene
25 Tue 4/19 Latent-Variable Structured SVM Learning Structural SVMs with Latent Variables by Yu et al. (link)
Marginal Structured SVM with Hidden Variables by Ping et al. (link)
Bert Huang
26 Thu 4/21 Latent-Variable Inner-Dual methods Efficient Structured Prediction with Latent Variables for General Graphical Models by Schwing et al. (link)
Paired-Dual Learning for Fast Training of Latent Variable Hinge-Loss MRFs by Bach et al. (link)
Subhodip Biswas
27 Tue 4/26 Learning Deep Structured Models Learning Deep Structured Models by Chen et al. (link) William Ryan Doan
28 Thu 4/28 Causality KF 21 Jason Louis Granstedt
29 Tue 5/3 Wrap Up Bert Huang

Disclaimer: This syllabus details the plans for the course, which are subject to change. I will make sure any changes are clearly announced and will always be intended for your benefit.

For visitors outside the course: You are welcome to use the course materials for educational purposes. Do not sell any of this content.