Dr. Rina Dechter - University of California at Irvine ZOT!
home | publications | book | courses | research Revised on Mar. 11, 2017

CompSci-276 Spring 2017, Reasoning in Probabilistic Graphical Models: Belief Networks
[ main | software | references ]


Course Reference

    Lecture
    Days: Monday/Wednesday
    Time: 11:00 am - 12:20 pm
    Room: DBH 1427

    Instructor: Rina Dechter
    Office hours: TBD

Course Description

One of the main challenges in building intelligent systems is the ability to reason under uncertainty, and one of the most successful approaches for dealing with this challenge is based on the framework of Bayesian networks, also called graphical models. Intelligent systems based on Bayesian networks are being used in a variety of real-world applications including diagnosis, sensor fusion, on-line help systems, credit assessment, bioinformatics and data mining.

The objective of this class is to provide an in-depth exposition of knowledge representation and reasoning under uncertainty using the framework of Bayesian networks.  Both theoretical underpinnings and practical considerations will be covered, with a special emphasis on dependency and independency models, on construction Bayesian graphical models and on exact and approximate probabilistic reasoning algorithms. Additional topics include: causal networks, learning Bayesian network parameters from data and dynamic Bayesian networks.

Prerequisites

  • Familiarity with basic concepts of probability theory.
  • Knowledge of basic computer science, algorithms and programming principles.
  • Previous exposure to AI is desirable but not essential.

Syllabus

Week       Date Topic Readings           Files        
Week 1 4/3
  • Introduction and Background.
(a) Pearl 1-2
(b) Darwiche 1-3
(c) Russell-Norvig 13
(d) Darwiche.
Bayesian Networks
  4/5
  • Markov networks: undirected graphical Models.
Week 2 4/10
  • Bayesian networks: directed graphical models.
  4/12
  • Bayesian networks: directed graphical models of independence.
Pearl Ch.3
Week 3 4/17
  • Building Bayesian networks.
Darwiche Ch. 5
  4/19
  • Building Bayesian networks.

Week 4 4/24
  • Exact inference by variable elimination.
Dechter Ch. 4,
Darwiche Ch. 6

4/26
  • Optimization queries: MPE and MAP.

Week 5 5/1
  • Exact inference by Tree-decompositions:
    Join-tree/Junction-tree algorithm. Cluster tree elimination.
Dechter Ch. 5,
Darwiche Ch. 7-8

5/3
  • Exact inference by tree-decomposition, cutset-conditioning scheme.

Week 6 5/8
  • AND/OR search spaces.

5/10
  • AND/OR search spaces.
Week 7 5/15
  • Approximate algorithms by Sampling: MCMC schemes.


5/17
  • Approximate algorithms by Sampling: advanced schemes.
Week 8 5/22
  • Approximate algorithms by Bounded Inference.

5/24
  • Approximate algorithms by Bounded Inference (continued).

Week 9 5/29
  • Approximate algorithms by Bounded Inference (continued).

5/31
  • Approximate algorithms by Bounded Inference (continued).

Week 10 6/5
  • Project presentations.



6/7
  • Project presentations.


Week 11 6/12
  • Project presentations.



Assignments:

There will be homework assignments and students will also be engaged in projects.

Grading Policy:

Homework and exam (75%), class project (25%)