Welcome to T3-2022!

We will have a mixed course, online and face-to-face. All lectures will be offered online in the form of topic-based videos, which you can watch in your own time, and meeting with the lecturer to review the content and solve exercises. Tutorials will be held in the timetabled time slots. We will have face-to-face and online tutorials through MS Teams.

Assignments and quizzes will be online. The Final Exam will be held online.

Course Timetable

Lectures

COMP9418/1UGA
COMP9418/1PGA
Week 1-5,7-10

Pre-recorded lectures on
Youtube and Echo360
Online
Gustavo Batista

COMP9418/1UGA
COMP9418/1PGA
Week 1-5,7-10 Mon 14:00 - 15:00
Thu 14:00 - 15:00
MS Teams
MS Teams
Gustavo Batista

Tutorials

M16A mon16a Week 1-5,7-10 Mon 16:00 - 18:00 Mathews 312 Saurav Jha
M16B mon16b Week 1-5,7-10 Mon 16:00 - 18:00 MS Teams
Payal Bawa
T11A tue11a Week 1-5,7-10 Tue 11:00 - 13:00 Mathews 307 Madeleine Nouri
H11A thu11a Week 1-5,7-10 Thu 11:00 - 13:00 Mathews 307 Madeleine Nouri
H16A thu16a Week 1-5,7-10 Thu 16:00 - 18:00 Quadrangle G301 Saurav Jha
H16B thu16b Week 1-5,7-10 Thu 16:00 - 18:00 MS Teams
Payal Bawa

Overview

In this course, we will study a class of statistical inference models known as Probabilistic Graphical Models (PGMs). PGMs are a great example of how Computer Science and Statistics can work together. PGMs use graph data structures to represent domains with large amounts of variables and specialised algorithms for efficient inference over these graphical models. Therefore, PGMs have pushed the limits of probability theory to the scale and rate necessary to provide automated reasoning in modern AI systems.

During this course, we will cover several graphical models, including Bayesian networks, Markov networks, Conditional Random Fields, Markov chains, Hidden Markov Models, Kalman Filters and Markov decision processes. We will have a clear understanding of how these models work as well as their main algorithms for inference and learning. We will also cover several algorithms used to learn parameters and make inferences such as Monte Carlo Markov Chains (MCMC), Gibbs Sampling, Viterbi and the Baum-Welch algorithms, among others.

Contents

Course Details

Course Code COMP9418
Course Title Advanced Topics in Statistical Machine Learning
Convenor Gustavo Batista
Admin Maryam Hashemi
Consultations Mondays 15:00-16:00, MS Teams
Units of Credit 6
Course Website http://cse.unsw.edu.au/~cs9418/22T3/
Handbook Entry http://www.handbook.unsw.edu.au/postgraduate/courses/current/COMP9418.html
Student Reps stureps@cse.unsw.edu.au
Email the stureps if you have any issues with the course.
They will pass these anonymously to the relevant people to get the issues resolved.

Course Summary

This course presents an in-depth study of statistical machine learning approaches. It aims to provide the student with a solid understanding of methods for learning and inference in structured probabilistic models, with a healthy balance of theory and practice. It will cover topics on the semantics of direct and undirected representations in probabilistic graphical models, exact and approximate inference, and learning of model parameters and structure.

Assumed Knowledge

Official Pre-requisites

Knowledge of machine learning at the level of COMP9417.

Programming

We will use Python in the practical part of our bring-your-own-device tutorials. Students will have to install Jupyter Notebook on their computers to execute the practical part of the tutorials. Alternatively, they can remotely access the VLAB computers using a VPN (if outside UNSW) and software such as TigerVNC . However, it is expected that students unfamiliar with Python can get up to speed if they are able to construct (i.e., design, implement and test) working software in a general-purpose language such as C/C++ or Java, at least to the level of a first-year computing course (e.g., COMP1927 Computing 2 or equivalent). Having a good working knowledge of and be able to construct working software in standard data analysis languages such as Matlab/Octave, or R can also be helpful.

Software Tools

Since an important part of practical machine learning is "data wrangling" (i.e. pre-processing, filtering, cleaning, etc.) of data files, students are expected to master Unix tools such as those taught in COMP2041 Software Construction, or equivalents such as those in Matlab/Octave or R.

Student Learning Outcomes

After completing this course, students will:

  1. Derive statistical independence assumptions from a given graphical representation of a probabilistic model
  2. Understand and implement exact inference methods in graphical models including variable elimination and the junction tree algorithm
  3. Derive and implement maximum likelihood learning approaches to latent variable probabilistic models
  4. Understand and implement approximate inference algorithms in graphical models, including sampling and loopy belief propagation.
  5. Understand and apply basic methods for structured prediction

This course contributes to the development of the following graduate capabilities:

Graduate Capability Acquired in
Scholars capable of independent and collaborative enquiry, rigorous in their analysis, critique and reflection, and able to innovate by applying their knowledge and skills to the solution of novel as well as routine problems Lectures, tutorials, assignments and exam
Entrepreneurial leaders capable of initiating and embracing innovation and change, as well as engaging and enabling others to contribute to change Assignments
Professionals capable of ethical, self-directed practice and independent lifelong learning Lectures, tutorials and assignments
Global citizens who are culturally adept and capable of respecting diversity and acting in a socially just and responsible way Lectures, tutorials, assignments and student interactions

Teaching Strategies

Machine learning is at the intersection of Artificial Intelligence, Computer Science and Statistics. While the main goal of this course is to go beyond the basics of machine learning as provided by COMP9417 (focused on probabilistic modelling and inference), we will adopt a similar teaching rationale, where theory, algorithms and empirical analysis are all important components of the course. Therefore, the lectures, tutorials and assessments are designed to address these components jointly.

The course involves lectures and practical work.

  • Lectures: Aim to summarise the concepts and present case studies.
  • Tutorials: Aim to reinforce the topics covered in lectures and will cover theoretical and practical exercises. The practical part of the tutorials will be based on a bring-your-own-device approach, where students will be introduced to the technology required for the assignments and follow a series of programming and data analysis questions. There will be no formal assessment of the tutorials.
  • Assignments: Aim the same as the tutorials at a higher degree of difficulty and will be assessed.
  • Final exam: There will be a final exam.

Engagement Tools and Blended Learning

  • All lectures (slides/recordings) will be on the Web.
  • All tutorial and lab materials (questions before, solutions after) will be on the Web.
  • All assignments will have specifications on the Web and online submission.
  • The final exam will likely be online.
  • Forum for answering questions using WebCMS3

Student Conduct

The Student Code of Conduct ( Information , Policy ) sets out what the University expects from students as members of the UNSW community. As well as the learning, teaching and research environment, the University aims to provide an environment that enables students to achieve their full potential and to provide an experience consistent with the University's values and guiding principles. A condition of enrolment is that students inform themselves of the University's rules and policies affecting them, and conduct themselves accordingly.

In particular, students have the responsibility to observe standards of equity and respect in dealing with every member of the University community. This applies to all activities on UNSW premises and all external activities related to study and research. This includes behaviour in person as well as behaviour on social media, for example, Facebook groups set up for the purpose of discussing UNSW courses or coursework. Behaviour that is considered in breach of the Student Code Policy as discriminatory, sexually inappropriate, bullying, harassing, invading another's privacy or causing any person to fear for their personal safety is serious misconduct and can lead to severe penalties, including suspension or exclusion from UNSW.

If you have any concerns, you may raise them with your lecturer, or approach the School Ethics Officer , Grievance Officer , or one of the student representatives.

Plagiarism is defined as using the words or ideas of others and presenting them as your own. UNSW and CSE treat plagiarism as academic misconduct, which means that it carries penalties as severe as being excluded from further study at UNSW. There are several on-line sources to help you understand what plagiarism is and how it is dealt with at UNSW:

Make sure that you read and understand this. Ignorance is not accepted as an excuse for plagiarism. In particular, you are also responsible that your assignment files are not accessible by anyone but you by setting the correct permissions in your CSE directory and code repository, if using. Note also that plagiarism includes paying or asking another person to do a piece of work for you and then submitting it as your own work.

UNSW has an ongoing commitment to fostering a culture of learning informed by academic integrity. All UNSW staff and students have a responsibility to adhere to this principle of academic integrity. Plagiarism undermines academic integrity and is not tolerated at UNSW. Plagiarism at UNSW is defined as using the words or ideas of others and passing them off as your own.

If you haven't done so yet, please take the time to read the full text of

The pages below describe the policies and procedures in more detail:

You should also read the following page which describes your rights and responsibilities in the CSE context:

Please note due dates are subject to change.

Item Topics Due Marks Contributes to
(learning outcomes)
Assignment 1
Weeks 1-2 Week 5 15% 1-2
Assignment 2
Weeks 3-7
Week 9 15%
3-4
Quizzes Weeks 2-5,7-10 Weeks 2,3,4,5,7,8,9,10 10% 1-5
Final Exam All topics Exam period 60% 1-5

Quizzes are multiple-choice questions used to check your understanding during the course. The final mark for quizzes is the simple average of the eight quiz marks. Each quiz mark is normalised in the range 0-10.

There is a hurdle on the Final Exam; very poor performance in the exam will result in a fail, even if all your other assessment marks have been satisfactory. The following formula describes precisely how the mark will be computed and how the hurdle will be enforced.

quizzes   = mark for quizzes (out of 10) 
ass1      = mark for assignment 1 (out of 15) 
ass2      = mark for assignment 2 (out of 15) 
exam      = mark for exam (out of 60)
okExam    = finalExam >= 24/60 
mark      = quizzes + ass1 + ass2 + exam 
grade     = HD|DN|CR|PS if mark >= 50 && okExam 
          = FL if mark < 50 
          = UF if mark >= 50 && !okExam

Course Schedule

Please note this is a tentative schedule. All dates are only indicative and subject to change.

Week Lecture Tutorial Assignment Quizzes
1 Course overview [Ch. 1], propositional logic [Ch. 2] and probability calculus [Ch. 3] Graph representation, traversal and common algorithms - -
2 Bayesian networks representation and semantics [Chs. 4 and 5] Probability calculus and factor implementation - Quiz 1
3 Exact inference [Ch. 6]. Bayesian networks as classifiers Bayesian networks Ass1 released Quiz 2
4 Markov chains and hidden Markov models Variable elimination - Quiz 3
5 MAP inference [Ch. 10]. Markov networks Markov chains and hidden Markov models Ass1 due Quiz 4
6 Flexibility Week - - -
7 The jointree algorithm [Chs. 7 and 9] Markov networks Ass2 released Quiz 5
8 Gaussian Bayesian Networks [Koller Ch. 7, 14.1 & 14.2]
Factor elimination and jointrees - Quiz 6
9 Belief propagation [Ch. 14]. Approximate inference by Sampling [Ch. 15]
Gaussian Bayesian networks - Quiz 7
10 Learning parameters and graph structure [Ch. 17] Belief propagation and sampling Ass2 released Quiz 8

Resources for Students

This website also has links to the auxiliary material/documentation that you will need for the course. Solutions for all tutorial questions and exercises will also be made available.

Prescribed Book

Recommended Books

Other resources

Course Evaluation and Development

This course is evaluated using the myExperience system.

In the previous offering of this course, students suggested some changes in the content sequence and the addition of new material covering continuous distributions. In conversation with the students, we also noted that the tutorial code needed to be faster to support their assessment implementations.

Based on their comments, we have placed the MAP lecture earlier in the course. We also reduced the content of this lecture to allow space for a new lecture covering Gaussian Bayesian networks. We reimplemented the tutorial code replacing an unordered dictionary with a NumPy array to increase code efficiency. We also improved code organisation using an objected-oriented implementation.

We thank all the students that provided feedback to this course through MyExperience, email and conversations. These students include Martin Eftimoski, Gareth Dando, and many others.

Resource created Saturday 27 August 2022, 01:09:46 PM, last modified Tuesday 27 September 2022, 03:14:04 PM.


Back to top

COMP9418 22T3 (Advanced Machine Learning) is powered by WebCMS3
CRICOS Provider No. 00098G