This class is an introductory undergraduate course in machine learning. The class will briefly cover topics in regression, classification, mixture models, neural networks, deep learning, ensemble methods and reinforcement learning.

Prerequisites: You should understand basic probability and statistics, (STA 107, 250), and college-level algebra and calculus. For example it is expected that you know about standard probability distributions (Gaussians, Poisson), and also how to calculate derivatives. Knowledge of linear algebra is also expected, and knowledge of mathematics underlying probability models (STA 255, 261) will be useful. For the programming assignments, you should have some background in programming (CSC 270), and it would be helpful if you know Python.

• August 25: Webpage created.

Lectures: Monday 11-1pm (section 1), Wednesday 11-1pm (section 2), Thursday 4-6pm (section 3), Friday 11-1pm (section 4)

Lecture Room: AH 400 (section 1), MS 2170 (section 2), KP 108 (section 3), MS 2172 (section 4)

Instructor: Raquel Urtasun (sections 1 and 2), Richard Zemel (sections 3 and 4)

Office hours: Monday 4-5; Wednesday 4-5; Thursday 2-3; Pratt 290C

Tutorials: Monday 3-4pm (section 1), Wednesday 3-4pm (section 2), Thursday 6-7pm (section 3), Friday 3-4pm (section 4)

Tutorial Room: AH 400 (section 1), BA 1170 (section 2), KP 108 (section 3), MS 2172 (section 4)

The format of the class will be lectures, with some discussion. We strongly encourage interaction and questions. There are assigned readings for each lecture that are intended to prepare you to participate in the class discussion for that day.

The final grade will consist of the following
`Assignments` 55% (3 assignments, first two worth 15%, last 25%)
`Mid-Term Exam` 20%
`Final Exam` 25%

### Homework assignments

The best way to learn about a machine learning method is to program it yourself and experiment with it. So the assignments will generally involve implementing machine learning algorithms, and experimentation to test your algorithms on some data. You will be asked to summarize your work, and analyze the results, in brief (3-4 page) write ups. The implementations may be done in any language, but Matlab and Python are reocmmended. A brief tutorial on Matlab is included here. You may also use Octave.

Collaboration on the assignments is not allowed. Each student is responsible for his or her own work. Discussion of assignments and programs should be limited to clarification of the handout itself, and should not involve any sharing of pseudocode or code or simulation results. Violation of this policy is grounds for a semester grade of F, in accordance with university regulations.

The schedule of assignments is included in the syllabus. Assignments are due at the beginning of class/tutorial on the due date. Because they may be discussed in class that day, it is important that you have completed them by that day. Assignments handed in late but before 5 pm of that day will be penalized by 5% (i.e., total points multiplied by 0.95); a late penalty of 10% per day will be assessed thereafter. Extensions will be granted only in special situations, and you will need a Student Medical Certificate or a written request approved by the instructor at least one week before the due date.

### Exams

There will be a mid-term in class during the 6th week of classes, which will be a closed book exam on all material covered up to that point in the lectures, tutorials, required readings, and assignments.

The final will not be cumulative, except insofar as concepts from the first half of the semester are essential for understanding the later material.

### Attendance

We expect students to attend all classes, and all tutorials. This is especially important because we will cover material in class that is not included in the textbook. Also, the tutorials will not only be for review and answering questions, but new material will also be covered.

There is no required textbook for this course. There are several recommended books. We will recommend specific chapters from Pattern Recognition and Machine Learning by Chris Bishop. We will also recommend other readings.

Click on the syllabus

#### Schedule (tentative)

Lecture 1Course Introduction lecture1 (handout)
lecture1 (animated)
Lecture 2Linear Regression Bishop, Chapter 1.0-1.1; 3.1 lecture2 (handout)
lecture2 (animated)
Tutorial 1Tutorial: Review on Probability tutorial1
Lecture 3Linear Classification Bishop: Pages 179-195 lecture3
Lecture 4Logistic Regression Bishop: Pages 203-207 lecture4
Lecture 5Non-parametric Methods Bishop: Pages 120-127 lecture5
Lecture 6Decision Trees lecture6
Tutorial 3Tutorial: K-NN and Decision Trees tutorial3
Lecture 7Multi-class Classification Bishop 4.1.2, 4.3.4 lecture7
Lecture 8Probabilistic Classifiers I Bishop 4.2.2 lecture8
Lecture 9Probabilistic Classifiers II Bishop, 380-381 lecture9
Tutorial 4Tutorial: Naive Bayes tutorial4
Lecture 10Neural Networks I Bishop 5.1 - 5.3 lecture10
Lecture 11Neural Networks II lecture11
Tutorial 5Tutorial: Neural Networks tutorial5
Tutorial 6Tutorial: Midterm Review
Lecture 12Clustering Bishop 9.1 lecture12
Lecture 13Mixture of Gaussians Bishop 9.2, 9.3 lecture13
Tutorial 6Tutorial: Clustering tutorial6
Lecture 14PCA & Autoencoders Bishop 12.1 lecture14
Lecture 15SVM Bishop: Chapter 7, pages 325-337 lecture15
Tutorial 7Tutorial: PCA tutorial7
Lecture 16Kernels lecture16
Lecture 17Ensemble Methods I Bishop 14.2 - 14.3 lecture17
Tutorial 8Tutorial: SVM tutorial8
Lecture 18Ensemble Methods II lecture18
Lecture 19Reinforcement Learning lecture19
Tutorial 9Tutorial: RL tutorial9