## Stat 231 / CS 276A Pattern Recognition and Machine Learning

### Course Description

This course introduces fundamental concepts, theories, and algorithms for pattern recognition and machine learning,
which are used in computer vision, speech recognition, data mining, statistics, information retrieval, and bioinformatics.
Topics include: Bayesian decision theory, parametric and non-parametric learning, data clustering, component analysis,
boosting techniques, support vector machine, and deep learning with neural networks.

### Prerequisites

You don't have to take exactly these courses as long as you know the materials.

• Math 33A Linear Algebra and Its Applications, Matrix Analysis.
• Stat 100B Intro to Mathematical Statistics.
• CS 180 Intro to Algorithms and Complexity.
• Programming skills in Matlab or Python.

### Textbook

Textbook is not mandatory if you can understand the lecture notes and handouts.

• R. Duda, et al., Pattern Classification, John Wiley & Sons, 2001. [Good for CS students]
• T. Hastie, et al.,The Elements of Statistical Learning, Spinger, 2009. [Good for Stat students]
• C. Bishop, Pattern Recognition and Machine Learning, Springer, 2006. [with advanced materials]

### Instructors

• Prof. Song-Chun Zhu, sczhu@stat.ucla.edu, Office: Boelter Hall 9404
Office Hours: Tuesday 2-3pm
• TAs for STAT 231A: Ruiqi Gao ruiqigao@ucla.edu, Erik Nijkamp, enijkamp@ucla.edu.
TAs for CS276A: Luyao Yuan, yuanluxu@cs.ucla.edu
Office Hours: Session 1: Tuesday 6-7pm and Session 2: 7-8pm for project overview and assistant.

 Five projects: 0. Exercise on Classification using Deep Learning and CNN Experience the training and testing of problems in machine learning 1. Face modeling by AAM Model, Auto-encoder and FLD How many bits do you need to represent a face? Or telling apart male from female faces? 2. Face detection by AdaBoost and RealBoost How many features do you need to detect a face in a crowd? 3. Object detection by Fast-RCNN How to detect the class of an object and its location? 4. Face social attributes and sentiment analysis by SVM How do we measure the social dimensions of faces in political elections and social network? 10% 15% 15% 10% 10% 40%

• You are encouraged to work and discuss in a group, but each person must finish his/her own project. Submit your report of the project, and your code through the CCLE website.
• You have a total of 5 late days (not including weekends) for the entire class (5 projects) to cover your various reasons, but after using the four late days, no credit will be given for late homework/project.

### Schedule

Tentative Schedule for 2018 (Course materials will be posted on the CCLE site.)

 Lecture Topics Handouts 1 Introduction to Pattern Recognition [Problems, applications, examples] Lecture1 pdf. 2 Bayesian Decision Theory I [Bayes rule, discriminant functions] Lecture2 pdf 3 Bayesian Decision Theory II [loss functions and Bayesian error analysis] 4 Component Analysis and Dimension Reduction I: [PCA, face modeling by Active Appearance Model (AAM), Auto-encoder] 5 Component Analysis and Dimension Reduction II [Fisher Linear Discriminant ] [Multi-dimensional scaling (MDS)] 6 Component Analysis and Dimension Reduction III [Local Linear Embedding (LLE), Intrinsic dimension] 7 Boosting Techniques I [perceptron, backpropagation and Adaboost] 8 Boosting Techniques II [RealBoost and Example on face detection] 9 Boosting Techniques III [analysis, logit boost, cascade and decision policy] 10 Boosting Techniques III [analysis, logit boost, cascade and decision policy] 11 Non-metric method I [Decision tree and random forrest] 12 Non-metric method II [Syntactic pattern recognition and example on human parsing] 13 Support vector machine I [Kernel-induced feature space] 14 Support vector machine II [Support vector classifier] 15 Support vector machine III [Loss functions, Latent SVM] 16 Parametric Learning [ Maximum Likelihood Estimation ] [ Sufficient Statistics and Maximum entropy] 17 Non-parametric Learning I [ Parzen window and K-nn classifer] 18 Non-parametric Learning II [K-nn classifer and Error analysis] 19 Deep Learning I 20 Deep Learning II