Coursebooks

Fundamentals in statistical pattern recognition

Anjos André
Marcel Sébastien

English

Every 2 years

Remarque

Next time: Spring 2021

Summary

This course provides in-depth understanding of the most fundamental algorithms in statistical pattern recognition as well as concrete tools (as source code) to PhD students for their work. It will cover regression, classification (MLP, SVM) and probability distribution modeling (k-Means, GMM, HMM).

Content

Learning outcomes: this course provides in-depth understanding in Statistical Pattern Recognition as well as concrete tools to PhD students for their work. This course could serve as a pre-requisite for more advanced courses such as Machine Learning, Graphical Models, Statistical Sequence Processing and Computational perception using multimodal sensors.

1. Introduction

- Data representation

- Supervised/unsupervised models (from regression and classification to probability distribution modelling)

- Overview: Linear models, Principal Component Analysis (PCA), Linear Discriminant Analysis (LDA), Multi-Layered Perceptron (MLP), Support Vector Machines (SVM), k-Means, Gaussian Mixture Models (GMM)  Hidden Markov Models (HMM), Session Variability Modelling (Joint Factor Analysis) and Total Variability Modelling (iVectors), Probabilistic Linear Discriminant Analysis (PLDA)

- Connection to other EPFL courses

- Evaluation methodologies and performance measures (precision, recall, FA, FR, EER, HTER, MSE, ROC, AUC, DCF, DET, EPC)

- Hypothesis testing and statistical significance

2. Classification and regression

- Application examples (digit recognition, face detection)

- k-Nearest Neighbors (k-NN)

- Linear Regression (univariate and multivariate) and the Gradient descent

- Logistic Linear Regression

- Multi-Layered Perceptron (MLP) and the BackPropagation

- Support Vector Machines (SVM)

3. Dimensionality reduction and clustering

- Application examples (data analysis and texture recognition)

- k-Means

- Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA)

4. Probability distribution modelling

- Application examples (speech recognition and biometrics)

- Histograms

- Gaussian Mixture Models (GMM) and the Expectation-Maximisation (EM) algorithm

- Hidden Markov Models (HMM)

Keywords

Pattern Recognition, Machine Learning, Linear models, PCA, LDA, MLP, SVM, GMM, HMM.

Learning Prerequisites

Recommended courses

Linear algebra, Probabilities and Statistics, Signal Processing, Python (for the Labs).

Assessment methods

Project report and oral presentation.

In the programs

• Semester
• Exam form
Multiple
• Credits
4
• Subject examined
Fundamentals in statistical pattern recognition
• Number of places
30
• Lecture
36 Hour(s)
• Practical work
20 Hour(s)

Reference week

Lecture
Exercise, TP
Project, other

legend

• Autumn semester
• Winter sessions
• Spring semester
• Summer sessions
• Lecture in French
• Lecture in English
• Lecture in German