1 Prerequisites

- Bachelor level knowledge and ability in calculus, probability, statistics, optimization, linear algebra.
- Basic programming skills in Python, including some familiarity with the Numpy and Matplotlib packages.

 

Recommended (from MasterMath; all are optional)

- Continuous optimization

- Measure Theoretic Probability

- Asymptotic Statistics

 

2 Aim of the course

 Statistical learning provides a probabilistic and statistical understanding of topics in machine learning.

 

The overarching goal of the course is to develop methods for estimating (or 'learning') an unknown function from data or making predictions for unseen function outputs. The course aims to empower the student to make a justified decision in adopting machine learning approaches for practical problems and even design their own machine learning methodology using sound mathematical principles. This is achieved by studying key ideas, models, algorithms and theories related to the subject of statistical learning.

 

In the first half of the course a collection of essential models, algorithms and techniques is introduced. In the second half of the course we adopt a Bayesian perspective on machine learning and study associated methods for analysing models and developing computational methodology.

 

Study Goals

By the end of the course, you are able to:

1. Explain the challenges of statistical learning (e.g., curse of dimensionality, overfitting, bias-variance trade-off).

2. Implement machine learning algorithms and computational methods on artificial and real-world data sets (e.g., linear regression, logistic regression, sparse regression, neural networks, decision trees).

3. Compare the theoretical properties and practical aspects of machine learning approaches.

4. Validate machine learning approaches on arbitrary data sets.

5. Design tailor-made computational approaches for new learning problems by adapting, combining and/or extending methods and/or by employing a Bayesian approach. 

3 Rules regarding Homework/Exam

The final grade of the course consists of the following components:
- Practical assignments during the course (30%)
- Final written exam (70%)

Final grade calculation: (0.3 * practical assignments + 0.7 * final written exam)

For both components, a sufficient grade (5.8) is required.

 

The results of the practical assignments are also used as part of the grade of the resit.

 

4 Lecture notes/literature

Lecture notes and slides will be made available.

 

You may wish to consult additional references for background reading.

 

Some suggestions of publicly available textbooks are (listed in order of suitability for this course):

- Bishop, C. M. (2006). Pattern recognition and machine learning. Springer, New York. https://www.microsoft.com/en-us/research/publication/pattern-recognition-machine-learning/ (very explicit in terms of derivations; focusses on a Bayesian approach)

- Hastie, T., Tibshirani, R., & Friedman, J. (2009). The elements of statistical learning (Second). Springer, New York. https://web.stanford.edu/~hastie/ElemStatLearn/ (more advanced reference; requires some mathematical and statistical maturity since it shows relatively few derivations; focusses on a non-Bayesian approach)

* Murphy, Kevin P (2022). Probabilistic machine learning: an introduction. MIT press, https://probml.github.io/pml-book/book1.html (very recent textbook on machine learning which discusses most topics)

 

5 Lecturers

Lecturer: Joris Bierkens

Instructions: Chris van Vliet