Search programme

​Use the search function to search amongst programmes at Chalmers. The study programme and the study programme syllabus relating to your studies are generally from the academic year you began your studies.

Syllabus for

Academic year
EEN100 - Statistics and machine learning in high dimensions
Statistik och maskininlärning i högre dimensioner
Syllabus adopted 2020-02-18 by Head of Programme (or corresponding)
Owner: MPCOM
7,5 Credits
Grading: UG - Pass, Fail
Education cycle: Second-cycle
Major subject: Electrical Engineering, Mathematics

Teaching language: English
Application code: 13121
Open for exchange students: Yes
Block schedule: A
Maximum participants: 40

Module   Credit distribution   Examination dates
Sp1 Sp2 Sp3 Sp4 Summer course No Sp
0120 Oral examination 6,0 c Grading: UG   6,0 c    
0220 Project 1,5 c Grading: UG   1,5 c    

In programs



Giuseppe Durisi

  Go to Course Homepage


General entry requirements for Master's level (second cycle)
Applicants enrolled in a programme at Chalmers where the course is included in the study programme are exempted from fulfilling the requirements above.

Specific entry requirements

English 6 (or by other approved means with the equivalent proficiency level)
Applicants enrolled in a programme at Chalmers where the course is included in the study programme are exempted from fulfilling the requirements above.

Course specific prerequisites

A solid foundation in probability and calculus


The explosion in the volume of data collected in all scientific disciplines and in industry requires students interested in statistical analyses and machine-learning and signal-processing algorithms to acquire more sophisticated probability tools than the ones taught in basic probability courses.

This course provides an introduction to the area of high-dimensional statistics, which deals with large scale problems where both the number of parameters and the sample size is large.

The course covers fundamental tools for the analysis of random vectors, random matrices and random projections, such as tail bounds and concentration inequalities.
It further provides concrete applications of such tools in the context of generalization-error analyses in statistical learning theory, sparse linear model, and matrix models with rank constraints.

Learning outcomes (after completion of the course the student should be able to)

  • State basic tails and concentration bounds for sums of independent random variables
  • Apply these bounds to provide guarantees on how accurately one can
    • estimate a covariance matrix from data
    • recover a sparse linear vector from noisy linear projections
    • estimate a low-rank matrix from few of its entries


  • Fundamental probability tools
    • Preliminaries on random variables: classical inequalities and limit theorems
    • Concentration of sums of independent random variables: Hoeffding, Chernoff, Bernstein, sub-Gaussian and sub-exponential distributions
    • Random vectors and random matrices in high dimensions
    • Concentration without independence
    • Uniform laws of large number: Rademacher complexity and VC dimension

  • Applications in machine learning, statistics, and signal processing
    • Covariance matrix estimation
    • Recovery of sparse signals
    • Principal component analysis
    • Low-rank matrix recovery
    • Sample complexity in statistical learning theory


Approximately 16 lectures, 8 exercise sessions, and one project where the tools learnt in the course will be used to solve a real-world machine-learning task


The course will be partly based on the following references
  • R. Vershynin, High-dimensional probability: an introduction with applications in data science. Cambridge Univ. Press, 2019. Available: Online
  • M. J. Wainwright, High-dimensional statistics: a nonasymptotic viewpoint. Cambridge, U.K.: Cambridge Univ. Press, 2019.

Examination including compulsory elements

Oral exam and project (pass or fail)

Page manager Published: Mon 28 Nov 2016.