Selected Problems in Machine Learning

Course focus

The course has been designed especially for PhD students with a deep interest in Machine Learning. The course is a flexible combination of lectures, discussions, exercises and literature reading, aimed at the following three topics:
  1. refreshing (and deeper understanding of) basic notions of Machine Learning
  2. introduction to Bayesian inference
  3. practising unsupervised ML (especially methods based on sampling)

Course prerequisities

Students are expected to be familiar with basic probabilistic and ML concepts, roughly in the extent of NPFL067/068 - Statistical methods in NLP I/II, and NPFL 054 - Introduction to Machine Learning (in NLP).

Course schedule

  1. "Calibration" test - let me know what you already know
  2. Patching the holes disclosed by the calibration test.
  3. Patching the holes disclosed by the calibration test, continued.
  4. Very studious excercise on Beta distribution.
    • let us admire two mighty parameters generating a broad family of different shapes
    • generalization to n-dimenzions - Dirichlet distribution
    • supplementary materials - mathematicalmonks's videos:
  5. Derivation of some simple Bayesian models - let's enjoy conjugacy!
  6. Assignment 1 - Word-alignment using Gibbs sampling
  7. Reading - Bayesian Inference
  8. Kernel methods
  9. Assignment 2 - Segmentation of dependency trees
  10. Gibbs sampling in NLP - two case studies

Other useful links

Course passing requirements

All students are required to actively participate in the classes.