Saarland University, Machine Learning Group, Fak. MI - Mathematik und Informatik, Campus E1 1, 66123 Saarbrücken, Germany     

Machine Learning Group
Department of Mathematics and Computer Science - Saarland University

TEACHING

MACHINE LEARNING

Wintersemester 2008/2009

GENERAL INFORMATION

In a broader perspective machine learning tries to automatize the process of empirical sciences - namely extracting knowledge about natural phenomena from measured data with the goal to either understand better the underlying processes or to make good predictions. Machine learning methods are therefore widely used in different fields: bioinformatics, computer vision, information retrieval, computer linguistics, robotics,...

The lecture gives a broad introduction into machine learning methods. After the lecture the students should be able to solve and analyze learning problems.

List of topics (tentative)

  • Reminder of probability theory
  • Maximum Likelihood/Maximum A Posteriori Estimators
  • Bayesian decision theory
  • Linear classification and regression
  • Kernel methods
  • Model selection and evaluation of learning methods
  • Feature selection
  • Nonparametric methods
  • Boosting, Decision trees
  • Neural networks
  • Structured Output
  • Semi-supervised learning
  • Unsupervised learning (Clustering, Independent Component Analysis)
  • Dimensionality Reduction and Manifold Learning
  • Statistical learning theory

Previous knowledge of machine learning is not required. The participants should be familiar with linear algebra, analysis and probability theory on the level of the local `Mathematics for Computer Scienticists I-III' lectures. In particular, attendees should be familiar with

Type: Core lecture (Stammvorlesung), 9 credit points

LECTURE MATERIAL

Incremental lecture notes (last update: 09.02.2009).

The practical exercises will be in Matlab.

SLIDES AND EXCERCISES

22.10. - Introduction
24.10. - Reminder: Probability Theory Exercise 1 Solution 1
29.10. - Bayesian Decision Theory Decision boundary demo
31.10. - (Reg.) Empirical risk minimization Exercise 2 Solution 2
05.11. - Linear Regression
07.11. - Introduction Optimization Exercise 3 Solution 3 Data for the Exercise
12.11. - Linear Classification
14.11. - Linear Classification (cont.) Exercise 4 Solution 4 Data and functions
19.11. - Kernels and RKHS
21.11. - Kernelized Algorithms Exercise 5 Solution 5 LIBSVM, data, etc.
26.11. - Regularization etc.
28.11. - Evaluation, Statistical Tests Exercise 6 Solution 6 Data for Exercise 14
03.12. - Statistical tests cont.
05.12. - Class. Comparison, Model selection Exercise 7 Solution 7 Data for Exercise 16/17
10.12. - Midterm exam
12.12. - Lecture canceled
17.12. - Feature selection I
19.12. - Lecture canceled
07.01. - Feature selection II
09.01. - Boosting Exercise 8 Solution 8 Data for Exercise 18 (=14)
14.01. - Decision Trees, Neural Networks
and Nearest Neighbor Methods
16.01. - Semi-supervised Learning Exercise 9 Solution 9 Data for the Competition
21.01. - Semi-supervised Learning (cont.)
23.01. - K-Means and Spectral Clustering Exercise 10 Solution 10
28.01. - Spectral Clustering II
30.01. - Clustering II Exercise 11 Solution 11
04.02. - Dimensionality Reduction
06.02. - Dimensionality Reduction II
11.02. - Statistical Learning Theory I
13.02. - Statistical Learning Theory II

LITERATURE AND OTHER RESOURCES

  • The lecture will be partially based on the following books and partially on recent research papers:
    • R.O. Duda, P.E. Hart, and D.G.Stork: Pattern Classification, Wiley, (2000).
    • B. Schoelkopf and A. J. Smola: Learning with Kernels, MIT Press, (2002).
    • J. Shawe-Taylor and N. Christianini: Kernel Methods for Pattern Analysis, Cambridge University Press, (2004).
    • C. M. Bishop: Pattern recognition and Machine Learning, Springer, (2006).
    • T. Hastie, R. Tibshirani, J. Friedman: The Elements of Statistical Learning, Springer, second edition, (2008).
    • L. Devroye, L. Gyoerfi, G. Lugosi: A Probabilistic Theory of Pattern Recognition, Springer, (1996).
    • L. Wasserman: All of Statistics, Springer, (2004).
    • S. Boyd and L. Vandenberghe: Convex Optimization, Cambridge University Press, (2004).
  • Other resources:

NEWS

You can pick up your course certificate ("Schein") in my office.

Re exam: Results

Final grades: Grades. The graded certificate ("Schein") can be picked up in my office from the 15.4.

Results of the competition on user-guided image segmentation.

Next semester there will be a course offered by Dr. Seeger on "Bayesian Machine Learning: Graphical Models and Approximate Inference". This course is complementary to this lecture and provides an in-depth introduction to Bayesian learning in particular graphical models.

TIME AND LOCATION

Lecture: We, 8.30 s.t. -10, and Fr, 10.15 s.t. -12, E1 3, HS III

Exercise groups:

  • Group A, Wednesday, 16-18, Seminar room 5, E2 4, tutor: Thomas Buehler
  • Group B, Thursday, 16-18, Seminar room 15, E1 3, tutor: Manuel Noll
  • Group C, Friday, 14-16, Seminar room 7, E2 4, tutor: Pavlo Lutsik

EXAMS AND GRADING

Exams: Mid-term: 10.12.    End-term: 18.2.    Re-exam: 1.4.

Grading:

  • 50% of the points in the exercises (up to that point) are needed to take part in the exams (end-term/re-exam). In order to being admitted for the endterm and re-exam, you need to have presented properly once a solution in the exercise groups.
  • An exam is passed if you get at least 50% of the points.
  • The grading is based on the best result of the end-term and re-exam

LECTURER

Jun.-Prof. Dr. Matthias Hein

Office Hours: Mo, 16-18, Do, 16-18