Black Board Lectures of the Year 2017

Machine Learning: The elegant way to extract information from data

Drs Jens Kauffmann, Veselina Kalinova, Dario Colombo and Tobias Albertson (MPIfR)

Overview

The human brain is able to recognize patterns in the data due to the millions of years of evolution. Today, we try to use and further develop computer algorithms that can reproduce this ability to distinguish patterns in large statistical samples and images, to find correlations between events, to classify objects based on their similar and different features in an objective and automatic way. Machine Learning is a sub-field in computer science that explores pattern recognition in the data analysis and characterize already known trends in sets of observations. This course focuses on the basic most applied up-to-date techniques in the machine learning that can be used in any branch of science and industry. We start with a broad introduction about the topic, followed by detailed discussion about the theory and application of the different types of the machine learning — supervised and unsupervised learning. In the field of supervised learning, for example, we will discuss regression, support vector machines, and neural networks, while the field of unsupervised learning we will cover clustering techniques, principal component analysis, and dimensionality reduction. The final goal of the course is to give the hands–on knowledge needed to pick and apply machine learning tools in Python. Discussions of the underlying mathematical principles will illustrate the inner workings of these tools.

Topics to be covered

Introduction — what is machine learning? 
Supervised Learning — regression, support vector machines, neural networks 
Unsupervised Learning — clustering, principal component analysis, dimensionality reduction 
Tools — Monte Carlo Markov chains, Bayesian inference

 

Schedule

Lecture 1: February 13 at 10:00 in 0.02: Introduction to Machine Learning: data mining, pattern recognition, dimensionality reduction, validation, objectivism | .mov  .mp4

Lecture 2: February 14 at 10:00 in 0.02: Principal Component Analysis (PCA): eigenvectors, eigenvalues, variance, covariance, eigenfaces, application to astronomy | .mov  .mp4

Lecture 3: February 15 at 10:00 in 0.02: Clustering I: distance definitions, hierarchical clustering, k-means | .mov  .mp4

Lecture 4: February 16 at 10:00 in 0.02: Clustering II: graph theory, spectral clustering, DBSCAN, expectation maximization via gaussian mixture models, affinity propagation | .mov  .mp4 

Lecture 5: February 20 at 10:00 in 0.02: Regression: chi-squared fitting, Markov Chain Monte Carlo (MCMC), Bayesian inference | .mov  .mp4

Lecture 6: February 21 at 10:00 in 0.02: Classification I: support vector machines, perceptron, artificial neural networks | .mov  .mp4

Lecture 7: February 22 at 10:00 in 0.02: Classification II: decision trees, naive Bayes, genetic algorithm | .mov  .mp4 

Lecture 8: February 23 at 10:00 in 0.02: Summary: review of methods, validation,connections to astronomy, resources for further learning | .mov  .mp4

 
loading content