Machine Learning Theory CS6783
News :
  1. Join piazza: here
  2. Welcome to first day of class!

Location and Time : Location : Phillips hall, 407
Time : Tue-Thu 1:25 PM to 2:40 PM
Office Hours : Wed, 2-3pm, Gates 424

Description : We will discuss both classical results and recent advances in both statistical (iid batch) and online learning theory. We will also touch upon results in computational learning theory. The course aims at providing students with tools and techniques to understand inherent mathematical complexities of learning problems, to analyze and prove performance guarantees for machine learning methods and to develop theoretically sound learning algorithms.

Pre-requisite : Student requires a basic level of mathematical maturity and ease/familiarity with theorems and proofs style material. Familiarity with probability theory, basics of algorithms and an introductory course on Machine Learning (CS 4780 or equivalent) are required. M.Eng. and undergraduate students require permission of instructor.

Tentative topics :

    1. Introduction
    Overview of the learning problem : statistical and online learning frameworks.
    PAC and agnostic PAC models, online learning protocol, (adaptive/ oblivious adversaries, various notions of regret) connections to optimization and statistical estimation

    2. Minimax formulation for learning, distribution free and adversarial learning settings, uniform guarantees and no free lunch theorems

    3. Statistical Learning Framework
    . Empirical risk minimization and Regularized empirical risk minimization
    . Uniform convergence (iid data)
    . finite classes, PAC Bayes theorem, compression bounds
    . VC dimension and growth function
    . Rademacher complexity, covering numbers, Dudley integral bounds, fat-shattering dimension
    . Supervised learnability : necessary and sufficient conditions via uniform convergence (iid)
    . Local Rademacher analysis and fast rates

    4. Online Learning Framework (sequential prediction/decision making)
    . Learning with expert advice, perceptron, winnow
    . Sequential minimax analysis for online learning
    . Uniform convergence over martingales
    . Sequential Rademacher complexity, sequential covering numbers, Sequential fat-shattering dimension
    . Supervised online learnability : necessary and sufficient conditions via martingale uniform convergence

    5. Deriving algorithms through relaxation and minimax analysis for online learning

    6. Additional topics
    . Connections between learning and approximation algorithms
    . Connections between learning and optimization
    . Algorithmic stability tools
    . Statistical estimation Vs statistical learning Vs Stochastic optimization

Grading :
Assignments : There will be a total of 5 assignments covering 55% of your grade.

Term project :
    There will be a small term project due by the end of the course. The project is worth 40% of your grade. The project could be your choice of research problem approved by me for which you will submit a report by end of term or,

Email: sridharan at cs dot cornell dot edu