Skip to content

jsharpna/AML

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Advanced Machine Learning

Prof. James Sharpnack

Winter 2019

Office hours: 10am-12pm MSB 4107 or by appointment

Introduction

Machine learning has undergone dramatic changes in the past decade. Empirical risk minimization, enabled by batch and online convex optimization, is well understood, and has been used in classification, regression, and unsupervised learning. Recently, stochastic optimization and a suite of tools has been used for non-convex optimization, which has enabled learning functions that are the composition of other simpler functions. This has enabled the learning of representations in deep neural networks, and has transformed the field of computer vision and natural language processing.

Decision making in Markovian environments is a separate field of research in machine learning. Passive learning in Markov models include the hidden markov models and recurrent neural networks. For active decision making, settings and methods range from multiarmed bandits to reinforcement learning. All of these developments have culminated to deep Q-networks and other deep reinforcement learning techniques. It is our ultimate goal to track these ideas from basic machine learning to deep reinforcement learning.

How to read this syllabus

You should read this syllabus as a list of topics that will take on average 30 minutes to go over, and some topics are dependent on other topics. The only rule for going through this material is that you must complete the dependencies before you go over a given topic. In this way the course forms a directed acyclic graph (DAG) and we will visit each vertex in an ordering that is consistent with the DAG structure. The format of the content section is the following:

topic (ABV) : dependency 1, dependency 2, ...

  • description
  • subtopics
  • references

The refences are some of my source material for a topic. You should in no way interpret this as required reading, and the course content are the lectures that I give and the content in this repository.

Prerequisites:

  • Linear Algebra (projections, matrix operations, singular value decomposition and spectrum)
  • Calculus and probability (differentiation, Newton's method, LLN, CLT, expectation, conditioning, Bayes theorem, etc.)
  • Statistical decision making (MLE, hypothesis testing, confidence intervals, linear and logistic regression)
  • Computer programming (Basic data structures, sorting, big-O notation, sequential programming language such as C, Python, R, Java)

Contents

Intro to machine learning (IML) :

  • Look into the basic terminology of machine learning, and a preview of what is to come
  • Definition of learning machine, supervised learning, loss, risk, cross-validation, basic APIs, computational complexity, OLS, KNN
  • "Machine Learning", Tom Mitchell, Chapter 1

Classification and Surrogate Loss (Class) : IML

  • Classification is an important subclass of learning, and it is the first example of a computational-statistical tradeoff with surrogate losses.
  • Hinge loss, logistic loss, surrogate losses, ROC and PR curves
  • "A Probabilistic Theory of Pattern Recognition", Devroye et al., Chapters 1-2
  • "Surrogate losses and F-divergences", Nguyen, Wainwright, Jordan, 2009.

Information Theory (IT) : Class

  • This is a continuation of Class with a look at F-divergences and other topics in information theory
  • F-divergences, differential entropy, and hypothesis testing
  • "Introduction to Non-parametric Estimation", Tsybakov, Chapter 2.
  • "Information Theory, Inference, and Learning Algorithms", Mackay, Chapter 1.
  • "A Probabilistic Theory of Pattern Recognition", Devroye et al., Chapter 3

Bias-Variance Tradeoff (BV) : IML

Directed Graphical Models and Hidden Markov Models (HMM) : IML

  • We will look at directed graphical models and how to learn in Markovian environments. In particular, we will see hidden Markov models (HMMs) and how to learn parameters in this context.
  • Forward backward algorithm, EM algorithm, DAGS
  • "An Introduction to Hidden Markov Models and Bayesian Networks", Zoubin Ghahramani.

Convex Optimization (Conv) : IML

  • We will see convex functions, subgradients, and why convexity is nice.
  • convexity, subgradient, subdifferential, first order conditions
  • "Convex Optimization", Boyd and Vandenberghe, Unit 1
  • "Subgradients", Boyd, Duchi, Vandenberghe.

Online Learning (Online) : Conv

  • In online learning, we read the data sequentially and need to make predictions in sequence. We will introduce the basic idea of online convex optimization and regret and see how this fits into the ideas we have learned in IML. We will see some basics of learning from experts.
  • "Introduction to Online Convex Optimization", Hazan, Chapter 1.
  • "Prediction, learning, and games", Cesa-Bianchi, Lugosi, Introduction.

Online convex optimization (OCO) : Online, Class

Unsupervised Learning (UL): Conv, Class

  • PCA is the first solvable non-convex programs that we will encounter. PCA can be used for learning latent factors and dimension reduction. We will cast it as a contrained loss minimization problem, and show it's connections to K-means clustering.
  • PCA, Frobenius norm, SVD, K-means, Lloyds algorithm
  • "Elements of Statistical Learning", Hastie, Tibshirani, Friedman, Chapter 14

Kernel Machines (Kern) : Class, UL

  • The kernel trick is a simple way to go from linear methods to non-linear methods, while ensuring computational feasibility. We will see the general kernel trick and apply it to SVMs and PCA.
  • Kernel trick, Kernel SVMs, Kernel PCA
  • "Kernel Methods in Machine Learning", Hoffman, Scholkopf, Smola, 2008.

Decision Trees and Random Forests (Tree) : Class

  • We will look at algorithms like CART and other decision trees. We will recall the bootstrap and bagging, then go over random forests. We will talk about random forests in the context of the bias-variance tradeoff.
  • CART, bagging, random forests
  • "Elements of Statistical Learning", Hastie, Tibshirani, Friedman, Chapter 15

Boosting (Boost) : OCO

Neural Nets (NNets) : Class, Conv

Non-convex Optimization (NonConv) : UL, NNets, Conv

Deep learning and optimization (Deep) : NonConv, NNets

Recurrent Neural Nets (RNN) : Deep, HMM

Convolutional Neural Nets (ConvNets) : Deep

Batch normalization (BatchNorm): Deep

Deep Autoencoders (Auto) : Deep, UL

  • Much like adding layers to linear classifiers can form non-linear classifiers, autoencoders add layers to PCA to perform non-linear dimension reduction. Time permitting we will look at variants such as sparse, and convolutional.
  • autoencoders, convolutional autoencoders
  • "Autoencoders", Deep Learning, Goodfellow, Bengio, Courville.

Generalized Adversarial Networks (GAN) : Auto

  • GANs is another deep unsupervised learner that seeks to make it hard to distinguish between fake and training images. This is accomplished by simultaneously learning and adversary and a generative model.
  • GANs
  • "Generalized Adversarial Networks", Goodfellow et al., 2014

Bandits (Bandit) : OCO

Contextual Bandits (Context) : Bandit

Reinforcement learning (RL) : Bandit

  • The bandit setting is a specific Markov decision process. Finding a policy that can perform well in the MDP setting is the focus of reinforcement learning. We will see the basic definitions of RL, Bellman iteration, and Monto Carlo methods.
  • Bellman iteration, RL, MDP, on/off-policy
  • "Reinforcement Learning", Sutton, Barto, Chapters 3-4

Policy gradients and REINFORCE (PGrad) : RL

Temporal difference learning (TD) : RL

Q-learning and SARSA (Qlearn) : TD

Proximal policy optimization (PPO) : PGrad, IT

Deep Q-learning (DQN) : Qlearn

Instruction Plan and Grading

  • the scribes have been assigned, you can find your lesson here
  • you will use LaTeX to write up the lectures, and should recieve an invite to a project on overleaf

We may not be able to get to every topic. My job is to present these methods and summarize the material in lecture. Each of you will act as scribe for one of these topics. Once this is completed to my satisfaction then you will pass the class with an A.

About

Advanced Machine Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published