Machine Learning (Spring 2020)

Announcements

Final Project Term Paper (5 page, double-column, IEEE style) is due by Wednesday, April 29, 2:00 pm.

Final Project Presentations to happen online during Monday, April 27 and Wednesday, April 29.

  • Lectures

    • (January 22, 2020) Lecture 1 : Introduction and Formal Model of Learning (cf. [R1], Chapter 1). Paper by Leo Breiman on the two cultures can be found here.

    • (January 27, 2020) Lecture 2 : PAC Learnability and Agnostic PAC Learnability (cf. [R1], Chapter 2)

    • (January 29, 2020) Lecture 3 : Learning via Uniform Convergence (cf. [R1], Chapter 3)

    • (February 3, 2020) Lecture 4 : No-Free Lunch Theorem and Bias-Complexity Tradeoff (cf. [R1], Theorem 5.1, Corollary 5.2)

    • (February 5, 2020) Lecture 5 : VC-Dimension Theory (cf. [R1], Chapter 6)

    • (February 10, 2020) Lecture 6 : The Fundamental Theorem of PAC Learning (cf. [R1], Theorem 6.7, Lemma 6.10, Theorem 6.11)

    • (February 12, 2020) Lecture 7 : Convexity and Learning (cf. [R1], Chapter 12)

    • (February 17, 2020) Lecture 8 : Regularized Loss Minimization and Stable Rules (cf. [R1], Sections 13.1, 13.2, 13.3)

    • (February 24, 2020) Lecture 9 : Overfitting-Stability Tradeoff and Gradient Descent (cf. [R1], Corollary 13.8, Corollary 13.9, Section 14.1)

    • (February 26, 2020) Lecture 10 : Gradient Descent (contd.) (cf. [R1], Sectino 14.1, Section 14.2)

    • (March 2, 2020) Lecture 11 : Stochastic Gradient Descent (cf. Stochastic Gradient Descent and Its Variants in Machine Learning)

    • (March 4, 2020) Lecture 12 : Linear Predictors and Linear Regression (cf. [R1], Section 9.1, Section 9.2)

    • (March 9, 2020) Lecture 13 : Logistic Regresstion, Probability Interpretation of Regression, Regression Trees (cf. Andrew Ng's Notes; [R3], Section 9.2)

    • (March 11, 2020) Lecture 14 : Classification Trees, Ensemble Learning I : Bagging, Random Forests (cf. [R1], Chapter 18; [R3], Section 9.2, Section 15.1, Section 15.2, Section 15.3)

    • (March 23, 2020) Lecture 15 : Boosting, AdaBoost, Validation (cf. [R1], Chapter 10, Chapter 11)

    • (March 30, 2020) Lecture 16 : Support Vector Machines (cf. [R1], Chapter 15)

    • (April 1, 2020) Lecture 17 : Learning with Kernels (cf. [R1], Chapter 16)

    • (April 6, 2020) Lecture 18 : Nearest Neigbor Methods (cf. [R1], Chapter 19. Here is an interesting paper on the application of kNN methods in the estimation of information theoretic quantities.)

    • (April 8, 2020) Lecture 19 : Unsupervised Learning : k-means Clustering (cf. [R1], Chapter 22)

    • (April 13, 2020) Lecture 20 : k-means Clustering (contd.), Spectral Clustering, Information Bottleneck (cf. [R1], Chapter 22)

    • (April 15, 2020) Lecture 21 : Dimentionality Reduction (PCA), Compressed Sensing (cf. [R1], Chapter 23)

    • (April 20, 2020) Lecture 22 : Generative Models, Neural Networks (cf. [R1], Chapter 24, Andrew Ng's Notes)

    • (April 27, 2020) Lecture 23 : Final Project Presentations

    • (April 29, 2020) Lecture 24 : Final Project Presentations

  • Final Project

    • Term Paper, due by Wednesday, April 29, 2020, 2:00 pm

    • Presentations on Monday, April 27, 2020 and Wednesday, April 29, 2020

  • General Information

    • Instructor : Himanshu Asnani

    • Venue : A-201

    • Class Timings : Mondays, 4:00 pm - 5:30 pm and Wednesdays, 2:00 pm - 3:30 pm

    • Grading : Class Participation (10%), Homeworks (60%), Final Project (30%)

    • Assignments : We will be using Jupyter Notebook for the programming aspect of the course. Programming language will be Python and the installation guide can be located here.

    • Tentative Schedule

      • Module I [Theoretical Foundations] PAC Learning, Bias-Complexity Trade-off, VC Theory, Generalization, Gradient Descent, Stochastic Gradient Descent

      • Module II [Supervised Learning] Regression, Classification, Support Vector Machines, Decision Trees, Bagging, Boosting, Kernel Methods, Nearest Neighbor Methods, Neural Networks, Backpropagation

      • Module III [Unsupervised Learning] k-means Clustering, Spectral Clustering, Dimensionality Reduction (PCA), Compressed Sensing, Information Bottleneck, Naive Bayes, Linear Discriminative Analysis, Expectation-Maximization Algorithm