Dimension Reduction and Metric Learning

STAT 946

 

News:

  • Register the date of your presentation here
 

Wikicoursenote
Project
Course Outline
Readings
Resources
Important Dates
Old lectures
Papers

Fall 2009
Department of Statistics and Actuarial Science
University of Waterloo
 

Instructor: Ali Ghodsi
Room: MC 4063
Time: TTh 01:00-02:20
Office Hours:  2:300-3:30 T or by appointment   (MC 6081G)

Readings:

Tutorials:

Kernel PCA:

Locally Linear Embedding (LLE):

Isomap:

MDS, Landmark MDS and Nystrom Approximation:

Maximum Variance Unfolding (MVU)  / Semidefinite Embedding (SDE):

Landmark SDE:

Action Respecting Embedding (ARE):

Clustering (Impossibility Theorem):

K-means Clustering:

Metric Learning:

Spectral Clustering:

 

Resources:


Old Lectures from Fall 2006:

Lecture 1 and 2 Motivation
Lecture 3 and 4 Principal Components Analysis (PCA)                                Slides
Lecture 5 PCA, Kernel function
Lecture 6 Dual PCA, Kernel PCA                                                          Slides
Lectures 7 and 8 Centering, Locally Linear Embedding  (LLE)                      Slides (Examples are taken from this paper.)
Lecture 9 Locally Linear Embedding 
Project Discussion  
Thanksgiving  
Lectures 10 and 11 Multidimensional Scaling (MDS), Isomap                           Slides
Lecture 12 Nystrom Approximation, Landmark MDS                         
Lecture 13 Landmark MDS
Lectures 14, 15 and 16 Unified Framework, Semidefinite Embedding (SDE)
Lecture 17 Landmark SDE
Lecture 18 Action Respecting Embedding (ARE)
Lecture 19 Clustering
Lectures 20 and 21 Combinatorial Algorithms, K-means clustering
Lectures 22 and 23 Mixture Models
Lectures 24 and 25 Learning a Metric (Class-Equivalence Side Information)
Lecture 26 Learning a Metric (Partial Distance Side Information)
     

Slides for some of the New Lectures

Motivation
Principal Components Analysis (PCA)
Dual PCA, Kernel PCA
Multidimensional Scaling (MDS), Isomap  
Spectral Clustering
Maximum Variance Unfolding
Action Respecting Embedding
Nonnegetive Matrix Factorization (NMF)
 

Important Dates:

June 23          Presentations will start
June 30        Proposal due
Aug .15        Final project reports due
   

Papers:

 Register the date of your presentation here

Note: You need to have a Wikicoursenote account.  To obtain a user account, you must request one.

 

          1. Graph Laplacian Regularization for Large-Scale Semidefinite Programming ( pdf )

     Nonlinear Dimensionality Reduction by Semidefinite Programming and Kernel Matrix Factorization ( pdf )

2.  A DIRECT FORMULATION FOR SPARSE PCA USING SEMIDEFINITE PROGRAMMING ( pdf )

3.  Learning Spectral Clustering,With Application To Speech Separation ( pdf )

4. Dimensionality Reduction for Supervised Learning with Reproducing Kernel Hilbert Spaces ( pdf )

5.  Regression on Manifolds Using Kernel Dimension Reduction ( pdf )

6. Visualizing Data using t-SNE ( pdf )

7.  Learning a Nonlinear Embedding by Preserving Class Neighbourhood Structure ( pdf )

8.  Visualizing Similarity Data with a Mixture of Maps ( pdf )

9.  Convex and Semi-Nonnegative Matrix Factorizations ( pdf )

10. Maximum-Margin Matrix Factorization ( pdf )

11. Maximum Margin Matrix Factorization for Collaborative Ranking ( pdf )

12. Kernelized Sorting ( pdf )

13.  Random Projections for Manifold Learning ( pdf )

14.  Learning Distance Functions using Equivalence Relations ( pdf

       Adjustment Learning and Relevant Component Analysis ( pdf )

15.  Neighbourhood Components Analysis ( pdf )

16. Measuring Statistical Dependence with Hilbert-Schmidt Norms ( pdf )

 

Project Report:

Final project reports (up to 8 pages of PDF) are worth 50% of your final grade .You are encouraged to chose a topic related to your research area. However,  you cannot  borrow part of an existing thesis work, nor can  you re-use a project from another course.  

Academic Dishonesty:

 Example:

Plagiarism is an act of “using ideas, plots, text and other intellectual property developed by someone else while claiming it is your original work.”1

References:

1. Tec Encyclopedia. http://www.answers.com/topic/plagiarism

Evidence of  copying or plagiarism will cause a failing mark in the course.

Please attach this cover page to your report.

I use this marking scheme to mark the projects.