Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn about dimensionality reduction techniques in this university lecture that explores random projections, PCA/SVD, and frequent directions algorithms. Begin with a thorough recap of random projection motivation and algorithmic implementation, including methods for choosing random unit vectors. Compare and contrast random projection with PCA/SVD approaches, before diving into an in-depth discussion of frequent directions. Understand the Misra-Gries algorithm for frequent items and its relationship to dimensionality reduction. Conclude with a practical example examining Linformer, which demonstrates the real-world application of SVD and random projection techniques in modern machine learning architectures.
Syllabus
Recording starts
Announcements
Random projection motivation recap
Random projection algorithm recap
Choosing random unit vectors
Random projection vs. PCA/SVD
Dimensionality reduction so far
Frequent directions motivation
Frequent items / Misra-Gries reminder
Frequent directions algorithm
Linformer an example of using SVD and random projection
Lecture ends
Taught by
UofU Data Science