Provably Secure Machine Learning Algorithms - Designing Reliable AI Systems
Paul G. Allen School via YouTube
Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a critical colloquium on machine learning security vulnerabilities presented by Jacob Steinhardt from Stanford University at the Paul G. Allen School. Delve into the challenges posed by malicious actors exploiting the statistical nature of learning algorithms, and discover why current defense mechanisms are inadequate. Learn about the need for provably secure algorithms designed with robust statistics and optimization in mind. Examine the brittleness of high-dimensional learning algorithms and uncover new approaches for achieving robustness. Gain insights into certifiably robust optimization for non-convex models like neural networks. Understand Steinhardt's vision for creating machine learning systems with the reliability of well-designed software, and his work on algorithms that can detect failures and generalize predictably in novel situations.
Syllabus
UW Allen School Colloquium: Jacob Steinhardt (Stanford University)
Taught by
Paul G. Allen School