Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Model Orthogonalization - Class Distance Hardening in Neural Networks for Better Security

IEEE via YouTube

Overview

Explore a 19-minute IEEE conference talk on enhancing neural network security through model orthogonalization and class distance hardening. Delve into the research conducted by a team from Purdue University, including Guanhong Tao, Yingqi Liu, Guangyu Shen, Qiuling Xu, Shengwei An, Zhuo Zhang, and Xiangyu Zhang. Learn about innovative techniques to improve the robustness and reliability of neural networks in the face of security challenges, and gain insights into the potential applications of these methods in various fields of artificial intelligence and machine learning.

Syllabus

Model Orthogonalization: Class Distance Hardening in Neural Networks for Better Security

Taught by

IEEE Symposium on Security and Privacy

Reviews

Start your review of Model Orthogonalization - Class Distance Hardening in Neural Networks for Better Security

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.