CMU Neural Nets for NLP - Model Interpretation

CMU Neural Nets for NLP - Model Interpretation

Graham Neubig via YouTube Direct link

Explanation Technique: Influence Functions

12 of 15

12 of 15

Explanation Technique: Influence Functions

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

CMU Neural Nets for NLP - Model Interpretation

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Intro
  2. 2 Why interpretability?
  3. 3 Dictionary definition
  4. 4 Two broad themes
  5. 5 Comparing two directions
  6. 6 Source Syntax in NMT
  7. 7 Why neural translations are the right length?
  8. 8 Fine grained analysis of sentence embeddings
  9. 9 What you can cram into a single vector: Probing sentence embeddings for linguistic properties
  10. 10 How to evaluate?
  11. 11 Automatic evaluation
  12. 12 Explanation Technique: Influence Functions
  13. 13 Explanation Techniques: gradient based importance scores
  14. 14 Explanation Technique: Extractive Rationale Generation
  15. 15 Future Directions

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.