Offline Reinforcement Learning and Model-Based Optimization

Offline Reinforcement Learning and Model-Based Optimization

Simons Institute via YouTube Direct link

Distribution shift in a nutshell

9 of 23

9 of 23

Distribution shift in a nutshell

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Offline Reinforcement Learning and Model-Based Optimization

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Intro
  2. 2 What makes modern machine learning word
  3. 3 Predictive models are very powerful!
  4. 4 Automated decision making is very powerf
  5. 5 First setting: data-driven reinforcement lear
  6. 6 Second setting: data-driven model-based optimization
  7. 7 Off-policy RL: a quick primer
  8. 8 What's the problem?
  9. 9 Distribution shift in a nutshell
  10. 10 How do prior methods address this?
  11. 11 Learning with Q-function lower bounds Algorithm
  12. 12 Does the bound hold in practice?
  13. 13 How does CQL compare?
  14. 14 Predictive modeling and design
  15. 15 What's wrong with just doing prediction?
  16. 16 The model-based optimization problem
  17. 17 Uncertainty and extrapolation
  18. 18 What can we do?
  19. 19 Model inversion networks (MINS)
  20. 20 Putting it all together
  21. 21 Experimental results
  22. 22 Some takeaways
  23. 23 Some concluding remarks

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.