Reinforcement Learning in Feature Space: Complexity and Regret

Reinforcement Learning in Feature Space: Complexity and Regret

Simons Institute via YouTube Direct link

Intro

1 of 23

1 of 23

Intro

Class Central Classrooms beta

YouTube playlists curated by Class Central.

Classroom Contents

Reinforcement Learning in Feature Space: Complexity and Regret

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Intro
  2. 2 Markov decision process
  3. 3 What does a sample mean?
  4. 4 Complexity and Regret for Tabular MDP
  5. 5 Rethinking Bellman equation
  6. 6 State Feature Map
  7. 7 Representing value function using linear combination of features
  8. 8 Reducing Bellman equation using features
  9. 9 Sample complexity of RL with features
  10. 10 Learning to Control On-The-Fly
  11. 11 Episodic Reinforcement Learning
  12. 12 Hilbert space embedding of transition kernel
  13. 13 The MatrixRL Algorithm
  14. 14 Regret Analysis
  15. 15 From feature to kernel
  16. 16 MatrixRL has a equivalent kernelization
  17. 17 Pros and cons for using features for RL
  18. 18 What could be good state features?
  19. 19 Finding Metastable State Clusters
  20. 20 Example: stochastic diffusion process
  21. 21 Unsupervised state aggregation learning
  22. 22 Soft state aggregation for NYC taxi data
  23. 23 Example: State Trajectories of Demon Attack

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.