Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Towards a Theory of Exploration in Continuous-Time Reinforcement Learning

Centre for Networked Intelligence, IISc via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Watch a comprehensive lecture where Professor Harsha Honnappa from Purdue University explores the theoretical framework of continuous-time reinforcement learning in stochastic environments. Delve into the complexities of exploration processes using rough path theory, examining how traditional reinforcement learning methods must be adapted for ultra-high frequency interactions. Learn about the challenges of continuous-time exploration, the application of relaxed controls in Wasserstein space, and the development of pathwise relaxed control frameworks. Discover how these concepts apply to real-world scenarios like stock market volatility modeling and controlled stochastic networks with heavy-tailed service. Understand the mathematical foundations behind establishing value function existence and uniqueness through rough Hamilton-Jacobi-Bellman equations, and explore the implications for entropy-regularized objectives in reinforcement learning.

Syllabus

Time: 4:00 to PM

Taught by

Centre for Networked Intelligence, IISc

Reviews

Start your review of Towards a Theory of Exploration in Continuous-Time Reinforcement Learning

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.