Presented By: Department of Mathematics
Financial/Actuarial Mathematics Seminar
Exploration-exploitation trade-off for continuous-time episodic reinforcement learning with linear-convex models
We develop a probabilistic framework for analysing model-based reinforcement learning in the episodic setting. We then apply it to study finite-time horizon stochastic control problems with linear dynamics but unknown coefficients and convex, but possibly irregular, objective function. Using probabilistic representations, we study regularity of the associated cost functions and establish precise estimates for the performance gap between applying optimal feedback control derived from estimated and true model parameters. Next, we propose a phase-based learning algorithm for which we show how to optimise exploration-exploitation trade-off. Our algorithm achieves sublinear (or even logarithmic) regrets in high probability and expectation, which matches the best possible results from the literature. Speaker(s): Yufei Zhang (LSE)