Skip to Content

Sponsors

No results

Keywords

No results

Types

No results

Search Results

Events

No results
Search events using: keywords, sponsors, locations or event type
When / Where
All occurrences of this event have passed.
This listing is displayed for historical purposes.

Presented By: Department of Mathematics

Financial/Actuarial Mathematics Seminar

Learning in Linear Quadratic Framework: From Single-agent to Multi-agent

Linear quadratic framework is widely studied in the literature of stochastic control and game theory due to its simple structure, tractable solution, and various real-world applications. In this talk we discuss the theoretical convergence of policy gradient methods, one of the most popular reinforcement learning algorithms, in several linear quadratic problems. In the single agent setting, we show the global convergence of such methods to the optimal solution in the setting of known and unknown parameters. We also illustrate the performance of the algorithms in the optimal liquidation problem. In the multi-agent linear quadratic games, we show that the policy gradient method enjoys global convergence to the Nash equilibrium provided that there is a certain amount of noise in the system. The noise can either come from the underlying dynamics, or carefully designed explorations from the agents. This talk is based on joint work with Prof. Ben Hambly (University of Oxford) and Prof. Renyuan Xu (University of Southern California). Speaker(s): Huining Yang (Oxford)

Explore Similar Events

  •  Loading Similar Events...

Keywords


Back to Main Content