Department of Mathematics
 Search | Help | Login

Math @ Duke





.......................

.......................


Publications [#340183] of Rong Ge

Papers Published

  1. Fazel, M; Ge, R; Kakade, SM; Mesbahi, M, Global Convergence of Policy Gradient Methods for the Linear Quadratic Regulator, 35th International Conference on Machine Learning Icml 2018, vol. 4 (January, 2018), pp. 2385-2413, ISBN 9781510867963
    (last updated on 2026/01/15)

    Abstract:
    Direct policy gradient methods for reinforcement learning and continuous control problems arc a popular approach for a variety of reasons: 1) they are easy to implement without explicit knowledge of the underlying model, 2) they are an "end- to-end" approach, directly optimizing the performance metric of interest, 3) they inherently allow for richly parameterized policies. A notable drawback is that even in the most basic continuous control problem (that of linear quadratic regulators), these methods must solve a non-convex optimization problem, where little is understood about their efficiency from both computational and statistical perspectives. In contrast, system identification and model based planning in opti- : Mal control theory have a much more solid theo- ! retical footing, where much is known with regards to their computational and statistical properties. , This work bridges this gap showing that (model ; free) policy gradient methods globally converge to the optimal solution and are efficient (polynomi- ' ally so in relevant problem dependent quantities) : With regards to their sample and computational complexities.

 

dept@math.duke.edu
ph: 919.660.2800
fax: 919.660.2821

Mathematics Department
Duke University, Box 90320
Durham, NC 27708-0320


x