Approximation benefits of policy gradient methods with aggregated states
Year of publication: |
2023
|
---|---|
Authors: | Russo, Daniel |
Published in: |
Management science : journal of the Institute for Operations Research and the Management Sciences. - Hanover, Md. : INFORMS, ISSN 1526-5501, ZDB-ID 2023019-9. - Vol. 69.2023, 11, p. 6898-6911
|
Subject: | approximate dynamic programming | policy gradient methods | reinforcement learning | state aggregation | Theorie | Theory | Mathematische Optimierung | Mathematical programming | Dynamische Optimierung | Dynamic programming |
-
Global optimality guarantees for policy gradient methods
Bhandari, Jalaj, (2024)
-
A low-rank approximation for MDPs via moment coupling
Zhang, Amy B. Z., (2024)
-
Risk-averse approximate dynamic programming with quantile-based risk measures
Jiang, Daniel R., (2018)
- More ...
-
Learning to optimize via posterior sampling
Russo, Daniel, (2014)
-
Learning to optimize via information-directed sampling
Russo, Daniel, (2018)
-
Simple Bayesian algorithms for best-arm identification
Russo, Daniel, (2020)
- More ...