Back to MJ's Publications
-
D. Ding,
K. Zhang,
J. Duan,
T. Basar,
and M. R. Jovanovic.
Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs.
J. Mach. Learn. Res.,
2022.
Note: Submitted; also arXiv:2206.02346.
Keyword(s): Constrained Markov decision processes,
Constrained nonconvex optimization,
Function approximation,
Natural policy gradient,
Policy gradient methods,
Primal-dual algorithms,
Sample complexity.
[bibtex-entry]
-
D. Ding,
K. Zhang,
T. Basar,
and M. R. Jovanovic.
Convergence and optimality of policy gradient primal-dual method for constrained Markov decision processes.
In Proceedings of the 2022 American Control Conference,
Atlanta, GA,
pages 2851-2856,
2022.
Keyword(s): Constrained Markov decision processes,
Policy gradient methods,
Primal-dual algorithms.
[bibtex-entry]
-
D. Ding,
K. Zhang,
T. Basar,
and M. R. Jovanovic.
Natural policy gradient primal-dual method for constrained Markov decision processes.
In Proceedings of the 34th Conference on Neural Information Processing Systems,
volume 33,
Vancouver, Canada,
pages 8378-8390,
2020.
Keyword(s): Constrained Markov decision processes,
Constrained nonconvex optimization,
Natural policy gradient,
Policy gradient methods,
Primal-dual algorithms.
[bibtex-entry]
Back to MJ's Publications
Disclaimer:
This material is presented to ensure timely dissemination of
scholarly and technical work. Copyright and all rights therein
are retained by authors or by other copyright holders.
All person copying this information are expected to adhere to
the terms and constraints invoked by each author's copyright.
In most cases, these works may not be reposted
without the explicit permission of the copyright holder.
Last modified: Sat Oct 5 22:00:41 2024
Author: mihailo.
This document was translated from BibTEX by
bibtex2html