Alert button
Picture for Dongsheng Ding

Dongsheng Ding

Alert button

Resilient Constrained Reinforcement Learning

Add code
Bookmark button
Alert button
Dec 29, 2023
Dongsheng Ding, Zhengyan Huan, Alejandro Ribeiro

Viaarxiv icon

Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs

Add code
Bookmark button
Alert button
Jun 20, 2023
Dongsheng Ding, Chen-Yu Wei, Kaiqing Zhang, Alejandro Ribeiro

Figure 1 for Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs
Figure 2 for Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs
Figure 3 for Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs
Figure 4 for Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs
Viaarxiv icon

Provably Efficient Generalized Lagrangian Policy Optimization for Safe Multi-Agent Reinforcement Learning

Add code
Bookmark button
Alert button
May 31, 2023
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Viaarxiv icon

Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs

Add code
Bookmark button
Alert button
Jun 06, 2022
Dongsheng Ding, Kaiqing Zhang, Jiali Duan, Tamer Başar, Mihailo R. Jovanović

Figure 1 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Figure 2 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Figure 3 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Figure 4 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Viaarxiv icon

Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence

Add code
Bookmark button
Alert button
Feb 08, 2022
Dongsheng Ding, Chen-Yu Wei, Kaiqing Zhang, Mihailo R. Jovanović

Figure 1 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Figure 2 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Figure 3 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Figure 4 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Viaarxiv icon

Provably Efficient Safe Exploration via Primal-Dual Policy Optimization

Add code
Bookmark button
Alert button
Mar 01, 2020
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Viaarxiv icon

Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach

Add code
Bookmark button
Alert button
Oct 02, 2019
Dongsheng Ding, Mihailo R. Jovanović

Figure 1 for Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach
Figure 2 for Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach
Figure 3 for Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach
Viaarxiv icon

Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization

Add code
Bookmark button
Alert button
Aug 24, 2019
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Figure 1 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Figure 2 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Figure 3 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Figure 4 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Viaarxiv icon

Fast multi-agent temporal-difference learning via homotopy stochastic primal-dual optimization

Add code
Bookmark button
Alert button
Aug 07, 2019
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Figure 1 for Fast multi-agent temporal-difference learning via homotopy stochastic primal-dual optimization
Figure 2 for Fast multi-agent temporal-difference learning via homotopy stochastic primal-dual optimization
Figure 3 for Fast multi-agent temporal-difference learning via homotopy stochastic primal-dual optimization
Figure 4 for Fast multi-agent temporal-difference learning via homotopy stochastic primal-dual optimization
Viaarxiv icon