Alert button
Picture for Mihailo R. Jovanović

Mihailo R. Jovanović

Alert button

Provably Efficient Generalized Lagrangian Policy Optimization for Safe Multi-Agent Reinforcement Learning

Add code
Bookmark button
Alert button
May 31, 2023
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Viaarxiv icon

Tradeoffs between convergence rate and noise amplification for momentum-based accelerated optimization algorithms

Add code
Bookmark button
Alert button
Sep 24, 2022
Hesameddin Mohammadi, Meisam Razaviyayn, Mihailo R. Jovanović

Figure 1 for Tradeoffs between convergence rate and noise amplification for momentum-based accelerated optimization algorithms
Figure 2 for Tradeoffs between convergence rate and noise amplification for momentum-based accelerated optimization algorithms
Figure 3 for Tradeoffs between convergence rate and noise amplification for momentum-based accelerated optimization algorithms
Figure 4 for Tradeoffs between convergence rate and noise amplification for momentum-based accelerated optimization algorithms
Viaarxiv icon

Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs

Add code
Bookmark button
Alert button
Jun 06, 2022
Dongsheng Ding, Kaiqing Zhang, Jiali Duan, Tamer Başar, Mihailo R. Jovanović

Figure 1 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Figure 2 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Figure 3 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Figure 4 for Convergence and sample complexity of natural policy gradient primal-dual methods for constrained MDPs
Viaarxiv icon

Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence

Add code
Bookmark button
Alert button
Feb 08, 2022
Dongsheng Ding, Chen-Yu Wei, Kaiqing Zhang, Mihailo R. Jovanović

Figure 1 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Figure 2 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Figure 3 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Figure 4 for Independent Policy Gradient for Large-Scale Markov Potential Games: Sharper Rates, Function Approximation, and Game-Agnostic Convergence
Viaarxiv icon

Transient growth of accelerated first-order methods for strongly convex optimization problems

Add code
Bookmark button
Alert button
Mar 14, 2021
Hesameddin Mohammadi, Samantha Samuelson, Mihailo R. Jovanović

Figure 1 for Transient growth of accelerated first-order methods for strongly convex optimization problems
Figure 2 for Transient growth of accelerated first-order methods for strongly convex optimization problems
Figure 3 for Transient growth of accelerated first-order methods for strongly convex optimization problems
Figure 4 for Transient growth of accelerated first-order methods for strongly convex optimization problems
Viaarxiv icon

Provably Efficient Safe Exploration via Primal-Dual Policy Optimization

Add code
Bookmark button
Alert button
Mar 01, 2020
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Viaarxiv icon

Convergence and sample complexity of gradient methods for the model-free linear quadratic regulator problem

Add code
Bookmark button
Alert button
Dec 26, 2019
Hesameddin Mohammadi, Armin Zare, Mahdi Soltanolkotabi, Mihailo R. Jovanović

Figure 1 for Convergence and sample complexity of gradient methods for the model-free linear quadratic regulator problem
Figure 2 for Convergence and sample complexity of gradient methods for the model-free linear quadratic regulator problem
Figure 3 for Convergence and sample complexity of gradient methods for the model-free linear quadratic regulator problem
Figure 4 for Convergence and sample complexity of gradient methods for the model-free linear quadratic regulator problem
Viaarxiv icon

Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach

Add code
Bookmark button
Alert button
Oct 02, 2019
Dongsheng Ding, Mihailo R. Jovanović

Figure 1 for Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach
Figure 2 for Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach
Figure 3 for Global exponential stability of primal-dual gradient flow dynamics based on the proximal augmented Lagrangian: A Lyapunov-based approach
Viaarxiv icon

Stochastic dynamical modeling of turbulent flows

Add code
Bookmark button
Alert button
Aug 26, 2019
Armin Zare, Tryphon T. Georgiou, Mihailo R. Jovanović

Figure 1 for Stochastic dynamical modeling of turbulent flows
Figure 2 for Stochastic dynamical modeling of turbulent flows
Figure 3 for Stochastic dynamical modeling of turbulent flows
Figure 4 for Stochastic dynamical modeling of turbulent flows
Viaarxiv icon

Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization

Add code
Bookmark button
Alert button
Aug 24, 2019
Dongsheng Ding, Xiaohan Wei, Zhuoran Yang, Zhaoran Wang, Mihailo R. Jovanović

Figure 1 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Figure 2 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Figure 3 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Figure 4 for Fast Multi-Agent Temporal-Difference Learning via Homotopy Stochastic Primal-Dual Optimization
Viaarxiv icon