Alert button
Picture for Eric Mazumdar

Eric Mazumdar

Alert button

Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets

Add code
Bookmark button
Alert button
Jun 06, 2022
Chinmay Maheshwari, Eric Mazumdar, Shankar Sastry

Figure 1 for Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets
Figure 2 for Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets
Figure 3 for Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets
Figure 4 for Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets
Viaarxiv icon

Who Leads and Who Follows in Strategic Classification?

Add code
Bookmark button
Alert button
Jun 23, 2021
Tijana Zrnic, Eric Mazumdar, S. Shankar Sastry, Michael I. Jordan

Figure 1 for Who Leads and Who Follows in Strategic Classification?
Figure 2 for Who Leads and Who Follows in Strategic Classification?
Figure 3 for Who Leads and Who Follows in Strategic Classification?
Figure 4 for Who Leads and Who Follows in Strategic Classification?
Viaarxiv icon

Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization

Add code
Bookmark button
Alert button
Jun 16, 2021
Chinmay Maheshwari, Chih-Yuan Chiu, Eric Mazumdar, S. Shankar Sastry, Lillian J. Ratliff

Figure 1 for Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization
Figure 2 for Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization
Figure 3 for Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization
Figure 4 for Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization
Viaarxiv icon

Fast Distributionally Robust Learning with Variance Reduced Min-Max Optimization

Add code
Bookmark button
Alert button
Apr 27, 2021
Yaodong Yu, Tianyi Lin, Eric Mazumdar, Michael I. Jordan

Figure 1 for Fast Distributionally Robust Learning with Variance Reduced Min-Max Optimization
Figure 2 for Fast Distributionally Robust Learning with Variance Reduced Min-Max Optimization
Figure 3 for Fast Distributionally Robust Learning with Variance Reduced Min-Max Optimization
Figure 4 for Fast Distributionally Robust Learning with Variance Reduced Min-Max Optimization
Viaarxiv icon

Expert Selection in High-Dimensional Markov Decision Processes

Add code
Bookmark button
Alert button
Oct 26, 2020
Vicenc Rubies-Royo, Eric Mazumdar, Roy Dong, Claire Tomlin, S. Shankar Sastry

Figure 1 for Expert Selection in High-Dimensional Markov Decision Processes
Figure 2 for Expert Selection in High-Dimensional Markov Decision Processes
Figure 3 for Expert Selection in High-Dimensional Markov Decision Processes
Figure 4 for Expert Selection in High-Dimensional Markov Decision Processes
Viaarxiv icon

Technical Report: Adaptive Control for Linearizable Systems Using On-Policy Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 06, 2020
Tyler Westenbroek, Eric Mazumdar, David Fridovich-Keil, Valmik Prabhu, Claire J. Tomlin, S. Shankar Sastry

Figure 1 for Technical Report: Adaptive Control for Linearizable Systems Using On-Policy Reinforcement Learning
Viaarxiv icon

On Thompson Sampling with Langevin Algorithms

Add code
Bookmark button
Alert button
Feb 23, 2020
Eric Mazumdar, Aldo Pacchiano, Yi-an Ma, Peter L. Bartlett, Michael I. Jordan

Figure 1 for On Thompson Sampling with Langevin Algorithms
Viaarxiv icon

Feedback Linearization for Unknown Systems via Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 29, 2019
Tyler Westenbroek, David Fridovich-Keil, Eric Mazumdar, Shreyas Arora, Valmik Prabhu, S. Shankar Sastry, Claire J. Tomlin

Figure 1 for Feedback Linearization for Unknown Systems via Reinforcement Learning
Figure 2 for Feedback Linearization for Unknown Systems via Reinforcement Learning
Figure 3 for Feedback Linearization for Unknown Systems via Reinforcement Learning
Figure 4 for Feedback Linearization for Unknown Systems via Reinforcement Learning
Viaarxiv icon

Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings

Add code
Bookmark button
Alert button
Jul 08, 2019
Eric Mazumdar, Lillian J. Ratliff, Michael I. Jordan, S. Shankar Sastry

Figure 1 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Figure 2 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Figure 3 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Figure 4 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Viaarxiv icon

Convergence Analysis of Gradient-Based Learning with Non-Uniform Learning Rates in Non-Cooperative Multi-Agent Settings

Add code
Bookmark button
Alert button
May 30, 2019
Benjamin Chasnov, Lillian J. Ratliff, Eric Mazumdar, Samuel A. Burden

Figure 1 for Convergence Analysis of Gradient-Based Learning with Non-Uniform Learning Rates in Non-Cooperative Multi-Agent Settings
Figure 2 for Convergence Analysis of Gradient-Based Learning with Non-Uniform Learning Rates in Non-Cooperative Multi-Agent Settings
Figure 3 for Convergence Analysis of Gradient-Based Learning with Non-Uniform Learning Rates in Non-Cooperative Multi-Agent Settings
Figure 4 for Convergence Analysis of Gradient-Based Learning with Non-Uniform Learning Rates in Non-Cooperative Multi-Agent Settings
Viaarxiv icon