Alert button
Picture for R. Srikant

R. Srikant

Alert button

Improved Algorithms for Misspecified Linear Markov Decision Processes

Add code
Bookmark button
Alert button
Sep 12, 2021
Daniel Vial, Advait Parulekar, Sanjay Shakkottai, R. Srikant

Figure 1 for Improved Algorithms for Misspecified Linear Markov Decision Processes
Viaarxiv icon

Linear Convergence of Entropy-Regularized Natural Policy Gradient with Linear Function Approximation

Add code
Bookmark button
Alert button
Jun 08, 2021
Semih Cayci, Niao He, R. Srikant

Figure 1 for Linear Convergence of Entropy-Regularized Natural Policy Gradient with Linear Function Approximation
Viaarxiv icon

Regret Bounds for Stochastic Shortest Path Problems with Linear Function Approximation

Add code
Bookmark button
Alert button
May 04, 2021
Daniel Vial, Advait Parulekar, Sanjay Shakkottai, R. Srikant

Figure 1 for Regret Bounds for Stochastic Shortest Path Problems with Linear Function Approximation
Figure 2 for Regret Bounds for Stochastic Shortest Path Problems with Linear Function Approximation
Viaarxiv icon

Achieving Small Test Error in Mildly Overparameterized Neural Networks

Add code
Bookmark button
Alert button
Apr 24, 2021
Shiyu Liang, Ruoyu Sun, R. Srikant

Viaarxiv icon

Sample Complexity and Overparameterization Bounds for Projection-Free Neural TD Learning

Add code
Bookmark button
Alert button
Mar 02, 2021
Semih Cayci, Siddhartha Satpathi, Niao He, R. Srikant

Figure 1 for Sample Complexity and Overparameterization Bounds for Projection-Free Neural TD Learning
Viaarxiv icon

Optimistic Policy Iteration for MDPs with Acyclic Transient State Structure

Add code
Bookmark button
Alert button
Feb 13, 2021
Joseph Lubars, Anna Winnicki, Michael Livesay, R. Srikant

Figure 1 for Optimistic Policy Iteration for MDPs with Acyclic Transient State Structure
Figure 2 for Optimistic Policy Iteration for MDPs with Acyclic Transient State Structure
Figure 3 for Optimistic Policy Iteration for MDPs with Acyclic Transient State Structure
Figure 4 for Optimistic Policy Iteration for MDPs with Acyclic Transient State Structure
Viaarxiv icon

One-bit feedback is sufficient for upper confidence bound policies

Add code
Bookmark button
Alert button
Dec 04, 2020
Daniel Vial, Sanjay Shakkottai, R. Srikant

Figure 1 for One-bit feedback is sufficient for upper confidence bound policies
Figure 2 for One-bit feedback is sufficient for upper confidence bound policies
Figure 3 for One-bit feedback is sufficient for upper confidence bound policies
Viaarxiv icon

Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging

Add code
Bookmark button
Alert button
Nov 17, 2020
Joseph Lubars, Harsh Gupta, Adnan Raja, R. Srikant, Liyun Li, Xinzhou Wu

Figure 1 for Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging
Figure 2 for Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging
Figure 3 for Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging
Figure 4 for Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging
Viaarxiv icon

On the Consistency of Maximum Likelihood Estimators for Causal Network Identification

Add code
Bookmark button
Alert button
Oct 17, 2020
Xiaotian Xie, Dimitrios Katselis, Carolyn L. Beck, R. Srikant

Figure 1 for On the Consistency of Maximum Likelihood Estimators for Causal Network Identification
Figure 2 for On the Consistency of Maximum Likelihood Estimators for Causal Network Identification
Figure 3 for On the Consistency of Maximum Likelihood Estimators for Causal Network Identification
Viaarxiv icon