Picture for Michael I. Jordan

Michael I. Jordan

Provably Efficient Reinforcement Learning with Linear Function Approximation

Add code
Aug 08, 2019
Viaarxiv icon

Learning Stages: Phenomenon, Root Cause, Mechanism Hypothesis, and Implications

Add code
Aug 05, 2019
Figure 1 for Learning Stages: Phenomenon, Root Cause, Mechanism Hypothesis, and Implications
Figure 2 for Learning Stages: Phenomenon, Root Cause, Mechanism Hypothesis, and Implications
Figure 3 for Learning Stages: Phenomenon, Root Cause, Mechanism Hypothesis, and Implications
Figure 4 for Learning Stages: Phenomenon, Root Cause, Mechanism Hypothesis, and Implications
Viaarxiv icon

A Higher-Order Swiss Army Infinitesimal Jackknife

Add code
Jul 28, 2019
Figure 1 for A Higher-Order Swiss Army Infinitesimal Jackknife
Viaarxiv icon

Bayesian Robustness: A Nonasymptotic Viewpoint

Add code
Jul 27, 2019
Figure 1 for Bayesian Robustness: A Nonasymptotic Viewpoint
Figure 2 for Bayesian Robustness: A Nonasymptotic Viewpoint
Figure 3 for Bayesian Robustness: A Nonasymptotic Viewpoint
Figure 4 for Bayesian Robustness: A Nonasymptotic Viewpoint
Viaarxiv icon

Quantitative $W_1$ Convergence of Langevin-Like Stochastic Processes with Non-Convex Potential State-Dependent Noise

Add code
Jul 13, 2019
Figure 1 for Quantitative $W_1$ Convergence of Langevin-Like Stochastic Processes with Non-Convex Potential State-Dependent Noise
Figure 2 for Quantitative $W_1$ Convergence of Langevin-Like Stochastic Processes with Non-Convex Potential State-Dependent Noise
Figure 3 for Quantitative $W_1$ Convergence of Langevin-Like Stochastic Processes with Non-Convex Potential State-Dependent Noise
Figure 4 for Quantitative $W_1$ Convergence of Langevin-Like Stochastic Processes with Non-Convex Potential State-Dependent Noise
Viaarxiv icon

Convergence Rates for Gaussian Mixtures of Experts

Add code
Jul 09, 2019
Viaarxiv icon

Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings

Add code
Jul 08, 2019
Figure 1 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Figure 2 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Figure 3 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Figure 4 for Policy-Gradient Algorithms Have No Guarantees of Convergence in Continuous Action and State Multi-Agent Settings
Viaarxiv icon

Approximate Sherali-Adams Relaxations for MAP Inference via Entropy Regularization

Add code
Jul 02, 2019
Figure 1 for Approximate Sherali-Adams Relaxations for MAP Inference via Entropy Regularization
Figure 2 for Approximate Sherali-Adams Relaxations for MAP Inference via Entropy Regularization
Figure 3 for Approximate Sherali-Adams Relaxations for MAP Inference via Entropy Regularization
Figure 4 for Approximate Sherali-Adams Relaxations for MAP Inference via Entropy Regularization
Viaarxiv icon

Wasserstein Reinforcement Learning

Add code
Jun 19, 2019
Figure 1 for Wasserstein Reinforcement Learning
Figure 2 for Wasserstein Reinforcement Learning
Figure 3 for Wasserstein Reinforcement Learning
Figure 4 for Wasserstein Reinforcement Learning
Viaarxiv icon

Competing Bandits in Matching Markets

Add code
Jun 12, 2019
Figure 1 for Competing Bandits in Matching Markets
Figure 2 for Competing Bandits in Matching Markets
Viaarxiv icon