Alert button
Picture for Michael Gimelfarb

Michael Gimelfarb

Alert button

Constraint-Generation Policy Optimization (CGPO): Nonlinear Programming for Policy Optimization in Mixed Discrete-Continuous MDPs

Add code
Bookmark button
Alert button
Jan 20, 2024
Michael Gimelfarb, Ayal Taitler, Scott Sanner

Viaarxiv icon

Thompson Sampling for Parameterized Markov Decision Processes with Uninformative Actions

Add code
Bookmark button
Alert button
May 13, 2023
Michael Gimelfarb, Michael Jong Kim

Viaarxiv icon

pyRDDLGym: From RDDL to Gym Environments

Add code
Bookmark button
Alert button
Nov 14, 2022
Ayal Taitler, Michael Gimelfarb, Sriram Gopalakrishnan, Xiaotian Liu, Scott Sanner

Figure 1 for pyRDDLGym: From RDDL to Gym Environments
Figure 2 for pyRDDLGym: From RDDL to Gym Environments
Figure 3 for pyRDDLGym: From RDDL to Gym Environments
Figure 4 for pyRDDLGym: From RDDL to Gym Environments
Viaarxiv icon

Conservative Bayesian Model-Based Value Expansion for Offline Policy Optimization

Add code
Bookmark button
Alert button
Oct 07, 2022
Jihwan Jeong, Xiaoyu Wang, Michael Gimelfarb, Hyunwoo Kim, Baher Abdulhai, Scott Sanner

Figure 1 for Conservative Bayesian Model-Based Value Expansion for Offline Policy Optimization
Figure 2 for Conservative Bayesian Model-Based Value Expansion for Offline Policy Optimization
Figure 3 for Conservative Bayesian Model-Based Value Expansion for Offline Policy Optimization
Figure 4 for Conservative Bayesian Model-Based Value Expansion for Offline Policy Optimization
Viaarxiv icon

RAPTOR: End-to-end Risk-Aware MDP Planning and Policy Learning by Backpropagation

Add code
Bookmark button
Alert button
Jun 14, 2021
Noah Patton, Jihwan Jeong, Michael Gimelfarb, Scott Sanner

Figure 1 for RAPTOR: End-to-end Risk-Aware MDP Planning and Policy Learning by Backpropagation
Figure 2 for RAPTOR: End-to-end Risk-Aware MDP Planning and Policy Learning by Backpropagation
Figure 3 for RAPTOR: End-to-end Risk-Aware MDP Planning and Policy Learning by Backpropagation
Figure 4 for RAPTOR: End-to-end Risk-Aware MDP Planning and Policy Learning by Backpropagation
Viaarxiv icon

Risk-Aware Transfer in Reinforcement Learning using Successor Features

Add code
Bookmark button
Alert button
May 28, 2021
Michael Gimelfarb, André Barreto, Scott Sanner, Chi-Guhn Lee

Figure 1 for Risk-Aware Transfer in Reinforcement Learning using Successor Features
Figure 2 for Risk-Aware Transfer in Reinforcement Learning using Successor Features
Figure 3 for Risk-Aware Transfer in Reinforcement Learning using Successor Features
Figure 4 for Risk-Aware Transfer in Reinforcement Learning using Successor Features
Viaarxiv icon

ε-BMC: A Bayesian Ensemble Approach to Epsilon-Greedy Exploration in Model-Free Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 02, 2020
Michael Gimelfarb, Scott Sanner, Chi-Guhn Lee

Figure 1 for ε-BMC: A Bayesian Ensemble Approach to Epsilon-Greedy Exploration in Model-Free Reinforcement Learning
Figure 2 for ε-BMC: A Bayesian Ensemble Approach to Epsilon-Greedy Exploration in Model-Free Reinforcement Learning
Figure 3 for ε-BMC: A Bayesian Ensemble Approach to Epsilon-Greedy Exploration in Model-Free Reinforcement Learning
Figure 4 for ε-BMC: A Bayesian Ensemble Approach to Epsilon-Greedy Exploration in Model-Free Reinforcement Learning
Viaarxiv icon

Bayesian Experience Reuse for Learning from Multiple Demonstrators

Add code
Bookmark button
Alert button
Jun 10, 2020
Michael Gimelfarb, Scott Sanner, Chi-Guhn Lee

Figure 1 for Bayesian Experience Reuse for Learning from Multiple Demonstrators
Figure 2 for Bayesian Experience Reuse for Learning from Multiple Demonstrators
Figure 3 for Bayesian Experience Reuse for Learning from Multiple Demonstrators
Figure 4 for Bayesian Experience Reuse for Learning from Multiple Demonstrators
Viaarxiv icon

Contextual Policy Reuse using Deep Mixture Models

Add code
Bookmark button
Alert button
Feb 29, 2020
Michael Gimelfarb, Scott Sanner, Chi-Guhn Lee

Figure 1 for Contextual Policy Reuse using Deep Mixture Models
Figure 2 for Contextual Policy Reuse using Deep Mixture Models
Figure 3 for Contextual Policy Reuse using Deep Mixture Models
Figure 4 for Contextual Policy Reuse using Deep Mixture Models
Viaarxiv icon