Alert button
Picture for Yinlam Chow

Yinlam Chow

Alert button

AlgaeDICE: Policy Gradient from Arbitrary Experience

Add code
Bookmark button
Alert button
Dec 04, 2019
Ofir Nachum, Bo Dai, Ilya Kostrikov, Yinlam Chow, Lihong Li, Dale Schuurmans

Figure 1 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Figure 2 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Figure 3 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Figure 4 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Viaarxiv icon

CAQL: Continuous Action Q-Learning

Add code
Bookmark button
Alert button
Oct 09, 2019
Moonkyung Ryu, Yinlam Chow, Ross Anderson, Christian Tjandraatmadja, Craig Boutilier

Figure 1 for CAQL: Continuous Action Q-Learning
Figure 2 for CAQL: Continuous Action Q-Learning
Figure 3 for CAQL: Continuous Action Q-Learning
Figure 4 for CAQL: Continuous Action Q-Learning
Viaarxiv icon

Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control

Add code
Bookmark button
Alert button
Sep 04, 2019
Nir Levine, Yinlam Chow, Rui Shu, Ang Li, Mohammad Ghavamzadeh, Hung Bui

Figure 1 for Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control
Figure 2 for Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control
Figure 3 for Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control
Figure 4 for Prediction, Consistency, Curvature: Representation Learning for Locally-Linear Control
Viaarxiv icon

DualDICE: Behavior-Agnostic Estimation of Discounted Stationary Distribution Corrections

Add code
Bookmark button
Alert button
Jun 10, 2019
Ofir Nachum, Yinlam Chow, Bo Dai, Lihong Li

Figure 1 for DualDICE: Behavior-Agnostic Estimation of Discounted Stationary Distribution Corrections
Figure 2 for DualDICE: Behavior-Agnostic Estimation of Discounted Stationary Distribution Corrections
Figure 3 for DualDICE: Behavior-Agnostic Estimation of Discounted Stationary Distribution Corrections
Figure 4 for DualDICE: Behavior-Agnostic Estimation of Discounted Stationary Distribution Corrections
Viaarxiv icon

Lyapunov-based Safe Policy Optimization for Continuous Control

Add code
Bookmark button
Alert button
Jan 28, 2019
Yinlam Chow, Ofir Nachum, Aleksandra Faust, Mohammad Ghavamzadeh, Edgar Duenez-Guzman

Figure 1 for Lyapunov-based Safe Policy Optimization for Continuous Control
Figure 2 for Lyapunov-based Safe Policy Optimization for Continuous Control
Figure 3 for Lyapunov-based Safe Policy Optimization for Continuous Control
Figure 4 for Lyapunov-based Safe Policy Optimization for Continuous Control
Viaarxiv icon

A Block Coordinate Ascent Algorithm for Mean-Variance Optimization

Add code
Bookmark button
Alert button
Nov 01, 2018
Bo Liu, Tengyang Xie, Yangyang Xu, Mohammad Ghavamzadeh, Yinlam Chow, Daoming Lyu, Daesub Yoon

Figure 1 for A Block Coordinate Ascent Algorithm for Mean-Variance Optimization
Figure 2 for A Block Coordinate Ascent Algorithm for Mean-Variance Optimization
Viaarxiv icon

Risk-Sensitive Generative Adversarial Imitation Learning

Add code
Bookmark button
Alert button
Aug 13, 2018
Jonathan Lacotte, Yinlam Chow, Mohammad Ghavamzadeh, Marco Pavone

Figure 1 for Risk-Sensitive Generative Adversarial Imitation Learning
Figure 2 for Risk-Sensitive Generative Adversarial Imitation Learning
Figure 3 for Risk-Sensitive Generative Adversarial Imitation Learning
Viaarxiv icon

More Robust Doubly Robust Off-policy Evaluation

Add code
Bookmark button
Alert button
May 23, 2018
Mehrdad Farajtabar, Yinlam Chow, Mohammad Ghavamzadeh

Figure 1 for More Robust Doubly Robust Off-policy Evaluation
Figure 2 for More Robust Doubly Robust Off-policy Evaluation
Figure 3 for More Robust Doubly Robust Off-policy Evaluation
Figure 4 for More Robust Doubly Robust Off-policy Evaluation
Viaarxiv icon

A Lyapunov-based Approach to Safe Reinforcement Learning

Add code
Bookmark button
Alert button
May 20, 2018
Yinlam Chow, Ofir Nachum, Edgar Duenez-Guzman, Mohammad Ghavamzadeh

Figure 1 for A Lyapunov-based Approach to Safe Reinforcement Learning
Figure 2 for A Lyapunov-based Approach to Safe Reinforcement Learning
Viaarxiv icon