Alert button
Picture for Dale Schuurmans

Dale Schuurmans

Alert button

Kernel Exponential Family Estimation via Doubly Dual Embedding

Add code
Bookmark button
Alert button
Nov 06, 2018
Bo Dai, Hanjun Dai, Arthur Gretton, Le Song, Dale Schuurmans, Niao He

Figure 1 for Kernel Exponential Family Estimation via Doubly Dual Embedding
Figure 2 for Kernel Exponential Family Estimation via Doubly Dual Embedding
Figure 3 for Kernel Exponential Family Estimation via Doubly Dual Embedding
Figure 4 for Kernel Exponential Family Estimation via Doubly Dual Embedding
Viaarxiv icon

Smoothed Action Value Functions for Learning Gaussian Policies

Add code
Bookmark button
Alert button
Jul 25, 2018
Ofir Nachum, Mohammad Norouzi, George Tucker, Dale Schuurmans

Figure 1 for Smoothed Action Value Functions for Learning Gaussian Policies
Viaarxiv icon

Planning and Learning with Stochastic Action Sets

Add code
Bookmark button
Alert button
May 07, 2018
Craig Boutilier, Alon Cohen, Amit Daniely, Avinatan Hassidim, Yishay Mansour, Ofer Meshi, Martin Mladenov, Dale Schuurmans

Figure 1 for Planning and Learning with Stochastic Action Sets
Figure 2 for Planning and Learning with Stochastic Action Sets
Viaarxiv icon

Variational Rejection Sampling

Add code
Bookmark button
Alert button
Apr 05, 2018
Aditya Grover, Ramki Gummadi, Miguel Lazaro-Gredilla, Dale Schuurmans, Stefano Ermon

Figure 1 for Variational Rejection Sampling
Figure 2 for Variational Rejection Sampling
Figure 3 for Variational Rejection Sampling
Figure 4 for Variational Rejection Sampling
Viaarxiv icon

Trust-PCL: An Off-Policy Trust Region Method for Continuous Control

Add code
Bookmark button
Alert button
Feb 22, 2018
Ofir Nachum, Mohammad Norouzi, Kelvin Xu, Dale Schuurmans

Figure 1 for Trust-PCL: An Off-Policy Trust Region Method for Continuous Control
Figure 2 for Trust-PCL: An Off-Policy Trust Region Method for Continuous Control
Figure 3 for Trust-PCL: An Off-Policy Trust Region Method for Continuous Control
Figure 4 for Trust-PCL: An Off-Policy Trust Region Method for Continuous Control
Viaarxiv icon

Safe Exploration for Identifying Linear Systems via Robust Optimization

Add code
Bookmark button
Alert button
Nov 30, 2017
Tyler Lu, Martin Zinkevich, Craig Boutilier, Binz Roy, Dale Schuurmans

Figure 1 for Safe Exploration for Identifying Linear Systems via Robust Optimization
Figure 2 for Safe Exploration for Identifying Linear Systems via Robust Optimization
Figure 3 for Safe Exploration for Identifying Linear Systems via Robust Optimization
Viaarxiv icon

Bridging the Gap Between Value and Policy Based Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 22, 2017
Ofir Nachum, Mohammad Norouzi, Kelvin Xu, Dale Schuurmans

Figure 1 for Bridging the Gap Between Value and Policy Based Reinforcement Learning
Figure 2 for Bridging the Gap Between Value and Policy Based Reinforcement Learning
Viaarxiv icon

Improving Policy Gradient by Exploring Under-appreciated Rewards

Add code
Bookmark button
Alert button
Mar 15, 2017
Ofir Nachum, Mohammad Norouzi, Dale Schuurmans

Figure 1 for Improving Policy Gradient by Exploring Under-appreciated Rewards
Figure 2 for Improving Policy Gradient by Exploring Under-appreciated Rewards
Figure 3 for Improving Policy Gradient by Exploring Under-appreciated Rewards
Figure 4 for Improving Policy Gradient by Exploring Under-appreciated Rewards
Viaarxiv icon

Reward Augmented Maximum Likelihood for Neural Structured Prediction

Add code
Bookmark button
Alert button
Jan 04, 2017
Mohammad Norouzi, Samy Bengio, Zhifeng Chen, Navdeep Jaitly, Mike Schuster, Yonghui Wu, Dale Schuurmans

Figure 1 for Reward Augmented Maximum Likelihood for Neural Structured Prediction
Figure 2 for Reward Augmented Maximum Likelihood for Neural Structured Prediction
Figure 3 for Reward Augmented Maximum Likelihood for Neural Structured Prediction
Viaarxiv icon

Stochastic Neural Networks with Monotonic Activation Functions

Add code
Bookmark button
Alert button
Jul 22, 2016
Siamak Ravanbakhsh, Barnabas Poczos, Jeff Schneider, Dale Schuurmans, Russell Greiner

Figure 1 for Stochastic Neural Networks with Monotonic Activation Functions
Figure 2 for Stochastic Neural Networks with Monotonic Activation Functions
Figure 3 for Stochastic Neural Networks with Monotonic Activation Functions
Figure 4 for Stochastic Neural Networks with Monotonic Activation Functions
Viaarxiv icon