Alert button
Picture for Mohammad Ghavamzadeh

Mohammad Ghavamzadeh

Alert button

Garbage In, Reward Out: Bootstrapping Exploration in Multi-Armed Bandits

Add code
Bookmark button
Alert button
Nov 13, 2018
Branislav Kveton, Csaba Szepesvari, Zheng Wen, Mohammad Ghavamzadeh, Tor Lattimore

Figure 1 for Garbage In, Reward Out: Bootstrapping Exploration in Multi-Armed Bandits
Figure 2 for Garbage In, Reward Out: Bootstrapping Exploration in Multi-Armed Bandits
Viaarxiv icon

A Block Coordinate Ascent Algorithm for Mean-Variance Optimization

Add code
Bookmark button
Alert button
Nov 01, 2018
Bo Liu, Tengyang Xie, Yangyang Xu, Mohammad Ghavamzadeh, Yinlam Chow, Daoming Lyu, Daesub Yoon

Figure 1 for A Block Coordinate Ascent Algorithm for Mean-Variance Optimization
Figure 2 for A Block Coordinate Ascent Algorithm for Mean-Variance Optimization
Viaarxiv icon

Risk-Sensitive Generative Adversarial Imitation Learning

Add code
Bookmark button
Alert button
Aug 13, 2018
Jonathan Lacotte, Yinlam Chow, Mohammad Ghavamzadeh, Marco Pavone

Figure 1 for Risk-Sensitive Generative Adversarial Imitation Learning
Figure 2 for Risk-Sensitive Generative Adversarial Imitation Learning
Figure 3 for Risk-Sensitive Generative Adversarial Imitation Learning
Viaarxiv icon

Model-Independent Online Learning for Influence Maximization

Add code
Bookmark button
Alert button
May 24, 2018
Sharan Vaswani, Branislav Kveton, Zheng Wen, Mohammad Ghavamzadeh, Laks Lakshmanan, Mark Schmidt

Figure 1 for Model-Independent Online Learning for Influence Maximization
Figure 2 for Model-Independent Online Learning for Influence Maximization
Figure 3 for Model-Independent Online Learning for Influence Maximization
Viaarxiv icon

More Robust Doubly Robust Off-policy Evaluation

Add code
Bookmark button
Alert button
May 23, 2018
Mehrdad Farajtabar, Yinlam Chow, Mohammad Ghavamzadeh

Figure 1 for More Robust Doubly Robust Off-policy Evaluation
Figure 2 for More Robust Doubly Robust Off-policy Evaluation
Figure 3 for More Robust Doubly Robust Off-policy Evaluation
Figure 4 for More Robust Doubly Robust Off-policy Evaluation
Viaarxiv icon

A Lyapunov-based Approach to Safe Reinforcement Learning

Add code
Bookmark button
Alert button
May 20, 2018
Yinlam Chow, Ofir Nachum, Edgar Duenez-Guzman, Mohammad Ghavamzadeh

Figure 1 for A Lyapunov-based Approach to Safe Reinforcement Learning
Figure 2 for A Lyapunov-based Approach to Safe Reinforcement Learning
Viaarxiv icon

Optimizing over a Restricted Policy Class in Markov Decision Processes

Add code
Bookmark button
Alert button
Feb 26, 2018
Ershad Banijamali, Yasin Abbasi-Yadkori, Mohammad Ghavamzadeh, Nikos Vlassis

Figure 1 for Optimizing over a Restricted Policy Class in Markov Decision Processes
Figure 2 for Optimizing over a Restricted Policy Class in Markov Decision Processes
Figure 3 for Optimizing over a Restricted Policy Class in Markov Decision Processes
Figure 4 for Optimizing over a Restricted Policy Class in Markov Decision Processes
Viaarxiv icon

Robust Locally-Linear Controllable Embedding

Add code
Bookmark button
Alert button
Feb 21, 2018
Ershad Banijamali, Rui Shu, Mohammad Ghavamzadeh, Hung Bui, Ali Ghodsi

Figure 1 for Robust Locally-Linear Controllable Embedding
Figure 2 for Robust Locally-Linear Controllable Embedding
Figure 3 for Robust Locally-Linear Controllable Embedding
Figure 4 for Robust Locally-Linear Controllable Embedding
Viaarxiv icon

Path Consistency Learning in Tsallis Entropy Regularized MDPs

Add code
Bookmark button
Alert button
Feb 10, 2018
Ofir Nachum, Yinlam Chow, Mohammad Ghavamzadeh

Figure 1 for Path Consistency Learning in Tsallis Entropy Regularized MDPs
Figure 2 for Path Consistency Learning in Tsallis Entropy Regularized MDPs
Figure 3 for Path Consistency Learning in Tsallis Entropy Regularized MDPs
Viaarxiv icon

Disentangling Dynamics and Content for Control and Planning

Add code
Bookmark button
Alert button
Nov 24, 2017
Ershad Banijamali, Ahmad Khajenezhad, Ali Ghodsi, Mohammad Ghavamzadeh

Figure 1 for Disentangling Dynamics and Content for Control and Planning
Figure 2 for Disentangling Dynamics and Content for Control and Planning
Figure 3 for Disentangling Dynamics and Content for Control and Planning
Viaarxiv icon