Picture for Shie Mannor

Shie Mannor

Faculty of Electrical Engineering, Technion, Israel Institute of Technology

Deep Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood Matching

Add code
Jan 24, 2019
Figure 1 for Deep Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood Matching
Figure 2 for Deep Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood Matching
Figure 3 for Deep Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood Matching
Figure 4 for Deep Neural Linear Bandits: Overcoming Catastrophic Forgetting through Likelihood Matching
Viaarxiv icon

Trust Region Value Optimization using Kalman Filtering

Add code
Jan 23, 2019
Figure 1 for Trust Region Value Optimization using Kalman Filtering
Figure 2 for Trust Region Value Optimization using Kalman Filtering
Figure 3 for Trust Region Value Optimization using Kalman Filtering
Figure 4 for Trust Region Value Optimization using Kalman Filtering
Viaarxiv icon

Multi Instance Learning For Unbalanced Data

Add code
Dec 17, 2018
Figure 1 for Multi Instance Learning For Unbalanced Data
Figure 2 for Multi Instance Learning For Unbalanced Data
Figure 3 for Multi Instance Learning For Unbalanced Data
Figure 4 for Multi Instance Learning For Unbalanced Data
Viaarxiv icon

Revisiting Exploration-Conscious Reinforcement Learning

Add code
Dec 13, 2018
Figure 1 for Revisiting Exploration-Conscious Reinforcement Learning
Figure 2 for Revisiting Exploration-Conscious Reinforcement Learning
Figure 3 for Revisiting Exploration-Conscious Reinforcement Learning
Figure 4 for Revisiting Exploration-Conscious Reinforcement Learning
Viaarxiv icon

Soft-Robust Actor-Critic Policy-Gradient

Add code
Oct 24, 2018
Figure 1 for Soft-Robust Actor-Critic Policy-Gradient
Figure 2 for Soft-Robust Actor-Critic Policy-Gradient
Viaarxiv icon

Multiple-Step Greedy Policies in Online and Approximate Reinforcement Learning

Add code
Sep 20, 2018
Figure 1 for Multiple-Step Greedy Policies in Online and Approximate Reinforcement Learning
Viaarxiv icon

Inspiration Learning through Preferences

Add code
Sep 16, 2018
Figure 1 for Inspiration Learning through Preferences
Figure 2 for Inspiration Learning through Preferences
Figure 3 for Inspiration Learning through Preferences
Figure 4 for Inspiration Learning through Preferences
Viaarxiv icon

On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters

Add code
Sep 16, 2018
Figure 1 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Figure 2 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Figure 3 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Figure 4 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Viaarxiv icon

Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning

Add code
Sep 06, 2018
Figure 1 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Figure 2 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Figure 3 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Figure 4 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Viaarxiv icon

How to Combine Tree-Search Methods in Reinforcement Learning

Add code
Sep 06, 2018
Figure 1 for How to Combine Tree-Search Methods in Reinforcement Learning
Figure 2 for How to Combine Tree-Search Methods in Reinforcement Learning
Figure 3 for How to Combine Tree-Search Methods in Reinforcement Learning
Figure 4 for How to Combine Tree-Search Methods in Reinforcement Learning
Viaarxiv icon