Alert button
Picture for Shie Mannor

Shie Mannor

Alert button

Faculty of Electrical Engineering, Technion, Israel Institute of Technology

Inspiration Learning through Preferences

Add code
Bookmark button
Alert button
Sep 16, 2018
Nir Baram, Shie Mannor

Figure 1 for Inspiration Learning through Preferences
Figure 2 for Inspiration Learning through Preferences
Figure 3 for Inspiration Learning through Preferences
Figure 4 for Inspiration Learning through Preferences
Viaarxiv icon

On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters

Add code
Bookmark button
Alert button
Sep 16, 2018
Mark Kozdoba, Jakub Marecek, Tigran Tchrakian, Shie Mannor

Figure 1 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Figure 2 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Figure 3 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Figure 4 for On-Line Learning of Linear Dynamical Systems: Exponential Forgetting in Kalman Filters
Viaarxiv icon

Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 06, 2018
Tom Zahavy, Matan Haroush, Nadav Merlis, Daniel J. Mankowitz, Shie Mannor

Figure 1 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Figure 2 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Figure 3 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Figure 4 for Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning
Viaarxiv icon

How to Combine Tree-Search Methods in Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 06, 2018
Yonathan Efroni, Gal Dalal, Bruno Scherrer, Shie Mannor

Figure 1 for How to Combine Tree-Search Methods in Reinforcement Learning
Figure 2 for How to Combine Tree-Search Methods in Reinforcement Learning
Figure 3 for How to Combine Tree-Search Methods in Reinforcement Learning
Figure 4 for How to Combine Tree-Search Methods in Reinforcement Learning
Viaarxiv icon

Multi-user Communication Networks: A Coordinated Multi-armed Bandit Approach

Add code
Bookmark button
Alert button
Aug 14, 2018
Orly Avner, Shie Mannor

Figure 1 for Multi-user Communication Networks: A Coordinated Multi-armed Bandit Approach
Figure 2 for Multi-user Communication Networks: A Coordinated Multi-armed Bandit Approach
Figure 3 for Multi-user Communication Networks: A Coordinated Multi-armed Bandit Approach
Figure 4 for Multi-user Communication Networks: A Coordinated Multi-armed Bandit Approach
Viaarxiv icon

Beyond the One Step Greedy Approach in Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 30, 2018
Yonathan Efroni, Gal Dalal, Bruno Scherrer, Shie Mannor

Figure 1 for Beyond the One Step Greedy Approach in Reinforcement Learning
Viaarxiv icon

A General Approach to Multi-Armed Bandits Under Risk Criteria

Add code
Bookmark button
Alert button
Jun 04, 2018
Asaf Cassel, Shie Mannor, Assaf Zeevi

Figure 1 for A General Approach to Multi-Armed Bandits Under Risk Criteria
Viaarxiv icon

Finite Sample Analysis of Two-Timescale Stochastic Approximation with Applications to Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 04, 2018
Gal Dalal, Balazs Szorenyi, Gugan Thoppe, Shie Mannor

Figure 1 for Finite Sample Analysis of Two-Timescale Stochastic Approximation with Applications to Reinforcement Learning
Figure 2 for Finite Sample Analysis of Two-Timescale Stochastic Approximation with Applications to Reinforcement Learning
Figure 3 for Finite Sample Analysis of Two-Timescale Stochastic Approximation with Applications to Reinforcement Learning
Figure 4 for Finite Sample Analysis of Two-Timescale Stochastic Approximation with Applications to Reinforcement Learning
Viaarxiv icon

Reward Constrained Policy Optimization

Add code
Bookmark button
Alert button
May 28, 2018
Chen Tessler, Daniel J. Mankowitz, Shie Mannor

Figure 1 for Reward Constrained Policy Optimization
Figure 2 for Reward Constrained Policy Optimization
Figure 3 for Reward Constrained Policy Optimization
Figure 4 for Reward Constrained Policy Optimization
Viaarxiv icon

Nonlinear Distributional Gradient Temporal-Difference Learning

Add code
Bookmark button
Alert button
May 20, 2018
Chao Qu, Shie Mannor, Huan Xu

Figure 1 for Nonlinear Distributional Gradient Temporal-Difference Learning
Figure 2 for Nonlinear Distributional Gradient Temporal-Difference Learning
Viaarxiv icon