Picture for Sergey Levine

Sergey Levine

Stanford University

Learning to Identify Object Instances by Touch: Tactile Recognition via Multimodal Matching

Add code
Mar 08, 2019
Figure 1 for Learning to Identify Object Instances by Touch: Tactile Recognition via Multimodal Matching
Figure 2 for Learning to Identify Object Instances by Touch: Tactile Recognition via Multimodal Matching
Figure 3 for Learning to Identify Object Instances by Touch: Tactile Recognition via Multimodal Matching
Figure 4 for Learning to Identify Object Instances by Touch: Tactile Recognition via Multimodal Matching
Viaarxiv icon

Learning Latent Plans from Play

Add code
Mar 05, 2019
Figure 1 for Learning Latent Plans from Play
Figure 2 for Learning Latent Plans from Play
Figure 3 for Learning Latent Plans from Play
Figure 4 for Learning Latent Plans from Play
Viaarxiv icon

Model-Based Reinforcement Learning for Atari

Add code
Mar 05, 2019
Figure 1 for Model-Based Reinforcement Learning for Atari
Figure 2 for Model-Based Reinforcement Learning for Atari
Figure 3 for Model-Based Reinforcement Learning for Atari
Figure 4 for Model-Based Reinforcement Learning for Atari
Viaarxiv icon

VideoFlow: A Flow-Based Generative Model for Video

Add code
Mar 04, 2019
Figure 1 for VideoFlow: A Flow-Based Generative Model for Video
Figure 2 for VideoFlow: A Flow-Based Generative Model for Video
Figure 3 for VideoFlow: A Flow-Based Generative Model for Video
Figure 4 for VideoFlow: A Flow-Based Generative Model for Video
Viaarxiv icon

Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning

Add code
Feb 27, 2019
Figure 1 for Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning
Figure 2 for Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning
Figure 3 for Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning
Figure 4 for Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning
Viaarxiv icon

Diagnosing Bottlenecks in Deep Q-learning Algorithms

Add code
Feb 26, 2019
Figure 1 for Diagnosing Bottlenecks in Deep Q-learning Algorithms
Figure 2 for Diagnosing Bottlenecks in Deep Q-learning Algorithms
Figure 3 for Diagnosing Bottlenecks in Deep Q-learning Algorithms
Figure 4 for Diagnosing Bottlenecks in Deep Q-learning Algorithms
Viaarxiv icon

Online Meta-Learning

Add code
Feb 22, 2019
Figure 1 for Online Meta-Learning
Figure 2 for Online Meta-Learning
Figure 3 for Online Meta-Learning
Figure 4 for Online Meta-Learning
Viaarxiv icon

SOLAR: Deep Structured Representations for Model-Based Reinforcement Learning

Add code
Feb 20, 2019
Figure 1 for SOLAR: Deep Structured Representations for Model-Based Reinforcement Learning
Figure 2 for SOLAR: Deep Structured Representations for Model-Based Reinforcement Learning
Figure 3 for SOLAR: Deep Structured Representations for Model-Based Reinforcement Learning
Figure 4 for SOLAR: Deep Structured Representations for Model-Based Reinforcement Learning
Viaarxiv icon

From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following

Add code
Feb 20, 2019
Figure 1 for From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following
Figure 2 for From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following
Figure 3 for From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following
Figure 4 for From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following
Viaarxiv icon

Hierarchical Policy Design for Sample-Efficient Learning of Robot Table Tennis Through Self-Play

Add code
Feb 17, 2019
Figure 1 for Hierarchical Policy Design for Sample-Efficient Learning of Robot Table Tennis Through Self-Play
Figure 2 for Hierarchical Policy Design for Sample-Efficient Learning of Robot Table Tennis Through Self-Play
Figure 3 for Hierarchical Policy Design for Sample-Efficient Learning of Robot Table Tennis Through Self-Play
Figure 4 for Hierarchical Policy Design for Sample-Efficient Learning of Robot Table Tennis Through Self-Play
Viaarxiv icon