Picture for Sergey Levine

Sergey Levine

Stanford University

First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization

Add code
May 24, 2022
Figure 1 for First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization
Figure 2 for First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization
Figure 3 for First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization
Figure 4 for First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization
Viaarxiv icon

Planning with Diffusion for Flexible Behavior Synthesis

Add code
May 20, 2022
Figure 1 for Planning with Diffusion for Flexible Behavior Synthesis
Figure 2 for Planning with Diffusion for Flexible Behavior Synthesis
Figure 3 for Planning with Diffusion for Flexible Behavior Synthesis
Figure 4 for Planning with Diffusion for Flexible Behavior Synthesis
Viaarxiv icon

Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space

Add code
May 17, 2022
Figure 1 for Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space
Figure 2 for Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space
Figure 3 for Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space
Figure 4 for Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space
Viaarxiv icon

ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters

Add code
May 05, 2022
Figure 1 for ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters
Figure 2 for ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters
Figure 3 for ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters
Figure 4 for ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters
Viaarxiv icon

Control-Aware Prediction Objectives for Autonomous Driving

Add code
Apr 28, 2022
Figure 1 for Control-Aware Prediction Objectives for Autonomous Driving
Figure 2 for Control-Aware Prediction Objectives for Autonomous Driving
Figure 3 for Control-Aware Prediction Objectives for Autonomous Driving
Figure 4 for Control-Aware Prediction Objectives for Autonomous Driving
Viaarxiv icon

Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning

Add code
Apr 28, 2022
Figure 1 for Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning
Figure 2 for Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning
Figure 3 for Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning
Figure 4 for Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning
Viaarxiv icon

Context-Aware Language Modeling for Goal-Oriented Dialogue Systems

Add code
Apr 22, 2022
Figure 1 for Context-Aware Language Modeling for Goal-Oriented Dialogue Systems
Figure 2 for Context-Aware Language Modeling for Goal-Oriented Dialogue Systems
Figure 3 for Context-Aware Language Modeling for Goal-Oriented Dialogue Systems
Figure 4 for Context-Aware Language Modeling for Goal-Oriented Dialogue Systems
Viaarxiv icon

INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL

Add code
Apr 18, 2022
Figure 1 for INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL
Figure 2 for INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL
Figure 3 for INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL
Figure 4 for INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL
Viaarxiv icon

CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning

Add code
Apr 18, 2022
Figure 1 for CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning
Figure 2 for CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning
Figure 3 for CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning
Figure 4 for CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning
Viaarxiv icon

When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?

Add code
Apr 12, 2022
Figure 1 for When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?
Figure 2 for When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?
Figure 3 for When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?
Figure 4 for When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?
Viaarxiv icon