Picture for Mohammad Norouzi

Mohammad Norouzi

Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse

Add code
Nov 06, 2019
Figure 1 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Figure 2 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Figure 3 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Figure 4 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Viaarxiv icon

Efficient Exploration with Self-Imitation Learning via Trajectory-Conditioned Policy

Add code
Jul 24, 2019
Figure 1 for Efficient Exploration with Self-Imitation Learning via Trajectory-Conditioned Policy
Figure 2 for Efficient Exploration with Self-Imitation Learning via Trajectory-Conditioned Policy
Figure 3 for Efficient Exploration with Self-Imitation Learning via Trajectory-Conditioned Policy
Figure 4 for Efficient Exploration with Self-Imitation Learning via Trajectory-Conditioned Policy
Viaarxiv icon

Striving for Simplicity in Off-policy Deep Reinforcement Learning

Add code
Jul 10, 2019
Figure 1 for Striving for Simplicity in Off-policy Deep Reinforcement Learning
Figure 2 for Striving for Simplicity in Off-policy Deep Reinforcement Learning
Figure 3 for Striving for Simplicity in Off-policy Deep Reinforcement Learning
Figure 4 for Striving for Simplicity in Off-policy Deep Reinforcement Learning
Viaarxiv icon

Similarity of Neural Network Representations Revisited

Add code
May 14, 2019
Figure 1 for Similarity of Neural Network Representations Revisited
Figure 2 for Similarity of Neural Network Representations Revisited
Figure 3 for Similarity of Neural Network Representations Revisited
Figure 4 for Similarity of Neural Network Representations Revisited
Viaarxiv icon

Learning to Generalize from Sparse and Underspecified Rewards

Add code
Feb 19, 2019
Figure 1 for Learning to Generalize from Sparse and Underspecified Rewards
Figure 2 for Learning to Generalize from Sparse and Underspecified Rewards
Figure 3 for Learning to Generalize from Sparse and Underspecified Rewards
Figure 4 for Learning to Generalize from Sparse and Underspecified Rewards
Viaarxiv icon

Understanding the impact of entropy on policy optimization

Add code
Nov 29, 2018
Figure 1 for Understanding the impact of entropy on policy optimization
Figure 2 for Understanding the impact of entropy on policy optimization
Figure 3 for Understanding the impact of entropy on policy optimization
Figure 4 for Understanding the impact of entropy on policy optimization
Viaarxiv icon

Contingency-Aware Exploration in Reinforcement Learning

Add code
Nov 05, 2018
Figure 1 for Contingency-Aware Exploration in Reinforcement Learning
Figure 2 for Contingency-Aware Exploration in Reinforcement Learning
Figure 3 for Contingency-Aware Exploration in Reinforcement Learning
Figure 4 for Contingency-Aware Exploration in Reinforcement Learning
Viaarxiv icon

Memory Augmented Policy Optimization for Program Synthesis and Semantic Parsing

Add code
Oct 31, 2018
Figure 1 for Memory Augmented Policy Optimization for Program Synthesis and Semantic Parsing
Figure 2 for Memory Augmented Policy Optimization for Program Synthesis and Semantic Parsing
Figure 3 for Memory Augmented Policy Optimization for Program Synthesis and Semantic Parsing
Figure 4 for Memory Augmented Policy Optimization for Program Synthesis and Semantic Parsing
Viaarxiv icon

Sequence to Sequence Mixture Model for Diverse Machine Translation

Add code
Oct 17, 2018
Figure 1 for Sequence to Sequence Mixture Model for Diverse Machine Translation
Figure 2 for Sequence to Sequence Mixture Model for Diverse Machine Translation
Figure 3 for Sequence to Sequence Mixture Model for Diverse Machine Translation
Figure 4 for Sequence to Sequence Mixture Model for Diverse Machine Translation
Viaarxiv icon

Optimal Completion Distillation for Sequence Learning

Add code
Oct 02, 2018
Figure 1 for Optimal Completion Distillation for Sequence Learning
Figure 2 for Optimal Completion Distillation for Sequence Learning
Figure 3 for Optimal Completion Distillation for Sequence Learning
Figure 4 for Optimal Completion Distillation for Sequence Learning
Viaarxiv icon