Picture for Quoc V. Le

Quoc V. Le

Attention Augmented Convolutional Networks

Add code
Apr 22, 2019
Figure 1 for Attention Augmented Convolutional Networks
Figure 2 for Attention Augmented Convolutional Networks
Figure 3 for Attention Augmented Convolutional Networks
Figure 4 for Attention Augmented Convolutional Networks
Viaarxiv icon

SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition

Add code
Apr 18, 2019
Figure 1 for SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition
Figure 2 for SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition
Figure 3 for SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition
Figure 4 for SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition
Viaarxiv icon

NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection

Add code
Apr 16, 2019
Figure 1 for NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection
Figure 2 for NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection
Figure 3 for NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection
Figure 4 for NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection
Viaarxiv icon

Soft Conditional Computation

Add code
Apr 10, 2019
Figure 1 for Soft Conditional Computation
Figure 2 for Soft Conditional Computation
Figure 3 for Soft Conditional Computation
Figure 4 for Soft Conditional Computation
Viaarxiv icon

The Evolved Transformer

Add code
Feb 15, 2019
Figure 1 for The Evolved Transformer
Figure 2 for The Evolved Transformer
Figure 3 for The Evolved Transformer
Figure 4 for The Evolved Transformer
Viaarxiv icon

Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context

Add code
Jan 18, 2019
Figure 1 for Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Figure 2 for Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Figure 3 for Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Figure 4 for Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Viaarxiv icon

GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

Add code
Dec 12, 2018
Figure 1 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 2 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 3 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 4 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Viaarxiv icon

Domain Adaptive Transfer Learning with Specialist Models

Add code
Dec 11, 2018
Figure 1 for Domain Adaptive Transfer Learning with Specialist Models
Figure 2 for Domain Adaptive Transfer Learning with Specialist Models
Figure 3 for Domain Adaptive Transfer Learning with Specialist Models
Figure 4 for Domain Adaptive Transfer Learning with Specialist Models
Viaarxiv icon

DropBlock: A regularization method for convolutional networks

Add code
Oct 30, 2018
Figure 1 for DropBlock: A regularization method for convolutional networks
Figure 2 for DropBlock: A regularization method for convolutional networks
Figure 3 for DropBlock: A regularization method for convolutional networks
Figure 4 for DropBlock: A regularization method for convolutional networks
Viaarxiv icon

Stochastic natural gradient descent draws posterior samples in function space

Add code
Oct 16, 2018
Figure 1 for Stochastic natural gradient descent draws posterior samples in function space
Figure 2 for Stochastic natural gradient descent draws posterior samples in function space
Figure 3 for Stochastic natural gradient descent draws posterior samples in function space
Figure 4 for Stochastic natural gradient descent draws posterior samples in function space
Viaarxiv icon