Picture for Rami Al-Rfou

Rami Al-Rfou

ByT5: Towards a token-free future with pre-trained byte-to-byte models

Add code
May 28, 2021
Figure 1 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Figure 2 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Figure 3 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Figure 4 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Viaarxiv icon

The Power of Scale for Parameter-Efficient Prompt Tuning

Add code
Apr 18, 2021
Figure 1 for The Power of Scale for Parameter-Efficient Prompt Tuning
Figure 2 for The Power of Scale for Parameter-Efficient Prompt Tuning
Figure 3 for The Power of Scale for Parameter-Efficient Prompt Tuning
Figure 4 for The Power of Scale for Parameter-Efficient Prompt Tuning
Viaarxiv icon

Large Scale Knowledge Graph Based Synthetic Corpus Generation for Knowledge-Enhanced Language Model Pre-training

Add code
Oct 23, 2020
Figure 1 for Large Scale Knowledge Graph Based Synthetic Corpus Generation for Knowledge-Enhanced Language Model Pre-training
Figure 2 for Large Scale Knowledge Graph Based Synthetic Corpus Generation for Knowledge-Enhanced Language Model Pre-training
Figure 3 for Large Scale Knowledge Graph Based Synthetic Corpus Generation for Knowledge-Enhanced Language Model Pre-training
Figure 4 for Large Scale Knowledge Graph Based Synthetic Corpus Generation for Knowledge-Enhanced Language Model Pre-training
Viaarxiv icon

mT5: A massively multilingual pre-trained text-to-text transformer

Add code
Oct 23, 2020
Figure 1 for mT5: A massively multilingual pre-trained text-to-text transformer
Figure 2 for mT5: A massively multilingual pre-trained text-to-text transformer
Figure 3 for mT5: A massively multilingual pre-trained text-to-text transformer
Figure 4 for mT5: A massively multilingual pre-trained text-to-text transformer
Viaarxiv icon

LAReQA: Language-agnostic answer retrieval from a multilingual pool

Add code
Apr 11, 2020
Figure 1 for LAReQA: Language-agnostic answer retrieval from a multilingual pool
Figure 2 for LAReQA: Language-agnostic answer retrieval from a multilingual pool
Figure 3 for LAReQA: Language-agnostic answer retrieval from a multilingual pool
Figure 4 for LAReQA: Language-agnostic answer retrieval from a multilingual pool
Viaarxiv icon

Bridging the Gap for Tokenizer-Free Language Models

Add code
Aug 27, 2019
Figure 1 for Bridging the Gap for Tokenizer-Free Language Models
Figure 2 for Bridging the Gap for Tokenizer-Free Language Models
Figure 3 for Bridging the Gap for Tokenizer-Free Language Models
Figure 4 for Bridging the Gap for Tokenizer-Free Language Models
Viaarxiv icon

DDGK: Learning Graph Representations for Deep Divergence Graph Kernels

Add code
Apr 21, 2019
Figure 1 for DDGK: Learning Graph Representations for Deep Divergence Graph Kernels
Figure 2 for DDGK: Learning Graph Representations for Deep Divergence Graph Kernels
Figure 3 for DDGK: Learning Graph Representations for Deep Divergence Graph Kernels
Figure 4 for DDGK: Learning Graph Representations for Deep Divergence Graph Kernels
Viaarxiv icon

Watch Your Step: Learning Node Embeddings via Graph Attention

Add code
Sep 12, 2018
Figure 1 for Watch Your Step: Learning Node Embeddings via Graph Attention
Figure 2 for Watch Your Step: Learning Node Embeddings via Graph Attention
Figure 3 for Watch Your Step: Learning Node Embeddings via Graph Attention
Figure 4 for Watch Your Step: Learning Node Embeddings via Graph Attention
Viaarxiv icon

Creating Virtual Universes Using Generative Adversarial Networks

Add code
Aug 17, 2018
Figure 1 for Creating Virtual Universes Using Generative Adversarial Networks
Figure 2 for Creating Virtual Universes Using Generative Adversarial Networks
Figure 3 for Creating Virtual Universes Using Generative Adversarial Networks
Figure 4 for Creating Virtual Universes Using Generative Adversarial Networks
Viaarxiv icon

Character-Level Language Modeling with Deeper Self-Attention

Add code
Aug 09, 2018
Figure 1 for Character-Level Language Modeling with Deeper Self-Attention
Figure 2 for Character-Level Language Modeling with Deeper Self-Attention
Figure 3 for Character-Level Language Modeling with Deeper Self-Attention
Figure 4 for Character-Level Language Modeling with Deeper Self-Attention
Viaarxiv icon