Picture for Adam Roberts

Adam Roberts

Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$

Add code
Mar 31, 2022
Figure 1 for Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$
Figure 2 for Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$
Viaarxiv icon

LaMDA: Language Models for Dialog Applications

Add code
Feb 10, 2022
Figure 1 for LaMDA: Language Models for Dialog Applications
Figure 2 for LaMDA: Language Models for Dialog Applications
Figure 3 for LaMDA: Language Models for Dialog Applications
Figure 4 for LaMDA: Language Models for Dialog Applications
Viaarxiv icon

ByT5: Towards a token-free future with pre-trained byte-to-byte models

Add code
May 28, 2021
Figure 1 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Figure 2 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Figure 3 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Figure 4 for ByT5: Towards a token-free future with pre-trained byte-to-byte models
Viaarxiv icon

Do Transformer Modifications Transfer Across Implementations and Applications?

Add code
Feb 23, 2021
Figure 1 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 2 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 3 for Do Transformer Modifications Transfer Across Implementations and Applications?
Viaarxiv icon

NeurIPS 2020 EfficientQA Competition: Systems, Analyses and Lessons Learned

Add code
Jan 01, 2021
Figure 1 for NeurIPS 2020 EfficientQA Competition: Systems, Analyses and Lessons Learned
Figure 2 for NeurIPS 2020 EfficientQA Competition: Systems, Analyses and Lessons Learned
Figure 3 for NeurIPS 2020 EfficientQA Competition: Systems, Analyses and Lessons Learned
Figure 4 for NeurIPS 2020 EfficientQA Competition: Systems, Analyses and Lessons Learned
Viaarxiv icon

Extracting Training Data from Large Language Models

Add code
Dec 14, 2020
Figure 1 for Extracting Training Data from Large Language Models
Figure 2 for Extracting Training Data from Large Language Models
Figure 3 for Extracting Training Data from Large Language Models
Figure 4 for Extracting Training Data from Large Language Models
Viaarxiv icon

mT5: A massively multilingual pre-trained text-to-text transformer

Add code
Oct 23, 2020
Figure 1 for mT5: A massively multilingual pre-trained text-to-text transformer
Figure 2 for mT5: A massively multilingual pre-trained text-to-text transformer
Figure 3 for mT5: A massively multilingual pre-trained text-to-text transformer
Figure 4 for mT5: A massively multilingual pre-trained text-to-text transformer
Viaarxiv icon

WT5?! Training Text-to-Text Models to Explain their Predictions

Add code
Apr 30, 2020
Figure 1 for WT5?! Training Text-to-Text Models to Explain their Predictions
Figure 2 for WT5?! Training Text-to-Text Models to Explain their Predictions
Figure 3 for WT5?! Training Text-to-Text Models to Explain their Predictions
Figure 4 for WT5?! Training Text-to-Text Models to Explain their Predictions
Viaarxiv icon

How Much Knowledge Can You Pack Into the Parameters of a Language Model?

Add code
Feb 24, 2020
Figure 1 for How Much Knowledge Can You Pack Into the Parameters of a Language Model?
Figure 2 for How Much Knowledge Can You Pack Into the Parameters of a Language Model?
Viaarxiv icon

DDSP: Differentiable Digital Signal Processing

Add code
Jan 14, 2020
Figure 1 for DDSP: Differentiable Digital Signal Processing
Figure 2 for DDSP: Differentiable Digital Signal Processing
Figure 3 for DDSP: Differentiable Digital Signal Processing
Figure 4 for DDSP: Differentiable Digital Signal Processing
Viaarxiv icon