Picture for Chris Dyer

Chris Dyer

Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context

Add code
Mar 08, 2024
Viaarxiv icon

Continuous diffusion for categorical data

Add code
Dec 15, 2022
Figure 1 for Continuous diffusion for categorical data
Figure 2 for Continuous diffusion for categorical data
Figure 3 for Continuous diffusion for categorical data
Figure 4 for Continuous diffusion for categorical data
Viaarxiv icon

MAD for Robust Reinforcement Learning in Machine Translation

Add code
Jul 18, 2022
Figure 1 for MAD for Robust Reinforcement Learning in Machine Translation
Figure 2 for MAD for Robust Reinforcement Learning in Machine Translation
Figure 3 for MAD for Robust Reinforcement Learning in Machine Translation
Figure 4 for MAD for Robust Reinforcement Learning in Machine Translation
Viaarxiv icon

Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale

Add code
Mar 01, 2022
Figure 1 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Figure 2 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Figure 3 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Figure 4 for Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Viaarxiv icon

Enabling arbitrary translation objectives with Adaptive Tree Search

Add code
Feb 23, 2022
Figure 1 for Enabling arbitrary translation objectives with Adaptive Tree Search
Figure 2 for Enabling arbitrary translation objectives with Adaptive Tree Search
Figure 3 for Enabling arbitrary translation objectives with Adaptive Tree Search
Figure 4 for Enabling arbitrary translation objectives with Adaptive Tree Search
Viaarxiv icon

Scaling Language Models: Methods, Analysis & Insights from Training Gopher

Add code
Dec 08, 2021
Figure 1 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 2 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 3 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 4 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Viaarxiv icon

End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering

Add code
Jun 09, 2021
Figure 1 for End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering
Figure 2 for End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering
Figure 3 for End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering
Figure 4 for End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering
Viaarxiv icon

Diverse Pretrained Context Encodings Improve Document Translation

Add code
Jun 07, 2021
Figure 1 for Diverse Pretrained Context Encodings Improve Document Translation
Figure 2 for Diverse Pretrained Context Encodings Improve Document Translation
Figure 3 for Diverse Pretrained Context Encodings Improve Document Translation
Figure 4 for Diverse Pretrained Context Encodings Improve Document Translation
Viaarxiv icon

Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings

Add code
Jun 04, 2021
Figure 1 for Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings
Figure 2 for Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings
Figure 3 for Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings
Figure 4 for Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings
Viaarxiv icon

Syntactic Structure Distillation Pretraining For Bidirectional Encoders

Add code
May 27, 2020
Figure 1 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Figure 2 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Figure 3 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Figure 4 for Syntactic Structure Distillation Pretraining For Bidirectional Encoders
Viaarxiv icon