Alert button
Picture for Michiel de Jong

Michiel de Jong

Alert button

MEMORY-VQ: Compression for Tractable Internet-Scale Memory

Add code
Bookmark button
Alert button
Aug 28, 2023
Yury Zemlyanskiy, Michiel de Jong, Luke Vilnis, Santiago Ontañón, William W. Cohen, Sumit Sanghai, Joshua Ainslie

Figure 1 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 2 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 3 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 4 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Viaarxiv icon

GLIMMER: generalized late-interaction memory reranker

Add code
Bookmark button
Alert button
Jun 17, 2023
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Sumit Sanghai, William W. Cohen, Joshua Ainslie

Figure 1 for GLIMMER: generalized late-interaction memory reranker
Figure 2 for GLIMMER: generalized late-interaction memory reranker
Figure 3 for GLIMMER: generalized late-interaction memory reranker
Figure 4 for GLIMMER: generalized late-interaction memory reranker
Viaarxiv icon

GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints

Add code
Bookmark button
Alert button
May 22, 2023
Joshua Ainslie, James Lee-Thorp, Michiel de Jong, Yury Zemlyanskiy, Federico Lebrón, Sumit Sanghai

Figure 1 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 2 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 3 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 4 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Viaarxiv icon

CoLT5: Faster Long-Range Transformers with Conditional Computation

Add code
Bookmark button
Alert button
Mar 17, 2023
Joshua Ainslie, Tao Lei, Michiel de Jong, Santiago Ontañón, Siddhartha Brahma, Yury Zemlyanskiy, David Uthus, Mandy Guo, James Lee-Thorp, Yi Tay, Yun-Hsuan Sung, Sumit Sanghai

Figure 1 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 2 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 3 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 4 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Viaarxiv icon

Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute

Add code
Bookmark button
Alert button
Jan 25, 2023
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Joshua Ainslie, Sumit Sanghai, Fei Sha, William Cohen

Figure 1 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 2 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 3 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 4 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Viaarxiv icon

FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference

Add code
Bookmark button
Alert button
Dec 15, 2022
Michiel de Jong, Yury Zemlyanskiy, Joshua Ainslie, Nicholas FitzGerald, Sumit Sanghai, Fei Sha, William Cohen

Figure 1 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 2 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 3 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 4 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Viaarxiv icon

Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing

Add code
Bookmark button
Alert button
Sep 29, 2022
Yury Zemlyanskiy, Michiel de Jong, Joshua Ainslie, Panupong Pasupat, Peter Shaw, Linlu Qiu, Sumit Sanghai, Fei Sha

Figure 1 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 2 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 3 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 4 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Viaarxiv icon

Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering

Add code
Bookmark button
Alert button
Apr 10, 2022
Wenhu Chen, Pat Verga, Michiel de Jong, John Wieting, William Cohen

Figure 1 for Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering
Figure 2 for Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering
Figure 3 for Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering
Figure 4 for Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering
Viaarxiv icon

Mention Memory: incorporating textual knowledge into Transformers through entity mention attention

Add code
Bookmark button
Alert button
Oct 12, 2021
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Fei Sha, William Cohen

Figure 1 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Figure 2 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Figure 3 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Figure 4 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Viaarxiv icon

Grounding Complex Navigational Instructions Using Scene Graphs

Add code
Bookmark button
Alert button
Jun 03, 2021
Michiel de Jong, Satyapriya Krishna, Anuva Agarwal

Figure 1 for Grounding Complex Navigational Instructions Using Scene Graphs
Figure 2 for Grounding Complex Navigational Instructions Using Scene Graphs
Figure 3 for Grounding Complex Navigational Instructions Using Scene Graphs
Figure 4 for Grounding Complex Navigational Instructions Using Scene Graphs
Viaarxiv icon