Alert button
Picture for Yury Zemlyanskiy

Yury Zemlyanskiy

Alert button

MEMORY-VQ: Compression for Tractable Internet-Scale Memory

Add code
Bookmark button
Alert button
Aug 28, 2023
Yury Zemlyanskiy, Michiel de Jong, Luke Vilnis, Santiago Ontañón, William W. Cohen, Sumit Sanghai, Joshua Ainslie

Figure 1 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 2 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 3 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 4 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Viaarxiv icon

GLIMMER: generalized late-interaction memory reranker

Add code
Bookmark button
Alert button
Jun 17, 2023
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Sumit Sanghai, William W. Cohen, Joshua Ainslie

Figure 1 for GLIMMER: generalized late-interaction memory reranker
Figure 2 for GLIMMER: generalized late-interaction memory reranker
Figure 3 for GLIMMER: generalized late-interaction memory reranker
Figure 4 for GLIMMER: generalized late-interaction memory reranker
Viaarxiv icon

GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints

Add code
Bookmark button
Alert button
May 22, 2023
Joshua Ainslie, James Lee-Thorp, Michiel de Jong, Yury Zemlyanskiy, Federico Lebrón, Sumit Sanghai

Figure 1 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 2 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 3 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 4 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Viaarxiv icon

CoLT5: Faster Long-Range Transformers with Conditional Computation

Add code
Bookmark button
Alert button
Mar 17, 2023
Joshua Ainslie, Tao Lei, Michiel de Jong, Santiago Ontañón, Siddhartha Brahma, Yury Zemlyanskiy, David Uthus, Mandy Guo, James Lee-Thorp, Yi Tay, Yun-Hsuan Sung, Sumit Sanghai

Figure 1 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 2 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 3 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 4 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Viaarxiv icon

Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute

Add code
Bookmark button
Alert button
Jan 25, 2023
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Joshua Ainslie, Sumit Sanghai, Fei Sha, William Cohen

Figure 1 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 2 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 3 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 4 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Viaarxiv icon

FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference

Add code
Bookmark button
Alert button
Dec 15, 2022
Michiel de Jong, Yury Zemlyanskiy, Joshua Ainslie, Nicholas FitzGerald, Sumit Sanghai, Fei Sha, William Cohen

Figure 1 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 2 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 3 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 4 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Viaarxiv icon

Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing

Add code
Bookmark button
Alert button
Sep 29, 2022
Yury Zemlyanskiy, Michiel de Jong, Joshua Ainslie, Panupong Pasupat, Peter Shaw, Linlu Qiu, Sumit Sanghai, Fei Sha

Figure 1 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 2 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 3 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 4 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Viaarxiv icon

Mention Memory: incorporating textual knowledge into Transformers through entity mention attention

Add code
Bookmark button
Alert button
Oct 12, 2021
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Fei Sha, William Cohen

Figure 1 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Figure 2 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Figure 3 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Figure 4 for Mention Memory: incorporating textual knowledge into Transformers through entity mention attention
Viaarxiv icon

ReadTwice: Reading Very Large Documents with Memories

Add code
Bookmark button
Alert button
May 11, 2021
Yury Zemlyanskiy, Joshua Ainslie, Michiel de Jong, Philip Pham, Ilya Eckstein, Fei Sha

Figure 1 for ReadTwice: Reading Very Large Documents with Memories
Figure 2 for ReadTwice: Reading Very Large Documents with Memories
Figure 3 for ReadTwice: Reading Very Large Documents with Memories
Figure 4 for ReadTwice: Reading Very Large Documents with Memories
Viaarxiv icon