Alert button
Picture for Sumit Sanghai

Sumit Sanghai

Alert button

Functional Interpolation for Relative Positions Improves Long Context Transformers

Add code
Bookmark button
Alert button
Oct 06, 2023
Shanda Li, Chong You, Guru Guruganesh, Joshua Ainslie, Santiago Ontanon, Manzil Zaheer, Sumit Sanghai, Yiming Yang, Sanjiv Kumar, Srinadh Bhojanapalli

Viaarxiv icon

MEMORY-VQ: Compression for Tractable Internet-Scale Memory

Add code
Bookmark button
Alert button
Aug 28, 2023
Yury Zemlyanskiy, Michiel de Jong, Luke Vilnis, Santiago Ontañón, William W. Cohen, Sumit Sanghai, Joshua Ainslie

Figure 1 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 2 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 3 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Figure 4 for MEMORY-VQ: Compression for Tractable Internet-Scale Memory
Viaarxiv icon

GLIMMER: generalized late-interaction memory reranker

Add code
Bookmark button
Alert button
Jun 17, 2023
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Sumit Sanghai, William W. Cohen, Joshua Ainslie

Figure 1 for GLIMMER: generalized late-interaction memory reranker
Figure 2 for GLIMMER: generalized late-interaction memory reranker
Figure 3 for GLIMMER: generalized late-interaction memory reranker
Figure 4 for GLIMMER: generalized late-interaction memory reranker
Viaarxiv icon

GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints

Add code
Bookmark button
Alert button
May 22, 2023
Joshua Ainslie, James Lee-Thorp, Michiel de Jong, Yury Zemlyanskiy, Federico Lebrón, Sumit Sanghai

Figure 1 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 2 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 3 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Figure 4 for GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints
Viaarxiv icon

CoLT5: Faster Long-Range Transformers with Conditional Computation

Add code
Bookmark button
Alert button
Mar 17, 2023
Joshua Ainslie, Tao Lei, Michiel de Jong, Santiago Ontañón, Siddhartha Brahma, Yury Zemlyanskiy, David Uthus, Mandy Guo, James Lee-Thorp, Yi Tay, Yun-Hsuan Sung, Sumit Sanghai

Figure 1 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 2 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 3 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Figure 4 for CoLT5: Faster Long-Range Transformers with Conditional Computation
Viaarxiv icon

Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute

Add code
Bookmark button
Alert button
Jan 25, 2023
Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Joshua Ainslie, Sumit Sanghai, Fei Sha, William Cohen

Figure 1 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 2 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 3 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Figure 4 for Pre-computed memory or on-the-fly encoding? A hybrid approach to retrieval augmentation makes the most of your compute
Viaarxiv icon

ImPaKT: A Dataset for Open-Schema Knowledge Base Construction

Add code
Bookmark button
Alert button
Dec 21, 2022
Luke Vilnis, Zach Fisher, Bhargav Kanagal, Patrick Murray, Sumit Sanghai

Figure 1 for ImPaKT: A Dataset for Open-Schema Knowledge Base Construction
Figure 2 for ImPaKT: A Dataset for Open-Schema Knowledge Base Construction
Figure 3 for ImPaKT: A Dataset for Open-Schema Knowledge Base Construction
Figure 4 for ImPaKT: A Dataset for Open-Schema Knowledge Base Construction
Viaarxiv icon

FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference

Add code
Bookmark button
Alert button
Dec 15, 2022
Michiel de Jong, Yury Zemlyanskiy, Joshua Ainslie, Nicholas FitzGerald, Sumit Sanghai, Fei Sha, William Cohen

Figure 1 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 2 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 3 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Figure 4 for FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference
Viaarxiv icon

Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing

Add code
Bookmark button
Alert button
Sep 29, 2022
Yury Zemlyanskiy, Michiel de Jong, Joshua Ainslie, Panupong Pasupat, Peter Shaw, Linlu Qiu, Sumit Sanghai, Fei Sha

Figure 1 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 2 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 3 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Figure 4 for Generate-and-Retrieve: use your predictions to improve retrieval for semantic parsing
Viaarxiv icon

MAVE: A Product Dataset for Multi-source Attribute Value Extraction

Add code
Bookmark button
Alert button
Dec 16, 2021
Li Yang, Qifan Wang, Zac Yu, Anand Kulkarni, Sumit Sanghai, Bin Shu, Jon Elsas, Bhargav Kanagal

Figure 1 for MAVE: A Product Dataset for Multi-source Attribute Value Extraction
Figure 2 for MAVE: A Product Dataset for Multi-source Attribute Value Extraction
Figure 3 for MAVE: A Product Dataset for Multi-source Attribute Value Extraction
Figure 4 for MAVE: A Product Dataset for Multi-source Attribute Value Extraction
Viaarxiv icon