Alert button
Picture for Dimitris Papailiopoulos

Dimitris Papailiopoulos

Alert button

CHAI: Clustered Head Attention for Efficient LLM Inference

Add code
Bookmark button
Alert button
Mar 12, 2024
Saurabh Agarwal, Bilge Acun, Basil Homer, Mostafa Elhoushi, Yejin Lee, Shivaram Venkataraman, Dimitris Papailiopoulos, Carole-Jean Wu

Figure 1 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 2 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 3 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 4 for CHAI: Clustered Head Attention for Efficient LLM Inference
Viaarxiv icon

How Well Can Transformers Emulate In-context Newton's Method?

Add code
Bookmark button
Alert button
Mar 05, 2024
Angeliki Giannou, Liu Yang, Tianhao Wang, Dimitris Papailiopoulos, Jason D. Lee

Figure 1 for How Well Can Transformers Emulate In-context Newton's Method?
Figure 2 for How Well Can Transformers Emulate In-context Newton's Method?
Figure 3 for How Well Can Transformers Emulate In-context Newton's Method?
Figure 4 for How Well Can Transformers Emulate In-context Newton's Method?
Viaarxiv icon

Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning Tasks

Add code
Bookmark button
Alert button
Feb 06, 2024
Jongho Park, Jaeseung Park, Zheyang Xiong, Nayoung Lee, Jaewoong Cho, Samet Oymak, Kangwook Lee, Dimitris Papailiopoulos

Viaarxiv icon

Looped Transformers are Better at Learning Learning Algorithms

Add code
Bookmark button
Alert button
Nov 21, 2023
Liu Yang, Kangwook Lee, Robert Nowak, Dimitris Papailiopoulos

Viaarxiv icon

Predictive Pipelined Decoding: A Compute-Latency Trade-off for Exact LLM Decoding

Add code
Bookmark button
Alert button
Jul 12, 2023
Seongjun Yang, Gibbeum Lee, Jaewoong Cho, Dimitris Papailiopoulos, Kangwook Lee

Figure 1 for Predictive Pipelined Decoding: A Compute-Latency Trade-off for Exact LLM Decoding
Figure 2 for Predictive Pipelined Decoding: A Compute-Latency Trade-off for Exact LLM Decoding
Figure 3 for Predictive Pipelined Decoding: A Compute-Latency Trade-off for Exact LLM Decoding
Figure 4 for Predictive Pipelined Decoding: A Compute-Latency Trade-off for Exact LLM Decoding
Viaarxiv icon

Mini-Batch Optimization of Contrastive Loss

Add code
Bookmark button
Alert button
Jul 12, 2023
Jaewoong Cho, Kartik Sreenivasan, Keon Lee, Kyunghoo Mun, Soheun Yi, Jeong-Gwan Lee, Anna Lee, Jy-yong Sohn, Dimitris Papailiopoulos, Kangwook Lee

Figure 1 for Mini-Batch Optimization of Contrastive Loss
Figure 2 for Mini-Batch Optimization of Contrastive Loss
Figure 3 for Mini-Batch Optimization of Contrastive Loss
Figure 4 for Mini-Batch Optimization of Contrastive Loss
Viaarxiv icon

Teaching Arithmetic to Small Transformers

Add code
Bookmark button
Alert button
Jul 07, 2023
Nayoung Lee, Kartik Sreenivasan, Jason D. Lee, Kangwook Lee, Dimitris Papailiopoulos

Figure 1 for Teaching Arithmetic to Small Transformers
Figure 2 for Teaching Arithmetic to Small Transformers
Figure 3 for Teaching Arithmetic to Small Transformers
Figure 4 for Teaching Arithmetic to Small Transformers
Viaarxiv icon

Dissecting Chain-of-Thought: A Study on Compositional In-Context Learning of MLPs

Add code
Bookmark button
Alert button
May 30, 2023
Yingcong Li, Kartik Sreenivasan, Angeliki Giannou, Dimitris Papailiopoulos, Samet Oymak

Figure 1 for Dissecting Chain-of-Thought: A Study on Compositional In-Context Learning of MLPs
Figure 2 for Dissecting Chain-of-Thought: A Study on Compositional In-Context Learning of MLPs
Figure 3 for Dissecting Chain-of-Thought: A Study on Compositional In-Context Learning of MLPs
Figure 4 for Dissecting Chain-of-Thought: A Study on Compositional In-Context Learning of MLPs
Viaarxiv icon

Prompted LLMs as Chatbot Modules for Long Open-domain Conversation

Add code
Bookmark button
Alert button
May 08, 2023
Gibbeum Lee, Volker Hartmann, Jongho Park, Dimitris Papailiopoulos, Kangwook Lee

Figure 1 for Prompted LLMs as Chatbot Modules for Long Open-domain Conversation
Figure 2 for Prompted LLMs as Chatbot Modules for Long Open-domain Conversation
Figure 3 for Prompted LLMs as Chatbot Modules for Long Open-domain Conversation
Figure 4 for Prompted LLMs as Chatbot Modules for Long Open-domain Conversation
Viaarxiv icon

Cuttlefish: Low-Rank Model Training without All the Tuning

Add code
Bookmark button
Alert button
May 05, 2023
Hongyi Wang, Saurabh Agarwal, Pongsakorn U-chupala, Yoshiki Tanaka, Eric P. Xing, Dimitris Papailiopoulos

Figure 1 for Cuttlefish: Low-Rank Model Training without All the Tuning
Figure 2 for Cuttlefish: Low-Rank Model Training without All the Tuning
Figure 3 for Cuttlefish: Low-Rank Model Training without All the Tuning
Figure 4 for Cuttlefish: Low-Rank Model Training without All the Tuning
Viaarxiv icon