Alert button
Picture for Piotr Nawrot

Piotr Nawrot

Alert button

Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference

Add code
Bookmark button
Alert button
Mar 14, 2024
Piotr Nawrot, Adrian Łańcucki, Marcin Chochowski, David Tarjan, Edoardo M. Ponti

Figure 1 for Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
Figure 2 for Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
Figure 3 for Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
Figure 4 for Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
Viaarxiv icon

nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources

Add code
Bookmark button
Alert button
Sep 05, 2023
Piotr Nawrot

Figure 1 for nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources
Figure 2 for nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources
Figure 3 for nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources
Figure 4 for nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources
Viaarxiv icon

No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models

Add code
Bookmark button
Alert button
Jul 26, 2023
Jean Kaddour, Oscar Key, Piotr Nawrot, Pasquale Minervini, Matt J. Kusner

Figure 1 for No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models
Figure 2 for No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models
Figure 3 for No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models
Figure 4 for No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models
Viaarxiv icon

Efficient Transformers with Dynamic Token Pooling

Add code
Bookmark button
Alert button
Nov 17, 2022
Piotr Nawrot, Jan Chorowski, Adrian Łańcucki, Edoardo M. Ponti

Figure 1 for Efficient Transformers with Dynamic Token Pooling
Figure 2 for Efficient Transformers with Dynamic Token Pooling
Figure 3 for Efficient Transformers with Dynamic Token Pooling
Figure 4 for Efficient Transformers with Dynamic Token Pooling
Viaarxiv icon

Hierarchical Transformers Are More Efficient Language Models

Add code
Bookmark button
Alert button
Oct 26, 2021
Piotr Nawrot, Szymon Tworkowski, Michał Tyrolski, Łukasz Kaiser, Yuhuai Wu, Christian Szegedy, Henryk Michalewski

Figure 1 for Hierarchical Transformers Are More Efficient Language Models
Figure 2 for Hierarchical Transformers Are More Efficient Language Models
Figure 3 for Hierarchical Transformers Are More Efficient Language Models
Figure 4 for Hierarchical Transformers Are More Efficient Language Models
Viaarxiv icon