Alert button
Picture for Yi Tay

Yi Tay

Alert button

PolyViT: Co-training Vision Transformers on Images, Videos and Audio

Add code
Bookmark button
Alert button
Nov 25, 2021
Valerii Likhosherstov, Anurag Arnab, Krzysztof Choromanski, Mario Lucic, Yi Tay, Adrian Weller, Mostafa Dehghani

Figure 1 for PolyViT: Co-training Vision Transformers on Images, Videos and Audio
Figure 2 for PolyViT: Co-training Vision Transformers on Images, Videos and Audio
Figure 3 for PolyViT: Co-training Vision Transformers on Images, Videos and Audio
Figure 4 for PolyViT: Co-training Vision Transformers on Images, Videos and Audio
Viaarxiv icon

ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning

Add code
Bookmark button
Alert button
Nov 22, 2021
Vamsi Aribandi, Yi Tay, Tal Schuster, Jinfeng Rao, Huaixiu Steven Zheng, Sanket Vaibhav Mehta, Honglei Zhuang, Vinh Q. Tran, Dara Bahri, Jianmo Ni, Jai Gupta, Kai Hui, Sebastian Ruder, Donald Metzler

Figure 1 for ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
Figure 2 for ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
Figure 3 for ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
Figure 4 for ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
Viaarxiv icon

The Efficiency Misnomer

Add code
Bookmark button
Alert button
Oct 25, 2021
Mostafa Dehghani, Anurag Arnab, Lucas Beyer, Ashish Vaswani, Yi Tay

Figure 1 for The Efficiency Misnomer
Figure 2 for The Efficiency Misnomer
Figure 3 for The Efficiency Misnomer
Viaarxiv icon

SCENIC: A JAX Library for Computer Vision Research and Beyond

Add code
Bookmark button
Alert button
Oct 18, 2021
Mostafa Dehghani, Alexey Gritsenko, Anurag Arnab, Matthias Minderer, Yi Tay

Figure 1 for SCENIC: A JAX Library for Computer Vision Research and Beyond
Viaarxiv icon

Sharpness-Aware Minimization Improves Language Model Generalization

Add code
Bookmark button
Alert button
Oct 16, 2021
Dara Bahri, Hossein Mobahi, Yi Tay

Figure 1 for Sharpness-Aware Minimization Improves Language Model Generalization
Figure 2 for Sharpness-Aware Minimization Improves Language Model Generalization
Figure 3 for Sharpness-Aware Minimization Improves Language Model Generalization
Figure 4 for Sharpness-Aware Minimization Improves Language Model Generalization
Viaarxiv icon

Improving Compositional Generalization with Self-Training for Data-to-Text Generation

Add code
Bookmark button
Alert button
Oct 16, 2021
Sanket Vaibhav Mehta, Jinfeng Rao, Yi Tay, Mihir Kale, Ankur Parikh, Hongtao Zhong, Emma Strubell

Figure 1 for Improving Compositional Generalization with Self-Training for Data-to-Text Generation
Figure 2 for Improving Compositional Generalization with Self-Training for Data-to-Text Generation
Figure 3 for Improving Compositional Generalization with Self-Training for Data-to-Text Generation
Figure 4 for Improving Compositional Generalization with Self-Training for Data-to-Text Generation
Viaarxiv icon

Born Again Neural Rankers

Add code
Bookmark button
Alert button
Sep 30, 2021
Zhen Qin, Le Yan, Yi Tay, Honglei Zhuang, Xuanhui Wang, Michael Bendersky, Marc Najork

Figure 1 for Born Again Neural Rankers
Figure 2 for Born Again Neural Rankers
Figure 3 for Born Again Neural Rankers
Figure 4 for Born Again Neural Rankers
Viaarxiv icon

Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers

Add code
Bookmark button
Alert button
Sep 22, 2021
Yi Tay, Mostafa Dehghani, Jinfeng Rao, William Fedus, Samira Abnar, Hyung Won Chung, Sharan Narang, Dani Yogatama, Ashish Vaswani, Donald Metzler

Figure 1 for Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers
Figure 2 for Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers
Figure 3 for Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers
Figure 4 for Scale Efficiently: Insights from Pre-training and Fine-tuning Transformers
Viaarxiv icon

The Benchmark Lottery

Add code
Bookmark button
Alert button
Jul 14, 2021
Mostafa Dehghani, Yi Tay, Alexey A. Gritsenko, Zhe Zhao, Neil Houlsby, Fernando Diaz, Donald Metzler, Oriol Vinyals

Figure 1 for The Benchmark Lottery
Figure 2 for The Benchmark Lottery
Figure 3 for The Benchmark Lottery
Figure 4 for The Benchmark Lottery
Viaarxiv icon

Charformer: Fast Character Transformers via Gradient-based Subword Tokenization

Add code
Bookmark button
Alert button
Jul 02, 2021
Yi Tay, Vinh Q. Tran, Sebastian Ruder, Jai Gupta, Hyung Won Chung, Dara Bahri, Zhen Qin, Simon Baumgartner, Cong Yu, Donald Metzler

Figure 1 for Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
Figure 2 for Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
Figure 3 for Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
Figure 4 for Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
Viaarxiv icon