Alert button
Picture for Payal Bajaj

Payal Bajaj

Alert button

Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers

Add code
Bookmark button
Alert button
May 21, 2023
Linyuan Gong, Chenyan Xiong, Xiaodong Liu, Payal Bajaj, Yiqing Xie, Alvin Cheung, Jianfeng Gao, Xia Song

Figure 1 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Figure 2 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Figure 3 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Figure 4 for Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers
Viaarxiv icon

Understanding the Effectiveness of Very Large Language Models on Dialog Evaluation

Add code
Bookmark button
Alert button
Jan 27, 2023
Jessica Huynh, Cathy Jiao, Prakhar Gupta, Shikib Mehri, Payal Bajaj, Vishrav Chaudhary, Maxine Eskenazi

Figure 1 for Understanding the Effectiveness of Very Large Language Models on Dialog Evaluation
Figure 2 for Understanding the Effectiveness of Very Large Language Models on Dialog Evaluation
Figure 3 for Understanding the Effectiveness of Very Large Language Models on Dialog Evaluation
Figure 4 for Understanding the Effectiveness of Very Large Language Models on Dialog Evaluation
Viaarxiv icon

Foundation Transformers

Add code
Bookmark button
Alert button
Oct 19, 2022
Hongyu Wang, Shuming Ma, Shaohan Huang, Li Dong, Wenhui Wang, Zhiliang Peng, Yu Wu, Payal Bajaj, Saksham Singhal, Alon Benhaim, Barun Patra, Zhun Liu, Vishrav Chaudhary, Xia Song, Furu Wei

Figure 1 for Foundation Transformers
Figure 2 for Foundation Transformers
Figure 3 for Foundation Transformers
Figure 4 for Foundation Transformers
Viaarxiv icon

On the Representation Collapse of Sparse Mixture of Experts

Add code
Bookmark button
Alert button
Apr 20, 2022
Zewen Chi, Li Dong, Shaohan Huang, Damai Dai, Shuming Ma, Barun Patra, Saksham Singhal, Payal Bajaj, Xia Song, Furu Wei

Figure 1 for On the Representation Collapse of Sparse Mixture of Experts
Figure 2 for On the Representation Collapse of Sparse Mixture of Experts
Figure 3 for On the Representation Collapse of Sparse Mixture of Experts
Figure 4 for On the Representation Collapse of Sparse Mixture of Experts
Viaarxiv icon

METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals

Add code
Bookmark button
Alert button
Apr 16, 2022
Payal Bajaj, Chenyan Xiong, Guolin Ke, Xiaodong Liu, Di He, Saurabh Tiwary, Tie-Yan Liu, Paul Bennett, Xia Song, Jianfeng Gao

Figure 1 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 2 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 3 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 4 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Viaarxiv icon

Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators

Add code
Bookmark button
Alert button
Apr 07, 2022
Yu Meng, Chenyan Xiong, Payal Bajaj, Saurabh Tiwary, Paul Bennett, Jiawei Han, Xia Song

Figure 1 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Figure 2 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Figure 3 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Figure 4 for Pretraining Text Encoders with Adversarial Mixture of Training Signal Generators
Viaarxiv icon

XLM-E: Cross-lingual Language Model Pre-training via ELECTRA

Add code
Bookmark button
Alert button
Jun 30, 2021
Zewen Chi, Shaohan Huang, Li Dong, Shuming Ma, Saksham Singhal, Payal Bajaj, Xia Song, Furu Wei

Figure 1 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Figure 2 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Figure 3 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Figure 4 for XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Viaarxiv icon

Language Scaling for Universal Suggested Replies Model

Add code
Bookmark button
Alert button
Jun 04, 2021
Qianlan Ying, Payal Bajaj, Budhaditya Deb, Yu Yang, Wei Wang, Bojia Lin, Milad Shokouhi, Xia Song, Yang Yang, Daxin Jiang

Figure 1 for Language Scaling for Universal Suggested Replies Model
Figure 2 for Language Scaling for Universal Suggested Replies Model
Figure 3 for Language Scaling for Universal Suggested Replies Model
Figure 4 for Language Scaling for Universal Suggested Replies Model
Viaarxiv icon