Alert button
Picture for Xiaodong Liu

Xiaodong Liu

Alert button

PTSEFormer: Progressive Temporal-Spatial Enhanced TransFormer Towards Video Object Detection

Add code
Bookmark button
Alert button
Sep 06, 2022
Han Wang, Jun Tang, Xiaodong Liu, Shanyan Guan, Rong Xie, Li Song

Figure 1 for PTSEFormer: Progressive Temporal-Spatial Enhanced TransFormer Towards Video Object Detection
Figure 2 for PTSEFormer: Progressive Temporal-Spatial Enhanced TransFormer Towards Video Object Detection
Figure 3 for PTSEFormer: Progressive Temporal-Spatial Enhanced TransFormer Towards Video Object Detection
Figure 4 for PTSEFormer: Progressive Temporal-Spatial Enhanced TransFormer Towards Video Object Detection
Viaarxiv icon

Deep Generative Modeling on Limited Data with Regularization by Nontransferable Pre-trained Models

Add code
Bookmark button
Alert button
Aug 30, 2022
Yong Zhong, Hongtao Liu, Xiaodong Liu, Fan Bao, Weiran Shen, Chongxuan Li

Figure 1 for Deep Generative Modeling on Limited Data with Regularization by Nontransferable Pre-trained Models
Figure 2 for Deep Generative Modeling on Limited Data with Regularization by Nontransferable Pre-trained Models
Figure 3 for Deep Generative Modeling on Limited Data with Regularization by Nontransferable Pre-trained Models
Figure 4 for Deep Generative Modeling on Limited Data with Regularization by Nontransferable Pre-trained Models
Viaarxiv icon

AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models

Add code
Bookmark button
Alert button
May 24, 2022
Yaqing Wang, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao

Figure 1 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Figure 2 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Figure 3 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Figure 4 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Viaarxiv icon

Visually-Augmented Language Modeling

Add code
Bookmark button
Alert button
May 20, 2022
Weizhi Wang, Li Dong, Hao Cheng, Haoyu Song, Xiaodong Liu, Xifeng Yan, Jianfeng Gao, Furu Wei

Figure 1 for Visually-Augmented Language Modeling
Figure 2 for Visually-Augmented Language Modeling
Figure 3 for Visually-Augmented Language Modeling
Figure 4 for Visually-Augmented Language Modeling
Viaarxiv icon

METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals

Add code
Bookmark button
Alert button
Apr 16, 2022
Payal Bajaj, Chenyan Xiong, Guolin Ke, Xiaodong Liu, Di He, Saurabh Tiwary, Tie-Yan Liu, Paul Bennett, Xia Song, Jianfeng Gao

Figure 1 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 2 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 3 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Figure 4 for METRO: Efficient Denoising Pretraining of Large Scale Autoencoding Language Models with Model Generated Signals
Viaarxiv icon

Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer

Add code
Bookmark button
Alert button
Mar 28, 2022
Greg Yang, Edward J. Hu, Igor Babuschkin, Szymon Sidor, Xiaodong Liu, David Farhi, Nick Ryder, Jakub Pachocki, Weizhu Chen, Jianfeng Gao

Figure 1 for Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Figure 2 for Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Figure 3 for Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Figure 4 for Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Viaarxiv icon

A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models

Add code
Bookmark button
Alert button
Feb 17, 2022
Da Yin, Li Dong, Hao Cheng, Xiaodong Liu, Kai-Wei Chang, Furu Wei, Jianfeng Gao

Figure 1 for A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models
Figure 2 for A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models
Figure 3 for A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models
Figure 4 for A Survey of Knowledge-Intensive NLP with Pre-Trained Language Models
Viaarxiv icon

No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models

Add code
Bookmark button
Alert button
Feb 14, 2022
Chen Liang, Haoming Jiang, Simiao Zuo, Pengcheng He, Xiaodong Liu, Jianfeng Gao, Weizhu Chen, Tuo Zhao

Figure 1 for No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models
Figure 2 for No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models
Figure 3 for No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models
Figure 4 for No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models
Viaarxiv icon