Alert button
Picture for Shuai Bai

Shuai Bai

Alert button

An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models

Add code
Bookmark button
Alert button
Mar 11, 2024
Liang Chen, Haozhe Zhao, Tianyu Liu, Shuai Bai, Junyang Lin, Chang Zhou, Baobao Chang

Figure 1 for An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models
Figure 2 for An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models
Figure 3 for An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models
Figure 4 for An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models
Viaarxiv icon

GD^2-NeRF: Generative Detail Compensation via GAN and Diffusion for One-shot Generalizable Neural Radiance Fields

Add code
Bookmark button
Alert button
Jan 02, 2024
Xiao Pan, Zongxin Yang, Shuai Bai, Yi Yang

Viaarxiv icon

Qwen Technical Report

Add code
Bookmark button
Alert button
Sep 28, 2023
Jinze Bai, Shuai Bai, Yunfei Chu, Zeyu Cui, Kai Dang, Xiaodong Deng, Yang Fan, Wenbin Ge, Yu Han, Fei Huang, Binyuan Hui, Luo Ji, Mei Li, Junyang Lin, Runji Lin, Dayiheng Liu, Gao Liu, Chengqiang Lu, Keming Lu, Jianxin Ma, Rui Men, Xingzhang Ren, Xuancheng Ren, Chuanqi Tan, Sinan Tan, Jianhong Tu, Peng Wang, Shijie Wang, Wei Wang, Shengguang Wu, Benfeng Xu, Jin Xu, An Yang, Hao Yang, Jian Yang, Shusheng Yang, Yang Yao, Bowen Yu, Hongyi Yuan, Zheng Yuan, Jianwei Zhang, Xingxuan Zhang, Yichang Zhang, Zhenru Zhang, Chang Zhou, Jingren Zhou, Xiaohuan Zhou, Tianhang Zhu

Figure 1 for Qwen Technical Report
Figure 2 for Qwen Technical Report
Figure 3 for Qwen Technical Report
Figure 4 for Qwen Technical Report
Viaarxiv icon

Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond

Add code
Bookmark button
Alert button
Sep 14, 2023
Jinze Bai, Shuai Bai, Shusheng Yang, Shijie Wang, Sinan Tan, Peng Wang, Junyang Lin, Chang Zhou, Jingren Zhou

Figure 1 for Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond
Figure 2 for Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond
Figure 3 for Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond
Figure 4 for Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond
Viaarxiv icon

TouchStone: Evaluating Vision-Language Models by Language Models

Add code
Bookmark button
Alert button
Sep 04, 2023
Shuai Bai, Shusheng Yang, Jinze Bai, Peng Wang, Xingxuan Zhang, Junyang Lin, Xinggang Wang, Chang Zhou, Jingren Zhou

Figure 1 for TouchStone: Evaluating Vision-Language Models by Language Models
Figure 2 for TouchStone: Evaluating Vision-Language Models by Language Models
Figure 3 for TouchStone: Evaluating Vision-Language Models by Language Models
Figure 4 for TouchStone: Evaluating Vision-Language Models by Language Models
Viaarxiv icon

Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities

Add code
Bookmark button
Alert button
Aug 24, 2023
Jinze Bai, Shuai Bai, Shusheng Yang, Shijie Wang, Sinan Tan, Peng Wang, Junyang Lin, Chang Zhou, Jingren Zhou

Figure 1 for Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities
Figure 2 for Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities
Figure 3 for Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities
Figure 4 for Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities
Viaarxiv icon

ONE-PEACE: Exploring One General Representation Model Toward Unlimited Modalities

Add code
Bookmark button
Alert button
May 18, 2023
Peng Wang, Shijie Wang, Junyang Lin, Shuai Bai, Xiaohuan Zhou, Jingren Zhou, Xinggang Wang, Chang Zhou

Figure 1 for ONE-PEACE: Exploring One General Representation Model Toward Unlimited Modalities
Figure 2 for ONE-PEACE: Exploring One General Representation Model Toward Unlimited Modalities
Figure 3 for ONE-PEACE: Exploring One General Representation Model Toward Unlimited Modalities
Figure 4 for ONE-PEACE: Exploring One General Representation Model Toward Unlimited Modalities
Viaarxiv icon

OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models

Add code
Bookmark button
Alert button
Dec 08, 2022
Jinze Bai, Rui Men, Hao Yang, Xuancheng Ren, Kai Dang, Yichang Zhang, Xiaohuan Zhou, Peng Wang, Sinan Tan, An Yang, Zeyu Cui, Yu Han, Shuai Bai, Wenbin Ge, Jianxin Ma, Junyang Lin, Jingren Zhou, Chang Zhou

Figure 1 for OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models
Figure 2 for OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models
Figure 3 for OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models
Figure 4 for OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist Models
Viaarxiv icon

Pretrained Diffusion Models for Unified Human Motion Synthesis

Add code
Bookmark button
Alert button
Dec 06, 2022
Jianxin Ma, Shuai Bai, Chang Zhou

Figure 1 for Pretrained Diffusion Models for Unified Human Motion Synthesis
Figure 2 for Pretrained Diffusion Models for Unified Human Motion Synthesis
Figure 3 for Pretrained Diffusion Models for Unified Human Motion Synthesis
Figure 4 for Pretrained Diffusion Models for Unified Human Motion Synthesis
Viaarxiv icon