Alert button
Picture for Nanyi Fei

Nanyi Fei

Alert button

CoTBal: Comprehensive Task Balancing for Multi-Task Visual Instruction Tuning

Add code
Bookmark button
Alert button
Mar 07, 2024
Yanqi Dai, Dong Jing, Nanyi Fei, Zhiwu Lu

Figure 1 for CoTBal: Comprehensive Task Balancing for Multi-Task Visual Instruction Tuning
Figure 2 for CoTBal: Comprehensive Task Balancing for Multi-Task Visual Instruction Tuning
Figure 3 for CoTBal: Comprehensive Task Balancing for Multi-Task Visual Instruction Tuning
Figure 4 for CoTBal: Comprehensive Task Balancing for Multi-Task Visual Instruction Tuning
Viaarxiv icon

Improvable Gap Balancing for Multi-Task Learning

Add code
Bookmark button
Alert button
Jul 28, 2023
Yanqi Dai, Nanyi Fei, Zhiwu Lu

Figure 1 for Improvable Gap Balancing for Multi-Task Learning
Figure 2 for Improvable Gap Balancing for Multi-Task Learning
Figure 3 for Improvable Gap Balancing for Multi-Task Learning
Figure 4 for Improvable Gap Balancing for Multi-Task Learning
Viaarxiv icon

VDT: An Empirical Study on Video Diffusion with Transformers

Add code
Bookmark button
Alert button
May 22, 2023
Haoyu Lu, Guoxing Yang, Nanyi Fei, Yuqi Huo, Zhiwu Lu, Ping Luo, Mingyu Ding

Figure 1 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 2 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 3 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 4 for VDT: An Empirical Study on Video Diffusion with Transformers
Viaarxiv icon

LGDN: Language-Guided Denoising Network for Video-Language Modeling

Add code
Bookmark button
Alert button
Oct 03, 2022
Haoyu Lu, Mingyu Ding, Nanyi Fei, Yuqi Huo, Zhiwu Lu

Figure 1 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Figure 2 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Figure 3 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Figure 4 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Viaarxiv icon

Multimodal foundation models are better simulators of the human brain

Add code
Bookmark button
Alert button
Aug 17, 2022
Haoyu Lu, Qiongyi Zhou, Nanyi Fei, Zhiwu Lu, Mingyu Ding, Jingyuan Wen, Changde Du, Xin Zhao, Hao Sun, Huiguang He, Ji-Rong Wen

Figure 1 for Multimodal foundation models are better simulators of the human brain
Figure 2 for Multimodal foundation models are better simulators of the human brain
Figure 3 for Multimodal foundation models are better simulators of the human brain
Figure 4 for Multimodal foundation models are better simulators of the human brain
Viaarxiv icon

COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval

Add code
Bookmark button
Alert button
Apr 15, 2022
Haoyu Lu, Nanyi Fei, Yuqi Huo, Yizhao Gao, Zhiwu Lu, Ji-Rong Wen

Figure 1 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Figure 2 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Figure 3 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Figure 4 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Viaarxiv icon

A Roadmap for Big Model

Add code
Bookmark button
Alert button
Apr 02, 2022
Sha Yuan, Hanyu Zhao, Shuai Zhao, Jiahong Leng, Yangxiao Liang, Xiaozhi Wang, Jifan Yu, Xin Lv, Zhou Shao, Jiaao He, Yankai Lin, Xu Han, Zhenghao Liu, Ning Ding, Yongming Rao, Yizhao Gao, Liang Zhang, Ming Ding, Cong Fang, Yisen Wang, Mingsheng Long, Jing Zhang, Yinpeng Dong, Tianyu Pang, Peng Cui, Lingxiao Huang, Zheng Liang, Huawei Shen, Hui Zhang, Quanshi Zhang, Qingxiu Dong, Zhixing Tan, Mingxuan Wang, Shuo Wang, Long Zhou, Haoran Li, Junwei Bao, Yingwei Pan, Weinan Zhang, Zhou Yu, Rui Yan, Chence Shi, Minghao Xu, Zuobai Zhang, Guoqiang Wang, Xiang Pan, Mengjie Li, Xiaoyu Chu, Zijun Yao, Fangwei Zhu, Shulin Cao, Weicheng Xue, Zixuan Ma, Zhengyan Zhang, Shengding Hu, Yujia Qin, Chaojun Xiao, Zheni Zeng, Ganqu Cui, Weize Chen, Weilin Zhao, Yuan Yao, Peng Li, Wenzhao Zheng, Wenliang Zhao, Ziyi Wang, Borui Zhang, Nanyi Fei, Anwen Hu, Zenan Ling, Haoyang Li, Boxi Cao, Xianpei Han, Weidong Zhan, Baobao Chang, Hao Sun, Jiawen Deng, Chujie Zheng, Juanzi Li, Lei Hou, Xigang Cao, Jidong Zhai, Zhiyuan Liu, Maosong Sun, Jiwen Lu, Zhiwu Lu, Qin Jin, Ruihua Song, Ji-Rong Wen, Zhouchen Lin, Liwei Wang, Hang Su, Jun Zhu, Zhifang Sui, Jiajun Zhang, Yang Liu, Xiaodong He, Minlie Huang, Jian Tang, Jie Tang

Figure 1 for A Roadmap for Big Model
Figure 2 for A Roadmap for Big Model
Figure 3 for A Roadmap for Big Model
Figure 4 for A Roadmap for Big Model
Viaarxiv icon

WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model

Add code
Bookmark button
Alert button
Oct 27, 2021
Nanyi Fei, Zhiwu Lu, Yizhao Gao, Guoxing Yang, Yuqi Huo, Jingyuan Wen, Haoyu Lu, Ruihua Song, Xin Gao, Tao Xiang, Hao Sun, Ji-Rong Wen

Figure 1 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Figure 2 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Figure 3 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Figure 4 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Viaarxiv icon

Contrastive Prototype Learning with Augmented Embeddings for Few-Shot Learning

Add code
Bookmark button
Alert button
Jan 23, 2021
Yizhao Gao, Nanyi Fei, Guangzhen Liu, Zhiwu Lu, Tao Xiang, Songfang Huang

Figure 1 for Contrastive Prototype Learning with Augmented Embeddings for Few-Shot Learning
Figure 2 for Contrastive Prototype Learning with Augmented Embeddings for Few-Shot Learning
Figure 3 for Contrastive Prototype Learning with Augmented Embeddings for Few-Shot Learning
Figure 4 for Contrastive Prototype Learning with Augmented Embeddings for Few-Shot Learning
Viaarxiv icon