Alert button
Picture for Yuqi Huo

Yuqi Huo

Alert button

VDT: An Empirical Study on Video Diffusion with Transformers

Add code
Bookmark button
Alert button
May 22, 2023
Haoyu Lu, Guoxing Yang, Nanyi Fei, Yuqi Huo, Zhiwu Lu, Ping Luo, Mingyu Ding

Figure 1 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 2 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 3 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 4 for VDT: An Empirical Study on Video Diffusion with Transformers
Viaarxiv icon

UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling

Add code
Bookmark button
Alert button
Feb 13, 2023
Haoyu Lu, Mingyu Ding, Yuqi Huo, Guoxing Yang, Zhiwu Lu, Masayoshi Tomizuka, Wei Zhan

Figure 1 for UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling
Figure 2 for UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling
Figure 3 for UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling
Figure 4 for UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling
Viaarxiv icon

LGDN: Language-Guided Denoising Network for Video-Language Modeling

Add code
Bookmark button
Alert button
Oct 03, 2022
Haoyu Lu, Mingyu Ding, Nanyi Fei, Yuqi Huo, Zhiwu Lu

Figure 1 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Figure 2 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Figure 3 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Figure 4 for LGDN: Language-Guided Denoising Network for Video-Language Modeling
Viaarxiv icon

COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval

Add code
Bookmark button
Alert button
Apr 15, 2022
Haoyu Lu, Nanyi Fei, Yuqi Huo, Yizhao Gao, Zhiwu Lu, Ji-Rong Wen

Figure 1 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Figure 2 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Figure 3 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Figure 4 for COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval
Viaarxiv icon

WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model

Add code
Bookmark button
Alert button
Oct 27, 2021
Nanyi Fei, Zhiwu Lu, Yizhao Gao, Guoxing Yang, Yuqi Huo, Jingyuan Wen, Haoyu Lu, Ruihua Song, Xin Gao, Tao Xiang, Hao Sun, Ji-Rong Wen

Figure 1 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Figure 2 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Figure 3 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Figure 4 for WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model
Viaarxiv icon

Pre-Trained Models: Past, Present and Future

Add code
Bookmark button
Alert button
Jun 15, 2021
Xu Han, Zhengyan Zhang, Ning Ding, Yuxian Gu, Xiao Liu, Yuqi Huo, Jiezhong Qiu, Liang Zhang, Wentao Han, Minlie Huang, Qin Jin, Yanyan Lan, Yang Liu, Zhiyuan Liu, Zhiwu Lu, Xipeng Qiu, Ruihua Song, Jie Tang, Ji-Rong Wen, Jinhui Yuan, Wayne Xin Zhao, Jun Zhu

Figure 1 for Pre-Trained Models: Past, Present and Future
Figure 2 for Pre-Trained Models: Past, Present and Future
Figure 3 for Pre-Trained Models: Past, Present and Future
Figure 4 for Pre-Trained Models: Past, Present and Future
Viaarxiv icon

Learning Versatile Neural Architectures by Propagating Network Codes

Add code
Bookmark button
Alert button
Mar 24, 2021
Mingyu Ding, Yuqi Huo, Haoyu Lu, Linjie Yang, Zhe Wang, Zhiwu Lu, Jingdong Wang, Ping Luo

Figure 1 for Learning Versatile Neural Architectures by Propagating Network Codes
Figure 2 for Learning Versatile Neural Architectures by Propagating Network Codes
Figure 3 for Learning Versatile Neural Architectures by Propagating Network Codes
Figure 4 for Learning Versatile Neural Architectures by Propagating Network Codes
Viaarxiv icon

WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training

Add code
Bookmark button
Alert button
Mar 19, 2021
Yuqi Huo, Manli Zhang, Guangzhen Liu, Haoyu Lu, Yizhao Gao, Guoxing Yang, Jingyuan Wen, Heng Zhang, Baogui Xu, Weihao Zheng, Zongzheng Xi, Yueqian Yang, Anwen Hu, Jinming Zhao, Ruichen Li, Yida Zhao, Liang Zhang, Yuqing Song, Xin Hong, Wanqing Cui, Danyang Hou, Yingyan Li, Junyi Li, Peiyu Liu, Zheng Gong, Chuhao Jin, Yuchong Sun, Shizhe Chen, Zhiwu Lu, Zhicheng Dou, Qin Jin, Yanyan Lan, Wayne Xin Zhao, Ruihua Song, Ji-Rong Wen

Figure 1 for WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training
Figure 2 for WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training
Figure 3 for WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training
Figure 4 for WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training
Viaarxiv icon