Alert button
Picture for Longtian Qiu

Longtian Qiu

Alert button

SPHINX-X: Scaling Data and Parameters for a Family of Multi-modal Large Language Models

Feb 08, 2024
Peng Gao, Renrui Zhang, Chris Liu, Longtian Qiu, Siyuan Huang, Weifeng Lin, Shitian Zhao, Shijie Geng, Ziyi Lin, Peng Jin, Kaipeng Zhang, Wenqi Shao, Chao Xu, Conghui He, Junjun He, Hao Shao, Pan Lu, Hongsheng Li, Yu Qiao

Viaarxiv icon

Mining Fine-Grained Image-Text Alignment for Zero-Shot Captioning via Text-Only Training

Jan 04, 2024
Longtian Qiu, Shan Ning, Xuming He

Viaarxiv icon

A Challenger to GPT-4V? Early Explorations of Gemini in Visual Expertise

Dec 20, 2023
Chaoyou Fu, Renrui Zhang, Zihan Wang, Yubo Huang, Zhengye Zhang, Longtian Qiu, Gaoxiang Ye, Yunhang Shen, Mengdan Zhang, Peixian Chen, Sirui Zhao, Shaohui Lin, Deqiang Jiang, Di Yin, Peng Gao, Ke Li, Hongsheng Li, Xing Sun

Viaarxiv icon

SPHINX: The Joint Mixing of Weights, Tasks, and Visual Embeddings for Multi-modal Large Language Models

Nov 13, 2023
Ziyi Lin, Chris Liu, Renrui Zhang, Peng Gao, Longtian Qiu, Han Xiao, Han Qiu, Chen Lin, Wenqi Shao, Keqin Chen, Jiaming Han, Siyuan Huang, Yichi Zhang, Xuming He, Hongsheng Li, Yu Qiao

Viaarxiv icon

HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models

Mar 29, 2023
Shan Ning, Longtian Qiu, Yongfei Liu, Xuming He

Figure 1 for HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models
Figure 2 for HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models
Figure 3 for HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models
Figure 4 for HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models
Viaarxiv icon

CALIP: Zero-Shot Enhancement of CLIP with Parameter-free Attention

Sep 28, 2022
Ziyu Guo, Renrui Zhang, Longtian Qiu, Xianzheng Ma, Xupeng Miao, Xuming He, Bin Cui

Figure 1 for CALIP: Zero-Shot Enhancement of CLIP with Parameter-free Attention
Figure 2 for CALIP: Zero-Shot Enhancement of CLIP with Parameter-free Attention
Figure 3 for CALIP: Zero-Shot Enhancement of CLIP with Parameter-free Attention
Figure 4 for CALIP: Zero-Shot Enhancement of CLIP with Parameter-free Attention
Viaarxiv icon

VT-CLIP: Enhancing Vision-Language Models with Visual-guided Texts

Dec 04, 2021
Renrui Zhang, Longtian Qiu, Wei Zhang, Ziyao Zeng

Figure 1 for VT-CLIP: Enhancing Vision-Language Models with Visual-guided Texts
Figure 2 for VT-CLIP: Enhancing Vision-Language Models with Visual-guided Texts
Figure 3 for VT-CLIP: Enhancing Vision-Language Models with Visual-guided Texts
Figure 4 for VT-CLIP: Enhancing Vision-Language Models with Visual-guided Texts
Viaarxiv icon