Alert button
Picture for Zhihong Chen

Zhihong Chen

Alert button

Large Multimodal Agents: A Survey

Feb 23, 2024
Junlin Xie, Zhihong Chen, Ruifei Zhang, Xiang Wan, Guanbin Li

Viaarxiv icon

ALLaVA: Harnessing GPT4V-synthesized Data for A Lite Vision-Language Model

Feb 18, 2024
Guiming Hardy Chen, Shunian Chen, Ruifei Zhang, Junying Chen, Xiangbo Wu, Zhiyi Zhang, Zhihong Chen, Jianquan Li, Xiang Wan, Benyou Wang

Viaarxiv icon

CheXagent: Towards a Foundation Model for Chest X-Ray Interpretation

Jan 22, 2024
Zhihong Chen, Maya Varma, Jean-Benoit Delbrouck, Magdalini Paschali, Louis Blankemeier, Dave Van Veen, Jeya Maria Jose Valanarasu, Alaa Youssef, Joseph Paul Cohen, Eduardo Pontes Reis, Emily B. Tsai, Andrew Johnston, Cameron Olsen, Tanishq Mathew Abraham, Sergios Gatidis, Akshay S. Chaudhari, Curtis Langlotz

Viaarxiv icon

MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V

Nov 23, 2023
Wentao Ge, Shunian Chen, Guiming Chen, Junying Chen, Zhihong Chen, Shuo Yan, Chenghao Zhu, Ziyue Lin, Wenya Xie, Xidong Wang, Anningzhe Gao, Zhiyi Zhang, Jianquan Li, Xiang Wan, Benyou Wang

Figure 1 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Figure 2 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Figure 3 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Figure 4 for MLLM-Bench, Evaluating Multi-modal LLMs using GPT-4V
Viaarxiv icon

Exploiting Low-confidence Pseudo-labels for Source-free Object Detection

Oct 19, 2023
Zhihong Chen, Zilei Wang, Yixin Zhang

Figure 1 for Exploiting Low-confidence Pseudo-labels for Source-free Object Detection
Figure 2 for Exploiting Low-confidence Pseudo-labels for Source-free Object Detection
Figure 3 for Exploiting Low-confidence Pseudo-labels for Source-free Object Detection
Figure 4 for Exploiting Low-confidence Pseudo-labels for Source-free Object Detection
Viaarxiv icon

AceGPT, Localizing Large Language Models in Arabic

Sep 22, 2023
Huang Huang, Fei Yu, Jianqing Zhu, Xuening Sun, Hao Cheng, Dingjie Song, Zhihong Chen, Abdulmohsen Alharthi, Bang An, Ziche Liu, Zhiyi Zhang, Junying Chen, Jianquan Li, Benyou Wang, Lian Zhang, Ruoyu Sun, Xiang Wan, Haizhou Li, Jinchao Xu

Figure 1 for AceGPT, Localizing Large Language Models in Arabic
Figure 2 for AceGPT, Localizing Large Language Models in Arabic
Figure 3 for AceGPT, Localizing Large Language Models in Arabic
Figure 4 for AceGPT, Localizing Large Language Models in Arabic
Viaarxiv icon

CMB: A Comprehensive Medical Benchmark in Chinese

Aug 17, 2023
Xidong Wang, Guiming Hardy Chen, Dingjie Song, Zhiyi Zhang, Zhihong Chen, Qingying Xiao, Feng Jiang, Jianquan Li, Xiang Wan, Benyou Wang, Haizhou Li

Figure 1 for CMB: A Comprehensive Medical Benchmark in Chinese
Figure 2 for CMB: A Comprehensive Medical Benchmark in Chinese
Figure 3 for CMB: A Comprehensive Medical Benchmark in Chinese
Figure 4 for CMB: A Comprehensive Medical Benchmark in Chinese
Viaarxiv icon

Advancing Visual Grounding with Scene Knowledge: Benchmark and Method

Jul 21, 2023
Zhihong Chen, Ruifei Zhang, Yibing Song, Xiang Wan, Guanbin Li

Figure 1 for Advancing Visual Grounding with Scene Knowledge: Benchmark and Method
Figure 2 for Advancing Visual Grounding with Scene Knowledge: Benchmark and Method
Figure 3 for Advancing Visual Grounding with Scene Knowledge: Benchmark and Method
Figure 4 for Advancing Visual Grounding with Scene Knowledge: Benchmark and Method
Viaarxiv icon

Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation

Jul 21, 2023
Zunnan Xu, Zhihong Chen, Yong Zhang, Yibing Song, Xiang Wan, Guanbin Li

Figure 1 for Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation
Figure 2 for Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation
Figure 3 for Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation
Figure 4 for Bridging Vision and Language Encoders: Parameter-Efficient Tuning for Referring Image Segmentation
Viaarxiv icon

On the Difference of BERT-style and CLIP-style Text Encoders

Jun 06, 2023
Zhihong Chen, Guiming Hardy Chen, Shizhe Diao, Xiang Wan, Benyou Wang

Figure 1 for On the Difference of BERT-style and CLIP-style Text Encoders
Figure 2 for On the Difference of BERT-style and CLIP-style Text Encoders
Figure 3 for On the Difference of BERT-style and CLIP-style Text Encoders
Figure 4 for On the Difference of BERT-style and CLIP-style Text Encoders
Viaarxiv icon