Alert button
Picture for Zefan Cai

Zefan Cai

Alert button

VeCAF: VLM-empowered Collaborative Active Finetuning with Training Objective Awareness

Jan 15, 2024
Rongyu Zhang, Zefan Cai, Huanrui Yang, Zidong Liu, Denis Gudovskiy, Tomoyuki Okuno, Yohei Nakata, Kurt Keutzer, Baobao Chang, Yuan Du, Li Du, Shanghang Zhang

Viaarxiv icon

ML-Bench: Large Language Models Leverage Open-source Libraries for Machine Learning Tasks

Nov 16, 2023
Yuliang Liu, Xiangru Tang, Zefan Cai, Junjie Lu, Yichi Zhang, Yanjun Shao, Zexuan Deng, Helan Hu, Zengxian Yang, Kaikai An, Ruijun Huang, Shuzheng Si, Sheng Chen, Haozhe Zhao, Zhengliang Li, Liang Chen, Yiming Zong, Yan Wang, Tianyu Liu, Zhiwei Jiang, Baobao Chang, Yujia Qin, Wangchunshu Zhou, Yilun Zhao, Arman Cohan, Mark Gerstein

Viaarxiv icon

Distantly-Supervised Named Entity Recognition with Uncertainty-aware Teacher Learning and Student-student Collaborative Learning

Nov 14, 2023
Helan Hu, Shuzheng Si, Haozhe Zhao, Shuang Zeng, Kaikai An, Zefan Cai, Baobao Chang

Viaarxiv icon

Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond

Oct 16, 2023
Liang Chen, Yichi Zhang, Shuhuai Ren, Haozhe Zhao, Zefan Cai, Yuchi Wang, Peiyi Wang, Tianyu Liu, Baobao Chang

Figure 1 for Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond
Figure 2 for Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond
Figure 3 for Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond
Figure 4 for Towards End-to-End Embodied Decision Making via Multi-modal Large Language Model: Explorations with GPT4-Vision and Beyond
Viaarxiv icon

MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning

Oct 02, 2023
Haozhe Zhao, Zefan Cai, Shuzheng Si, Xiaojian Ma, Kaikai An, Liang Chen, Zixuan Liu, Sheng Wang, Wenjuan Han, Baobao Chang

Figure 1 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Figure 2 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Figure 3 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Figure 4 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Viaarxiv icon

Human-in-the-Loop through Chain-of-Thought

Jun 23, 2023
Zefan Cai, Baobao Chang, Wenjuan Han

Figure 1 for Human-in-the-Loop through Chain-of-Thought
Figure 2 for Human-in-the-Loop through Chain-of-Thought
Figure 3 for Human-in-the-Loop through Chain-of-Thought
Figure 4 for Human-in-the-Loop through Chain-of-Thought
Viaarxiv icon

DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade

May 24, 2023
Zefan Cai, Xin Zheng, Tianyu Liu, Xu Wang, Haoran Meng, Jiaqi Han, Gang Yuan, Binghuai Lin, Baobao Chang, Yunbo Cao

Figure 1 for DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade
Figure 2 for DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade
Figure 3 for DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade
Figure 4 for DialogVCS: Robust Natural Language Understanding in Dialogue System Upgrade
Viaarxiv icon