Alert button
Picture for Chenfei Wu

Chenfei Wu

Alert button

LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models

Add code
Bookmark button
Alert button
Apr 03, 2024
Gabriela Ben Melech Stan, Raanan Yehezkel Rohekar, Yaniv Gurwicz, Matthew Lyle Olson, Anahita Bhiwandiwalla, Estelle Aflalo, Chenfei Wu, Nan Duan, Shao-Yen Tseng, Vasudev Lal

Viaarxiv icon

Using Left and Right Brains Together: Towards Vision and Language Planning

Add code
Bookmark button
Alert button
Feb 16, 2024
Jun Cen, Chenfei Wu, Xiao Liu, Shengming Yin, Yixuan Pei, Jinglong Yang, Qifeng Chen, Nan Duan, Jianguo Zhang

Viaarxiv icon

StrokeNUWA: Tokenizing Strokes for Vector Graphic Synthesis

Add code
Bookmark button
Alert button
Jan 30, 2024
Zecheng Tang, Chenfei Wu, Zekai Zhang, Mingheng Ni, Shengming Yin, Yu Liu, Zhengyuan Yang, Lijuan Wang, Zicheng Liu, Juntao Li, Nan Duan

Viaarxiv icon

EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation

Add code
Bookmark button
Alert button
Oct 12, 2023
Wang You, Wenshan Wu, Yaobo Liang, Shaoguang Mao, Chenfei Wu, Maosong Cao, Yuzhe Cai, Yiduo Guo, Yan Xia, Furu Wei, Nan Duan

Figure 1 for EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation
Figure 2 for EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation
Figure 3 for EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation
Figure 4 for EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation
Viaarxiv icon

LayoutNUWA: Revealing the Hidden Layout Expertise of Large Language Models

Add code
Bookmark button
Alert button
Sep 19, 2023
Zecheng Tang, Chenfei Wu, Juntao Li, Nan Duan

Figure 1 for LayoutNUWA: Revealing the Hidden Layout Expertise of Large Language Models
Figure 2 for LayoutNUWA: Revealing the Hidden Layout Expertise of Large Language Models
Figure 3 for LayoutNUWA: Revealing the Hidden Layout Expertise of Large Language Models
Figure 4 for LayoutNUWA: Revealing the Hidden Layout Expertise of Large Language Models
Viaarxiv icon

ORES: Open-vocabulary Responsible Visual Synthesis

Add code
Bookmark button
Alert button
Aug 26, 2023
Minheng Ni, Chenfei Wu, Xiaodong Wang, Shengming Yin, Lijuan Wang, Zicheng Liu, Nan Duan

Figure 1 for ORES: Open-vocabulary Responsible Visual Synthesis
Figure 2 for ORES: Open-vocabulary Responsible Visual Synthesis
Figure 3 for ORES: Open-vocabulary Responsible Visual Synthesis
Figure 4 for ORES: Open-vocabulary Responsible Visual Synthesis
Viaarxiv icon

GameEval: Evaluating LLMs on Conversational Games

Add code
Bookmark button
Alert button
Aug 19, 2023
Dan Qiao, Chenfei Wu, Yaobo Liang, Juntao Li, Nan Duan

Figure 1 for GameEval: Evaluating LLMs on Conversational Games
Figure 2 for GameEval: Evaluating LLMs on Conversational Games
Figure 3 for GameEval: Evaluating LLMs on Conversational Games
Figure 4 for GameEval: Evaluating LLMs on Conversational Games
Viaarxiv icon

DragNUWA: Fine-grained Control in Video Generation by Integrating Text, Image, and Trajectory

Add code
Bookmark button
Alert button
Aug 16, 2023
Shengming Yin, Chenfei Wu, Jian Liang, Jie Shi, Houqiang Li, Gong Ming, Nan Duan

Figure 1 for DragNUWA: Fine-grained Control in Video Generation by Integrating Text, Image, and Trajectory
Figure 2 for DragNUWA: Fine-grained Control in Video Generation by Integrating Text, Image, and Trajectory
Figure 3 for DragNUWA: Fine-grained Control in Video Generation by Integrating Text, Image, and Trajectory
Figure 4 for DragNUWA: Fine-grained Control in Video Generation by Integrating Text, Image, and Trajectory
Viaarxiv icon

ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning

Add code
Bookmark button
Alert button
May 31, 2023
Xiao Xu, Bei Li, Chenfei Wu, Shao-Yen Tseng, Anahita Bhiwandiwalla, Shachar Rosenman, Vasudev Lal, Wanxiang Che, Nan Duan

Figure 1 for ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning
Figure 2 for ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning
Figure 3 for ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning
Figure 4 for ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning
Viaarxiv icon

Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining

Add code
Bookmark button
Alert button
Apr 26, 2023
Bingqian Lin, Zicong Chen, Mingjie Li, Haokun Lin, Hang Xu, Yi Zhu, Jianzhuang Liu, Wenjia Cai, Lei Yang, Shen Zhao, Chenfei Wu, Ling Chen, Xiaojun Chang, Yi Yang, Lei Xing, Xiaodan Liang

Figure 1 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 2 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 3 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 4 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Viaarxiv icon