Alert button
Picture for Xiaodan Liang

Xiaodan Liang

Alert button

Boosting Visual-Language Models by Exploiting Hard Samples

Add code
Bookmark button
Alert button
May 09, 2023
Haonan Wang, Minbin Huang, Runhui Huang, Lanqing Hong, Hang Xu, Tianyang Hu, Xiaodan Liang, Zhenguo Li

Figure 1 for Boosting Visual-Language Models by Exploiting Hard Samples
Figure 2 for Boosting Visual-Language Models by Exploiting Hard Samples
Figure 3 for Boosting Visual-Language Models by Exploiting Hard Samples
Figure 4 for Boosting Visual-Language Models by Exploiting Hard Samples
Viaarxiv icon

Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining

Add code
Bookmark button
Alert button
Apr 26, 2023
Bingqian Lin, Zicong Chen, Mingjie Li, Haokun Lin, Hang Xu, Yi Zhu, Jianzhuang Liu, Wenjia Cai, Lei Yang, Shen Zhao, Chenfei Wu, Ling Chen, Xiaojun Chang, Yi Yang, Lei Xing, Xiaodan Liang

Figure 1 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 2 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 3 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 4 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Viaarxiv icon

LiDAR-NeRF: Novel LiDAR View Synthesis via Neural Radiance Fields

Add code
Bookmark button
Alert button
Apr 20, 2023
Tang Tao, Longfei Gao, Guangrun Wang, Peng Chen, Dayang Hao, Xiaodan Liang, Mathieu Salzmann, Kaicheng Yu

Figure 1 for LiDAR-NeRF: Novel LiDAR View Synthesis via Neural Radiance Fields
Figure 2 for LiDAR-NeRF: Novel LiDAR View Synthesis via Neural Radiance Fields
Figure 3 for LiDAR-NeRF: Novel LiDAR View Synthesis via Neural Radiance Fields
Figure 4 for LiDAR-NeRF: Novel LiDAR View Synthesis via Neural Radiance Fields
Viaarxiv icon

DetCLIPv2: Scalable Open-Vocabulary Object Detection Pre-training via Word-Region Alignment

Add code
Bookmark button
Alert button
Apr 10, 2023
Lewei Yao, Jianhua Han, Xiaodan Liang, Dan Xu, Wei Zhang, Zhenguo Li, Hang Xu

Figure 1 for DetCLIPv2: Scalable Open-Vocabulary Object Detection Pre-training via Word-Region Alignment
Figure 2 for DetCLIPv2: Scalable Open-Vocabulary Object Detection Pre-training via Word-Region Alignment
Figure 3 for DetCLIPv2: Scalable Open-Vocabulary Object Detection Pre-training via Word-Region Alignment
Figure 4 for DetCLIPv2: Scalable Open-Vocabulary Object Detection Pre-training via Word-Region Alignment
Viaarxiv icon

CLIP$^2$: Contrastive Language-Image-Point Pretraining from Real-World Point Cloud Data

Add code
Bookmark button
Alert button
Mar 26, 2023
Yihan Zeng, Chenhan Jiang, Jiageng Mao, Jianhua Han, Chaoqiang Ye, Qingqiu Huang, Dit-Yan Yeung, Zhen Yang, Xiaodan Liang, Hang Xu

Figure 1 for CLIP$^2$: Contrastive Language-Image-Point Pretraining from Real-World Point Cloud Data
Figure 2 for CLIP$^2$: Contrastive Language-Image-Point Pretraining from Real-World Point Cloud Data
Figure 3 for CLIP$^2$: Contrastive Language-Image-Point Pretraining from Real-World Point Cloud Data
Figure 4 for CLIP$^2$: Contrastive Language-Image-Point Pretraining from Real-World Point Cloud Data
Viaarxiv icon

GP-VTON: Towards General Purpose Virtual Try-on via Collaborative Local-Flow Global-Parsing Learning

Add code
Bookmark button
Alert button
Mar 24, 2023
Zhenyu Xie, Zaiyu Huang, Xin Dong, Fuwei Zhao, Haoye Dong, Xijin Zhang, Feida Zhu, Xiaodan Liang

Figure 1 for GP-VTON: Towards General Purpose Virtual Try-on via Collaborative Local-Flow Global-Parsing Learning
Figure 2 for GP-VTON: Towards General Purpose Virtual Try-on via Collaborative Local-Flow Global-Parsing Learning
Figure 3 for GP-VTON: Towards General Purpose Virtual Try-on via Collaborative Local-Flow Global-Parsing Learning
Figure 4 for GP-VTON: Towards General Purpose Virtual Try-on via Collaborative Local-Flow Global-Parsing Learning
Viaarxiv icon

Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation

Add code
Bookmark button
Alert button
Mar 18, 2023
Mingjie Li, Bingqian Lin, Zicong Chen, Haokun Lin, Xiaodan Liang, Xiaojun Chang

Figure 1 for Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation
Figure 2 for Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation
Figure 3 for Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation
Figure 4 for Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation
Viaarxiv icon

CapDet: Unifying Dense Captioning and Open-World Detection Pretraining

Add code
Bookmark button
Alert button
Mar 15, 2023
Yanxin Long, Youpeng Wen, Jianhua Han, Hang Xu, Pengzhen Ren, Wei Zhang, Shen Zhao, Xiaodan Liang

Figure 1 for CapDet: Unifying Dense Captioning and Open-World Detection Pretraining
Figure 2 for CapDet: Unifying Dense Captioning and Open-World Detection Pretraining
Figure 3 for CapDet: Unifying Dense Captioning and Open-World Detection Pretraining
Figure 4 for CapDet: Unifying Dense Captioning and Open-World Detection Pretraining
Viaarxiv icon

Visual Exemplar Driven Task-Prompting for Unified Perception in Autonomous Driving

Add code
Bookmark button
Alert button
Mar 03, 2023
Xiwen Liang, Minzhe Niu, Jianhua Han, Hang Xu, Chunjing Xu, Xiaodan Liang

Figure 1 for Visual Exemplar Driven Task-Prompting for Unified Perception in Autonomous Driving
Figure 2 for Visual Exemplar Driven Task-Prompting for Unified Perception in Autonomous Driving
Figure 3 for Visual Exemplar Driven Task-Prompting for Unified Perception in Autonomous Driving
Figure 4 for Visual Exemplar Driven Task-Prompting for Unified Perception in Autonomous Driving
Viaarxiv icon

Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation

Add code
Bookmark button
Alert button
Feb 13, 2023
Bingqian Lin, Yi Zhu, Xiaodan Liang, Liang Lin, Jianzhuang Liu

Figure 1 for Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation
Figure 2 for Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation
Figure 3 for Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation
Figure 4 for Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation
Viaarxiv icon