Alert button
Picture for Pengchuan Zhang

Pengchuan Zhang

Alert button

DIME-FM: DIstilling Multimodal and Efficient Foundation Models

Add code
Bookmark button
Alert button
Mar 31, 2023
Ximeng Sun, Pengchuan Zhang, Peizhao Zhang, Hardik Shah, Kate Saenko, Xide Xia

Figure 1 for DIME-FM: DIstilling Multimodal and Efficient Foundation Models
Figure 2 for DIME-FM: DIstilling Multimodal and Efficient Foundation Models
Figure 3 for DIME-FM: DIstilling Multimodal and Efficient Foundation Models
Figure 4 for DIME-FM: DIstilling Multimodal and Efficient Foundation Models
Viaarxiv icon

A Unified Model for Tracking and Image-Video Detection Has More Power

Add code
Bookmark button
Alert button
Nov 20, 2022
Peirong Liu, Rui Wang, Pengchuan Zhang, Omid Poursaeed, Yipin Zhou, Xuefei Cao, Sreya Dutta Roy, Ashish Shah, Ser-Nam Lim

Figure 1 for A Unified Model for Tracking and Image-Video Detection Has More Power
Figure 2 for A Unified Model for Tracking and Image-Video Detection Has More Power
Figure 3 for A Unified Model for Tracking and Image-Video Detection Has More Power
Figure 4 for A Unified Model for Tracking and Image-Video Detection Has More Power
Viaarxiv icon

Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone

Add code
Bookmark button
Alert button
Jun 15, 2022
Zi-Yi Dou, Aishwarya Kamath, Zhe Gan, Pengchuan Zhang, Jianfeng Wang, Linjie Li, Zicheng Liu, Ce Liu, Yann LeCun, Nanyun Peng, Jianfeng Gao, Lijuan Wang

Figure 1 for Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
Figure 2 for Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
Figure 3 for Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
Figure 4 for Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
Viaarxiv icon

GLIPv2: Unifying Localization and Vision-Language Understanding

Add code
Bookmark button
Alert button
Jun 12, 2022
Haotian Zhang, Pengchuan Zhang, Xiaowei Hu, Yen-Chun Chen, Liunian Harold Li, Xiyang Dai, Lijuan Wang, Lu Yuan, Jenq-Neng Hwang, Jianfeng Gao

Figure 1 for GLIPv2: Unifying Localization and Vision-Language Understanding
Figure 2 for GLIPv2: Unifying Localization and Vision-Language Understanding
Figure 3 for GLIPv2: Unifying Localization and Vision-Language Understanding
Figure 4 for GLIPv2: Unifying Localization and Vision-Language Understanding
Viaarxiv icon

Detection Hub: Unifying Object Detection Datasets via Query Adaptation on Language Embedding

Add code
Bookmark button
Alert button
Jun 07, 2022
Lingchen Meng, Xiyang Dai, Yinpeng Chen, Pengchuan Zhang, Dongdong Chen, Mengchen Liu, Jianfeng Wang, Zuxuan Wu, Lu Yuan, Yu-Gang Jiang

Figure 1 for Detection Hub: Unifying Object Detection Datasets via Query Adaptation on Language Embedding
Figure 2 for Detection Hub: Unifying Object Detection Datasets via Query Adaptation on Language Embedding
Figure 3 for Detection Hub: Unifying Object Detection Datasets via Query Adaptation on Language Embedding
Figure 4 for Detection Hub: Unifying Object Detection Datasets via Query Adaptation on Language Embedding
Viaarxiv icon

ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models

Add code
Bookmark button
Alert button
Apr 20, 2022
Chunyuan Li, Haotian Liu, Liunian Harold Li, Pengchuan Zhang, Jyoti Aneja, Jianwei Yang, Ping Jin, Yong Jae Lee, Houdong Hu, Zicheng Liu, Jianfeng Gao

Figure 1 for ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models
Figure 2 for ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models
Figure 3 for ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models
Figure 4 for ELEVATER: A Benchmark and Toolkit for Evaluating Language-Augmented Visual Models
Viaarxiv icon

K-LITE: Learning Transferable Visual Models with External Knowledge

Add code
Bookmark button
Alert button
Apr 20, 2022
Sheng Shen, Chunyuan Li, Xiaowei Hu, Yujia Xie, Jianwei Yang, Pengchuan Zhang, Anna Rohrbach, Zhe Gan, Lijuan Wang, Lu Yuan, Ce Liu, Kurt Keutzer, Trevor Darrell, Jianfeng Gao

Figure 1 for K-LITE: Learning Transferable Visual Models with External Knowledge
Figure 2 for K-LITE: Learning Transferable Visual Models with External Knowledge
Figure 3 for K-LITE: Learning Transferable Visual Models with External Knowledge
Figure 4 for K-LITE: Learning Transferable Visual Models with External Knowledge
Viaarxiv icon

Missingness Bias in Model Debugging

Add code
Bookmark button
Alert button
Apr 19, 2022
Saachi Jain, Hadi Salman, Eric Wong, Pengchuan Zhang, Vibhav Vineet, Sai Vemprala, Aleksander Madry

Figure 1 for Missingness Bias in Model Debugging
Figure 2 for Missingness Bias in Model Debugging
Figure 3 for Missingness Bias in Model Debugging
Figure 4 for Missingness Bias in Model Debugging
Viaarxiv icon

Unified Contrastive Learning in Image-Text-Label Space

Add code
Bookmark button
Alert button
Apr 07, 2022
Jianwei Yang, Chunyuan Li, Pengchuan Zhang, Bin Xiao, Ce Liu, Lu Yuan, Jianfeng Gao

Figure 1 for Unified Contrastive Learning in Image-Text-Label Space
Figure 2 for Unified Contrastive Learning in Image-Text-Label Space
Figure 3 for Unified Contrastive Learning in Image-Text-Label Space
Figure 4 for Unified Contrastive Learning in Image-Text-Label Space
Viaarxiv icon

Parameter-efficient Fine-tuning for Vision Transformers

Add code
Bookmark button
Alert button
Mar 29, 2022
Xuehai He, Chunyuan Li, Pengchuan Zhang, Jianwei Yang, Xin Eric Wang

Figure 1 for Parameter-efficient Fine-tuning for Vision Transformers
Figure 2 for Parameter-efficient Fine-tuning for Vision Transformers
Figure 3 for Parameter-efficient Fine-tuning for Vision Transformers
Figure 4 for Parameter-efficient Fine-tuning for Vision Transformers
Viaarxiv icon