Picture for Lei Li

Lei Li

Carnegie Mellon University

MedFLIP: Medical Vision-and-Language Self-supervised Fast Pre-Training with Masked Autoencoder

Add code
Mar 07, 2024
Viaarxiv icon

SnapNTell: Enhancing Entity-Centric Visual Question Answering with Retrieval Augmented Multimodal LLM

Add code
Mar 07, 2024
Figure 1 for SnapNTell: Enhancing Entity-Centric Visual Question Answering with Retrieval Augmented Multimodal LLM
Figure 2 for SnapNTell: Enhancing Entity-Centric Visual Question Answering with Retrieval Augmented Multimodal LLM
Figure 3 for SnapNTell: Enhancing Entity-Centric Visual Question Answering with Retrieval Augmented Multimodal LLM
Figure 4 for SnapNTell: Enhancing Entity-Centric Visual Question Answering with Retrieval Augmented Multimodal LLM
Viaarxiv icon

ImgTrojan: Jailbreaking Vision-Language Models with ONE Image

Add code
Mar 06, 2024
Viaarxiv icon

Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models

Add code
Mar 04, 2024
Figure 1 for Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
Figure 2 for Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
Figure 3 for Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
Figure 4 for Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models
Viaarxiv icon

TempCompass: Do Video LLMs Really Understand Videos?

Add code
Mar 01, 2024
Figure 1 for TempCompass: Do Video LLMs Really Understand Videos?
Figure 2 for TempCompass: Do Video LLMs Really Understand Videos?
Figure 3 for TempCompass: Do Video LLMs Really Understand Videos?
Figure 4 for TempCompass: Do Video LLMs Really Understand Videos?
Viaarxiv icon

Hire a Linguist!: Learning Endangered Languages with In-Context Linguistic Descriptions

Add code
Feb 28, 2024
Viaarxiv icon

Structure-Based Drug Design via 3D Molecular Generative Pre-training and Sampling

Add code
Feb 22, 2024
Viaarxiv icon

Where It Really Matters: Few-Shot Environmental Conservation Media Monitoring for Low-Resource Languages

Add code
Feb 19, 2024
Figure 1 for Where It Really Matters: Few-Shot Environmental Conservation Media Monitoring for Low-Resource Languages
Figure 2 for Where It Really Matters: Few-Shot Environmental Conservation Media Monitoring for Low-Resource Languages
Figure 3 for Where It Really Matters: Few-Shot Environmental Conservation Media Monitoring for Low-Resource Languages
Figure 4 for Where It Really Matters: Few-Shot Environmental Conservation Media Monitoring for Low-Resource Languages
Viaarxiv icon

Perils of Self-Feedback: Self-Bias Amplifies in Large Language Models

Add code
Feb 18, 2024
Viaarxiv icon

DE-COP: Detecting Copyrighted Content in Language Models Training Data

Add code
Feb 15, 2024
Figure 1 for DE-COP: Detecting Copyrighted Content in Language Models Training Data
Figure 2 for DE-COP: Detecting Copyrighted Content in Language Models Training Data
Figure 3 for DE-COP: Detecting Copyrighted Content in Language Models Training Data
Figure 4 for DE-COP: Detecting Copyrighted Content in Language Models Training Data
Viaarxiv icon