Picture for Chunyuan Li

Chunyuan Li

Multimodal Foundation Models: From Specialists to General-Purpose Assistants

Add code
Sep 18, 2023
Viaarxiv icon

Benchmarking and Analyzing Generative Data for Visual Recognition

Add code
Jul 25, 2023
Viaarxiv icon

Semantic-SAM: Segment and Recognize Anything at Any Granularity

Add code
Jul 10, 2023
Figure 1 for Semantic-SAM: Segment and Recognize Anything at Any Granularity
Figure 2 for Semantic-SAM: Segment and Recognize Anything at Any Granularity
Figure 3 for Semantic-SAM: Segment and Recognize Anything at Any Granularity
Figure 4 for Semantic-SAM: Segment and Recognize Anything at Any Granularity
Viaarxiv icon

Large Multimodal Models: Notes on CVPR 2023 Tutorial

Add code
Jun 26, 2023
Viaarxiv icon

MIMIC-IT: Multi-Modal In-Context Instruction Tuning

Add code
Jun 08, 2023
Figure 1 for MIMIC-IT: Multi-Modal In-Context Instruction Tuning
Figure 2 for MIMIC-IT: Multi-Modal In-Context Instruction Tuning
Figure 3 for MIMIC-IT: Multi-Modal In-Context Instruction Tuning
Figure 4 for MIMIC-IT: Multi-Modal In-Context Instruction Tuning
Viaarxiv icon

LLaVA-Med: Training a Large Language-and-Vision Assistant for Biomedicine in One Day

Add code
Jun 01, 2023
Viaarxiv icon

On the Hidden Mystery of OCR in Large Multimodal Models

Add code
May 13, 2023
Viaarxiv icon

Towards Building the Federated GPT: Federated Instruction Tuning

Add code
May 09, 2023
Figure 1 for Towards Building the Federated GPT: Federated Instruction Tuning
Figure 2 for Towards Building the Federated GPT: Federated Instruction Tuning
Figure 3 for Towards Building the Federated GPT: Federated Instruction Tuning
Figure 4 for Towards Building the Federated GPT: Federated Instruction Tuning
Viaarxiv icon

Visual Instruction Tuning

Add code
Apr 17, 2023
Viaarxiv icon

Instruction Tuning with GPT-4

Add code
Apr 06, 2023
Viaarxiv icon