Picture for Yong Jae Lee

Yong Jae Lee

Yo'LLaVA: Your Personalized Language and Vision Assistant

Add code
Jun 13, 2024
Viaarxiv icon

Matryoshka Multimodal Models

Add code
May 27, 2024
Viaarxiv icon

LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Models

Add code
Apr 01, 2024
Viaarxiv icon

LLM Inference Unveiled: Survey and Roofline Model Insights

Add code
Mar 11, 2024
Viaarxiv icon

Cohere3D: Exploiting Temporal Coherence for Unsupervised Representation Learning of Vision-based Autonomous Driving

Feb 23, 2024
Viaarxiv icon

CounterCurate: Enhancing Physical and Semantic Visio-Linguistic Compositional Reasoning via Counterfactual Examples

Add code
Feb 20, 2024
Figure 1 for CounterCurate: Enhancing Physical and Semantic Visio-Linguistic Compositional Reasoning via Counterfactual Examples
Figure 2 for CounterCurate: Enhancing Physical and Semantic Visio-Linguistic Compositional Reasoning via Counterfactual Examples
Figure 3 for CounterCurate: Enhancing Physical and Semantic Visio-Linguistic Compositional Reasoning via Counterfactual Examples
Figure 4 for CounterCurate: Enhancing Physical and Semantic Visio-Linguistic Compositional Reasoning via Counterfactual Examples
Viaarxiv icon

Edit One for All: Interactive Batch Image Editing

Add code
Jan 18, 2024
Viaarxiv icon

Interfacing Foundation Models' Embeddings

Add code
Dec 12, 2023
Viaarxiv icon

Diversify, Don't Fine-Tune: Scaling Up Visual Recognition Training with Synthetic Images

Add code
Dec 04, 2023
Viaarxiv icon

Making Large Multimodal Models Understand Arbitrary Visual Prompts

Add code
Dec 01, 2023
Figure 1 for Making Large Multimodal Models Understand Arbitrary Visual Prompts
Figure 2 for Making Large Multimodal Models Understand Arbitrary Visual Prompts
Figure 3 for Making Large Multimodal Models Understand Arbitrary Visual Prompts
Figure 4 for Making Large Multimodal Models Understand Arbitrary Visual Prompts
Viaarxiv icon