Picture for Hong-You Chen

Hong-You Chen

DREAM: Where Visual Understanding Meets Text-to-Image Generation

Add code
Mar 03, 2026
Viaarxiv icon

Xray-Visual Models: Scaling Vision models on Industry Scale Data

Add code
Feb 18, 2026
Viaarxiv icon

Principled Synthetic Data Enables the First Scaling Laws for LLMs in Recommendation

Add code
Feb 07, 2026
Viaarxiv icon

Federated Inverse Probability Treatment Weighting for Individual Treatment Effect Estimation

Add code
Mar 06, 2025
Viaarxiv icon

CLIP-UP: A Simple and Efficient Mixture-of-Experts CLIP Training Recipe with Sparse Upcycling

Add code
Feb 03, 2025
Figure 1 for CLIP-UP: A Simple and Efficient Mixture-of-Experts CLIP Training Recipe with Sparse Upcycling
Figure 2 for CLIP-UP: A Simple and Efficient Mixture-of-Experts CLIP Training Recipe with Sparse Upcycling
Figure 3 for CLIP-UP: A Simple and Efficient Mixture-of-Experts CLIP Training Recipe with Sparse Upcycling
Figure 4 for CLIP-UP: A Simple and Efficient Mixture-of-Experts CLIP Training Recipe with Sparse Upcycling
Viaarxiv icon

Contrastive Localized Language-Image Pre-Training

Add code
Oct 03, 2024
Figure 1 for Contrastive Localized Language-Image Pre-Training
Figure 2 for Contrastive Localized Language-Image Pre-Training
Figure 3 for Contrastive Localized Language-Image Pre-Training
Figure 4 for Contrastive Localized Language-Image Pre-Training
Viaarxiv icon

Revisit Large-Scale Image-Caption Data in Pre-training Multimodal Foundation Models

Add code
Oct 03, 2024
Viaarxiv icon

MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning

Add code
Sep 30, 2024
Figure 1 for MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning
Figure 2 for MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning
Figure 3 for MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning
Figure 4 for MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-tuning
Viaarxiv icon

Fine-Tuning is Fine, if Calibrated

Add code
Sep 24, 2024
Figure 1 for Fine-Tuning is Fine, if Calibrated
Figure 2 for Fine-Tuning is Fine, if Calibrated
Figure 3 for Fine-Tuning is Fine, if Calibrated
Figure 4 for Fine-Tuning is Fine, if Calibrated
Viaarxiv icon

Lessons Learned from a Unifying Empirical Study of Parameter-Efficient Transfer Learning (PETL) in Visual Recognition

Add code
Sep 24, 2024
Figure 1 for Lessons Learned from a Unifying Empirical Study of Parameter-Efficient Transfer Learning (PETL) in Visual Recognition
Figure 2 for Lessons Learned from a Unifying Empirical Study of Parameter-Efficient Transfer Learning (PETL) in Visual Recognition
Figure 3 for Lessons Learned from a Unifying Empirical Study of Parameter-Efficient Transfer Learning (PETL) in Visual Recognition
Figure 4 for Lessons Learned from a Unifying Empirical Study of Parameter-Efficient Transfer Learning (PETL) in Visual Recognition
Viaarxiv icon