Picture for Honglak Lee

Honglak Lee

University of Michigan, Ann Arbor

Towards Scalable Language-Image Pre-training for 3D Medical Imaging

Add code
May 28, 2025
Viaarxiv icon

SafeDPO: A Simple Approach to Direct Preference Optimization with Enhanced Safety

Add code
May 26, 2025
Viaarxiv icon

Scalable Video-to-Dataset Generation for Cross-Platform Mobile Agents

Add code
May 19, 2025
Viaarxiv icon

Visual Test-time Scaling for GUI Agent Grounding

Add code
May 01, 2025
Figure 1 for Visual Test-time Scaling for GUI Agent Grounding
Figure 2 for Visual Test-time Scaling for GUI Agent Grounding
Figure 3 for Visual Test-time Scaling for GUI Agent Grounding
Figure 4 for Visual Test-time Scaling for GUI Agent Grounding
Viaarxiv icon

Process Reward Models That Think

Add code
Apr 23, 2025
Figure 1 for Process Reward Models That Think
Figure 2 for Process Reward Models That Think
Figure 3 for Process Reward Models That Think
Figure 4 for Process Reward Models That Think
Viaarxiv icon

MLRC-Bench: Can Language Agents Solve Machine Learning Research Challenges?

Add code
Apr 13, 2025
Figure 1 for MLRC-Bench: Can Language Agents Solve Machine Learning Research Challenges?
Figure 2 for MLRC-Bench: Can Language Agents Solve Machine Learning Research Challenges?
Figure 3 for MLRC-Bench: Can Language Agents Solve Machine Learning Research Challenges?
Figure 4 for MLRC-Bench: Can Language Agents Solve Machine Learning Research Challenges?
Viaarxiv icon

EXAONE Deep: Reasoning Enhanced Language Models

Add code
Mar 16, 2025
Viaarxiv icon

Do Not Trust Licenses You See -- Dataset Compliance Requires Massive-Scale AI-Powered Lifecycle Tracing

Add code
Mar 04, 2025
Viaarxiv icon

Subtask-Aware Visual Reward Learning from Segmented Demonstrations

Add code
Feb 28, 2025
Figure 1 for Subtask-Aware Visual Reward Learning from Segmented Demonstrations
Figure 2 for Subtask-Aware Visual Reward Learning from Segmented Demonstrations
Figure 3 for Subtask-Aware Visual Reward Learning from Segmented Demonstrations
Figure 4 for Subtask-Aware Visual Reward Learning from Segmented Demonstrations
Viaarxiv icon

KL Penalty Control via Perturbation for Direct Preference Optimization

Add code
Feb 18, 2025
Figure 1 for KL Penalty Control via Perturbation for Direct Preference Optimization
Figure 2 for KL Penalty Control via Perturbation for Direct Preference Optimization
Figure 3 for KL Penalty Control via Perturbation for Direct Preference Optimization
Figure 4 for KL Penalty Control via Perturbation for Direct Preference Optimization
Viaarxiv icon