Picture for Mingyi Hong

Mingyi Hong

AgentDS Technical Report: Benchmarking the Future of Human-AI Collaboration in Domain-Specific Data Science

Add code
Mar 19, 2026
Viaarxiv icon

StitchCUDA: An Automated Multi-Agents End-to-End GPU Programing Framework with Rubric-based Agentic Reinforcement Learning

Add code
Mar 03, 2026
Viaarxiv icon

Powering Up Zeroth-Order Training via Subspace Gradient Orthogonalization

Add code
Feb 19, 2026
Viaarxiv icon

HiPER: Hierarchical Reinforcement Learning with Explicit Credit Assignment for Large Language Model Agents

Add code
Feb 18, 2026
Viaarxiv icon

DISPO: Enhancing Training Efficiency and Stability in Reinforcement Learning for Large Language Model Mathematical Reasoning

Add code
Feb 01, 2026
Viaarxiv icon

Scaling Unverifiable Rewards: A Case Study on Visual Insights

Add code
Dec 27, 2025
Viaarxiv icon

Can Agentic AI Match the Performance of Human Data Scientists?

Add code
Dec 24, 2025
Figure 1 for Can Agentic AI Match the Performance of Human Data Scientists?
Figure 2 for Can Agentic AI Match the Performance of Human Data Scientists?
Figure 3 for Can Agentic AI Match the Performance of Human Data Scientists?
Figure 4 for Can Agentic AI Match the Performance of Human Data Scientists?
Viaarxiv icon

Leak@$k$: Unlearning Does Not Make LLMs Forget Under Probabilistic Decoding

Add code
Nov 07, 2025
Viaarxiv icon

A Framework for Quantifying How Pre-Training and Context Benefit In-Context Learning

Add code
Oct 26, 2025
Figure 1 for A Framework for Quantifying How Pre-Training and Context Benefit In-Context Learning
Figure 2 for A Framework for Quantifying How Pre-Training and Context Benefit In-Context Learning
Figure 3 for A Framework for Quantifying How Pre-Training and Context Benefit In-Context Learning
Figure 4 for A Framework for Quantifying How Pre-Training and Context Benefit In-Context Learning
Viaarxiv icon

BLUR: A Bi-Level Optimization Approach for LLM Unlearning

Add code
Jun 09, 2025
Viaarxiv icon