Picture for Yueting Zhuang

Yueting Zhuang

AnyMS: Bottom-up Attention Decoupling for Layout-guided and Training-free Multi-subject Customization

Add code
Dec 29, 2025
Viaarxiv icon

FlowDC: Flow-Based Decoupling-Decay for Complex Image Editing

Add code
Dec 12, 2025
Viaarxiv icon

OmniEAR: Benchmarking Agent Reasoning in Embodied Tasks

Add code
Aug 07, 2025
Viaarxiv icon

What Limits Virtual Agent Application? OmniBench: A Scalable Multi-Dimensional Benchmark for Essential Virtual Agent Capabilities

Add code
Jun 10, 2025
Figure 1 for What Limits Virtual Agent Application? OmniBench: A Scalable Multi-Dimensional Benchmark for Essential Virtual Agent Capabilities
Figure 2 for What Limits Virtual Agent Application? OmniBench: A Scalable Multi-Dimensional Benchmark for Essential Virtual Agent Capabilities
Figure 3 for What Limits Virtual Agent Application? OmniBench: A Scalable Multi-Dimensional Benchmark for Essential Virtual Agent Capabilities
Figure 4 for What Limits Virtual Agent Application? OmniBench: A Scalable Multi-Dimensional Benchmark for Essential Virtual Agent Capabilities
Viaarxiv icon

Heartcare Suite: Multi-dimensional Understanding of ECG with Raw Multi-lead Signal Modeling

Add code
Jun 06, 2025
Viaarxiv icon

MoA: Heterogeneous Mixture of Adapters for Parameter-Efficient Fine-Tuning of Large Language Models

Add code
Jun 06, 2025
Viaarxiv icon

EOC-Bench: Can MLLMs Identify, Recall, and Forecast Objects in an Egocentric World?

Add code
Jun 05, 2025
Figure 1 for EOC-Bench: Can MLLMs Identify, Recall, and Forecast Objects in an Egocentric World?
Figure 2 for EOC-Bench: Can MLLMs Identify, Recall, and Forecast Objects in an Egocentric World?
Figure 3 for EOC-Bench: Can MLLMs Identify, Recall, and Forecast Objects in an Egocentric World?
Figure 4 for EOC-Bench: Can MLLMs Identify, Recall, and Forecast Objects in an Egocentric World?
Viaarxiv icon

FocusDiff: Advancing Fine-Grained Text-Image Alignment for Autoregressive Visual Generation through RL

Add code
Jun 05, 2025
Viaarxiv icon

ViewSpatial-Bench: Evaluating Multi-perspective Spatial Localization in Vision-Language Models

Add code
May 27, 2025
Figure 1 for ViewSpatial-Bench: Evaluating Multi-perspective Spatial Localization in Vision-Language Models
Figure 2 for ViewSpatial-Bench: Evaluating Multi-perspective Spatial Localization in Vision-Language Models
Figure 3 for ViewSpatial-Bench: Evaluating Multi-perspective Spatial Localization in Vision-Language Models
Figure 4 for ViewSpatial-Bench: Evaluating Multi-perspective Spatial Localization in Vision-Language Models
Viaarxiv icon

Let LLMs Break Free from Overthinking via Self-Braking Tuning

Add code
May 21, 2025
Viaarxiv icon