Picture for Zhendong Mao

Zhendong Mao

EmoVerse: A MLLMs-Driven Emotion Representation Dataset for Interpretable Visual Emotion Analysis

Add code
Nov 16, 2025
Figure 1 for EmoVerse: A MLLMs-Driven Emotion Representation Dataset for Interpretable Visual Emotion Analysis
Figure 2 for EmoVerse: A MLLMs-Driven Emotion Representation Dataset for Interpretable Visual Emotion Analysis
Figure 3 for EmoVerse: A MLLMs-Driven Emotion Representation Dataset for Interpretable Visual Emotion Analysis
Figure 4 for EmoVerse: A MLLMs-Driven Emotion Representation Dataset for Interpretable Visual Emotion Analysis
Viaarxiv icon

LayerEdit: Disentangled Multi-Object Editing via Conflict-Aware Multi-Layer Learning

Add code
Nov 11, 2025
Figure 1 for LayerEdit: Disentangled Multi-Object Editing via Conflict-Aware Multi-Layer Learning
Figure 2 for LayerEdit: Disentangled Multi-Object Editing via Conflict-Aware Multi-Layer Learning
Figure 3 for LayerEdit: Disentangled Multi-Object Editing via Conflict-Aware Multi-Layer Learning
Figure 4 for LayerEdit: Disentangled Multi-Object Editing via Conflict-Aware Multi-Layer Learning
Viaarxiv icon

SparseRM: A Lightweight Preference Modeling with Sparse Autoencoder

Add code
Nov 11, 2025
Figure 1 for SparseRM: A Lightweight Preference Modeling with Sparse Autoencoder
Figure 2 for SparseRM: A Lightweight Preference Modeling with Sparse Autoencoder
Figure 3 for SparseRM: A Lightweight Preference Modeling with Sparse Autoencoder
Figure 4 for SparseRM: A Lightweight Preference Modeling with Sparse Autoencoder
Viaarxiv icon

Video-LevelGauge: Investigating Contextual Positional Bias in Large Video Language Models

Add code
Aug 28, 2025
Figure 1 for Video-LevelGauge: Investigating Contextual Positional Bias in Large Video Language Models
Figure 2 for Video-LevelGauge: Investigating Contextual Positional Bias in Large Video Language Models
Figure 3 for Video-LevelGauge: Investigating Contextual Positional Bias in Large Video Language Models
Figure 4 for Video-LevelGauge: Investigating Contextual Positional Bias in Large Video Language Models
Viaarxiv icon

LongAnimation: Long Animation Generation with Dynamic Global-Local Memory

Add code
Jul 02, 2025
Viaarxiv icon

DeepResearch Bench: A Comprehensive Benchmark for Deep Research Agents

Add code
Jun 13, 2025
Viaarxiv icon

From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding

Add code
Jun 04, 2025
Figure 1 for From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding
Figure 2 for From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding
Figure 3 for From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding
Figure 4 for From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding
Viaarxiv icon

Rationales Are Not Silver Bullets: Measuring the Impact of Rationales on Model Performance and Reliability

Add code
May 30, 2025
Figure 1 for Rationales Are Not Silver Bullets: Measuring the Impact of Rationales on Model Performance and Reliability
Figure 2 for Rationales Are Not Silver Bullets: Measuring the Impact of Rationales on Model Performance and Reliability
Figure 3 for Rationales Are Not Silver Bullets: Measuring the Impact of Rationales on Model Performance and Reliability
Figure 4 for Rationales Are Not Silver Bullets: Measuring the Impact of Rationales on Model Performance and Reliability
Viaarxiv icon

MIRROR: Multi-agent Intra- and Inter-Reflection for Optimized Reasoning in Tool Learning

Add code
May 27, 2025
Viaarxiv icon

Leveraging Importance Sampling to Detach Alignment Modules from Large Language Models

Add code
May 26, 2025
Figure 1 for Leveraging Importance Sampling to Detach Alignment Modules from Large Language Models
Figure 2 for Leveraging Importance Sampling to Detach Alignment Modules from Large Language Models
Figure 3 for Leveraging Importance Sampling to Detach Alignment Modules from Large Language Models
Figure 4 for Leveraging Importance Sampling to Detach Alignment Modules from Large Language Models
Viaarxiv icon