Picture for Zheng Lian

Zheng Lian

Are MLMs Trapped in the Visual Room?

Add code
May 29, 2025
Viaarxiv icon

MMAR: A Challenging Benchmark for Deep Reasoning in Speech, Audio, Music, and Their Mix

Add code
May 19, 2025
Viaarxiv icon

$\mathcal{A}LLM4ADD$: Unlocking the Capabilities of Audio Large Language Models for Audio Deepfake Detection

Add code
May 16, 2025
Viaarxiv icon

Emotion-Qwen: Training Hybrid Experts for Unified Emotion and General Vision-Language Understanding

Add code
May 10, 2025
Viaarxiv icon

EmoBench-M: Benchmarking Emotional Intelligence for Multimodal Large Language Models

Add code
Feb 06, 2025
Figure 1 for EmoBench-M: Benchmarking Emotional Intelligence for Multimodal Large Language Models
Figure 2 for EmoBench-M: Benchmarking Emotional Intelligence for Multimodal Large Language Models
Figure 3 for EmoBench-M: Benchmarking Emotional Intelligence for Multimodal Large Language Models
Figure 4 for EmoBench-M: Benchmarking Emotional Intelligence for Multimodal Large Language Models
Viaarxiv icon

SarcasmBench: Towards Evaluating Large Language Models on Sarcasm Understanding

Add code
Aug 24, 2024
Figure 1 for SarcasmBench: Towards Evaluating Large Language Models on Sarcasm Understanding
Figure 2 for SarcasmBench: Towards Evaluating Large Language Models on Sarcasm Understanding
Figure 3 for SarcasmBench: Towards Evaluating Large Language Models on Sarcasm Understanding
Figure 4 for SarcasmBench: Towards Evaluating Large Language Models on Sarcasm Understanding
Viaarxiv icon

Towards Evaluating Large Language Models on Sarcasm Understanding

Add code
Aug 21, 2024
Figure 1 for Towards Evaluating Large Language Models on Sarcasm Understanding
Figure 2 for Towards Evaluating Large Language Models on Sarcasm Understanding
Figure 3 for Towards Evaluating Large Language Models on Sarcasm Understanding
Figure 4 for Towards Evaluating Large Language Models on Sarcasm Understanding
Viaarxiv icon

Emotion and Intent Joint Understanding in Multimodal Conversation: A Benchmarking Dataset

Add code
Jul 03, 2024
Viaarxiv icon

Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning

Add code
Jun 17, 2024
Figure 1 for Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning
Figure 2 for Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning
Figure 3 for Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning
Figure 4 for Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning
Viaarxiv icon

MER 2024: Semi-Supervised Learning, Noise Robustness, and Open-Vocabulary Multimodal Emotion Recognition

Add code
Apr 29, 2024
Figure 1 for MER 2024: Semi-Supervised Learning, Noise Robustness, and Open-Vocabulary Multimodal Emotion Recognition
Figure 2 for MER 2024: Semi-Supervised Learning, Noise Robustness, and Open-Vocabulary Multimodal Emotion Recognition
Figure 3 for MER 2024: Semi-Supervised Learning, Noise Robustness, and Open-Vocabulary Multimodal Emotion Recognition
Figure 4 for MER 2024: Semi-Supervised Learning, Noise Robustness, and Open-Vocabulary Multimodal Emotion Recognition
Viaarxiv icon