Picture for Lang Gao

Lang Gao

M3MAD-Bench: Are Multi-Agent Debates Really Effective Across Domains and Modalities?

Add code
Jan 06, 2026
Viaarxiv icon

When Personalization Tricks Detectors: The Feature-Inversion Trap in Machine-Generated Text Detection

Add code
Oct 14, 2025
Figure 1 for When Personalization Tricks Detectors: The Feature-Inversion Trap in Machine-Generated Text Detection
Figure 2 for When Personalization Tricks Detectors: The Feature-Inversion Trap in Machine-Generated Text Detection
Figure 3 for When Personalization Tricks Detectors: The Feature-Inversion Trap in Machine-Generated Text Detection
Figure 4 for When Personalization Tricks Detectors: The Feature-Inversion Trap in Machine-Generated Text Detection
Viaarxiv icon

SocialMaze: A Benchmark for Evaluating Social Reasoning in Large Language Models

Add code
May 29, 2025
Figure 1 for SocialMaze: A Benchmark for Evaluating Social Reasoning in Large Language Models
Figure 2 for SocialMaze: A Benchmark for Evaluating Social Reasoning in Large Language Models
Figure 3 for SocialMaze: A Benchmark for Evaluating Social Reasoning in Large Language Models
Figure 4 for SocialMaze: A Benchmark for Evaluating Social Reasoning in Large Language Models
Viaarxiv icon

Evaluate Bias without Manual Test Sets: A Concept Representation Perspective for LLMs

Add code
May 21, 2025
Figure 1 for Evaluate Bias without Manual Test Sets: A Concept Representation Perspective for LLMs
Figure 2 for Evaluate Bias without Manual Test Sets: A Concept Representation Perspective for LLMs
Figure 3 for Evaluate Bias without Manual Test Sets: A Concept Representation Perspective for LLMs
Figure 4 for Evaluate Bias without Manual Test Sets: A Concept Representation Perspective for LLMs
Viaarxiv icon

Audio Jailbreak: An Open Comprehensive Benchmark for Jailbreaking Large Audio-Language Models

Add code
May 21, 2025
Figure 1 for Audio Jailbreak: An Open Comprehensive Benchmark for Jailbreaking Large Audio-Language Models
Figure 2 for Audio Jailbreak: An Open Comprehensive Benchmark for Jailbreaking Large Audio-Language Models
Figure 3 for Audio Jailbreak: An Open Comprehensive Benchmark for Jailbreaking Large Audio-Language Models
Figure 4 for Audio Jailbreak: An Open Comprehensive Benchmark for Jailbreaking Large Audio-Language Models
Viaarxiv icon

Adversarial Cooperative Rationalization: The Risk of Spurious Correlations in Even Clean Datasets

Add code
May 04, 2025
Figure 1 for Adversarial Cooperative Rationalization: The Risk of Spurious Correlations in Even Clean Datasets
Figure 2 for Adversarial Cooperative Rationalization: The Risk of Spurious Correlations in Even Clean Datasets
Figure 3 for Adversarial Cooperative Rationalization: The Risk of Spurious Correlations in Even Clean Datasets
Figure 4 for Adversarial Cooperative Rationalization: The Risk of Spurious Correlations in Even Clean Datasets
Viaarxiv icon

Word Form Matters: LLMs' Semantic Reconstruction under Typoglycemia

Add code
Mar 03, 2025
Viaarxiv icon

Shaping the Safety Boundaries: Understanding and Defending Against Jailbreaks in Large Language Models

Add code
Dec 22, 2024
Figure 1 for Shaping the Safety Boundaries: Understanding and Defending Against Jailbreaks in Large Language Models
Figure 2 for Shaping the Safety Boundaries: Understanding and Defending Against Jailbreaks in Large Language Models
Figure 3 for Shaping the Safety Boundaries: Understanding and Defending Against Jailbreaks in Large Language Models
Figure 4 for Shaping the Safety Boundaries: Understanding and Defending Against Jailbreaks in Large Language Models
Viaarxiv icon

MedTrinity-25M: A Large-scale Multimodal Dataset with Multigranular Annotations for Medicine

Add code
Aug 06, 2024
Viaarxiv icon