Picture for Qingcheng Zeng

Qingcheng Zeng

NAACL: Noise-AwAre Verbal Confidence Calibration for LLMs in RAG Systems

Add code
Jan 16, 2026
Viaarxiv icon

The Confidence Dichotomy: Analyzing and Mitigating Miscalibration in Tool-Use Agents

Add code
Jan 12, 2026
Viaarxiv icon

Toward Global Large Language Models in Medicine

Add code
Jan 05, 2026
Viaarxiv icon

Good Intentions Beyond ACL: Who Does NLP for Social Good, and Where?

Add code
Oct 06, 2025
Viaarxiv icon

DeepSieve: Information Sieving via LLM-as-a-Knowledge-Router

Add code
Jul 30, 2025
Viaarxiv icon

Seeing is Believing, but How Much? A Comprehensive Analysis of Verbalized Calibration in Vision-Language Models

Add code
May 26, 2025
Viaarxiv icon

The Pragmatic Mind of Machines: Tracing the Emergence of Pragmatic Competence in Large Language Models

Add code
May 24, 2025
Viaarxiv icon

CARES: Comprehensive Evaluation of Safety and Adversarial Robustness in Medical LLMs

Add code
May 16, 2025
Viaarxiv icon

Large Language Models Are More Persuasive Than Incentivized Human Persuaders

Add code
May 14, 2025
Figure 1 for Large Language Models Are More Persuasive Than Incentivized Human Persuaders
Figure 2 for Large Language Models Are More Persuasive Than Incentivized Human Persuaders
Figure 3 for Large Language Models Are More Persuasive Than Incentivized Human Persuaders
Figure 4 for Large Language Models Are More Persuasive Than Incentivized Human Persuaders
Viaarxiv icon

Do Reasoning Models Show Better Verbalized Calibration?

Add code
Apr 09, 2025
Viaarxiv icon