Picture for Yufang Hou

Yufang Hou

Beyond Outcome Verification: Verifiable Process Reward Models for Structured Reasoning

Add code
Jan 23, 2026
Viaarxiv icon

Face-Voice Association with Inductive Bias for Maximum Class Separation

Add code
Jan 20, 2026
Viaarxiv icon

FactCorrector: A Graph-Inspired Approach to Long-Form Factuality Correction of Large Language Models

Add code
Jan 16, 2026
Viaarxiv icon

Where Knowledge Collides: A Mechanistic Study of Intra-Memory Knowledge Conflict in Language Models

Add code
Jan 14, 2026
Viaarxiv icon

LLMs as Science Journalists: Supporting Early-stage Researchers in Communicating Their Science to the Public

Add code
Jan 09, 2026
Viaarxiv icon

MathBuddy: A Multimodal System for Affective Math Tutoring

Add code
Aug 27, 2025
Viaarxiv icon

Enhancing Study-Level Inference from Clinical Trial Papers via RL-based Numeric Reasoning

Add code
May 28, 2025
Viaarxiv icon

A Position Paper on the Automatic Generation of Machine Learning Leaderboards

Add code
May 23, 2025
Viaarxiv icon

Query-driven Document-level Scientific Evidence Extraction from Biomedical Studies

Add code
May 09, 2025
Figure 1 for Query-driven Document-level Scientific Evidence Extraction from Biomedical Studies
Figure 2 for Query-driven Document-level Scientific Evidence Extraction from Biomedical Studies
Figure 3 for Query-driven Document-level Scientific Evidence Extraction from Biomedical Studies
Figure 4 for Query-driven Document-level Scientific Evidence Extraction from Biomedical Studies
Viaarxiv icon

FactReasoner: A Probabilistic Approach to Long-Form Factuality Assessment for Large Language Models

Add code
Feb 25, 2025
Figure 1 for FactReasoner: A Probabilistic Approach to Long-Form Factuality Assessment for Large Language Models
Figure 2 for FactReasoner: A Probabilistic Approach to Long-Form Factuality Assessment for Large Language Models
Figure 3 for FactReasoner: A Probabilistic Approach to Long-Form Factuality Assessment for Large Language Models
Figure 4 for FactReasoner: A Probabilistic Approach to Long-Form Factuality Assessment for Large Language Models
Viaarxiv icon