Picture for Xuansheng Wu

Xuansheng Wu

Investigating CoT Monitorability in Large Reasoning Models

Add code
Nov 13, 2025
Figure 1 for Investigating CoT Monitorability in Large Reasoning Models
Figure 2 for Investigating CoT Monitorability in Large Reasoning Models
Figure 3 for Investigating CoT Monitorability in Large Reasoning Models
Figure 4 for Investigating CoT Monitorability in Large Reasoning Models
Viaarxiv icon

AutoSCORE: Enhancing Automated Scoring with Multi-Agent Large Language Models via Structured Component Recognition

Add code
Sep 26, 2025
Figure 1 for AutoSCORE: Enhancing Automated Scoring with Multi-Agent Large Language Models via Structured Component Recognition
Figure 2 for AutoSCORE: Enhancing Automated Scoring with Multi-Agent Large Language Models via Structured Component Recognition
Figure 3 for AutoSCORE: Enhancing Automated Scoring with Multi-Agent Large Language Models via Structured Component Recognition
Viaarxiv icon

Is Long-to-Short a Free Lunch? Investigating Inconsistency and Reasoning Efficiency in LRMs

Add code
Jun 24, 2025
Viaarxiv icon

Denoising Concept Vectors with Sparse Autoencoders for Improved Language Model Steering

Add code
May 21, 2025
Viaarxiv icon

Artificial Intelligence Bias on English Language Learners in Automatic Scoring

Add code
May 15, 2025
Viaarxiv icon

Beyond Input Activations: Identifying Influential Latents by Gradient Sparse Autoencoders

Add code
May 12, 2025
Viaarxiv icon

A Survey on Sparse Autoencoders: Interpreting the Internal Mechanisms of Large Language Models

Add code
Mar 07, 2025
Viaarxiv icon

Interpreting and Steering LLMs with Mutual Information-based Explanations on Sparse Autoencoders

Add code
Feb 21, 2025
Figure 1 for Interpreting and Steering LLMs with Mutual Information-based Explanations on Sparse Autoencoders
Figure 2 for Interpreting and Steering LLMs with Mutual Information-based Explanations on Sparse Autoencoders
Figure 3 for Interpreting and Steering LLMs with Mutual Information-based Explanations on Sparse Autoencoders
Figure 4 for Interpreting and Steering LLMs with Mutual Information-based Explanations on Sparse Autoencoders
Viaarxiv icon

Self-Regularization with Latent Space Explanations for Controllable LLM-based Classification

Add code
Feb 19, 2025
Viaarxiv icon

LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models

Add code
Oct 02, 2024
Figure 1 for LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models
Figure 2 for LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models
Figure 3 for LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models
Figure 4 for LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models
Viaarxiv icon