Picture for Yirong Sun

Yirong Sun

Think-as-You-See: Streaming Chain-of-Thought Reasoning for Large Vision-Language Models

Add code
Mar 03, 2026
Viaarxiv icon

Rethinking the Role of LLMs in Time Series Forecasting

Add code
Feb 16, 2026
Viaarxiv icon

SonicBench: Dissecting the Physical Perception Bottleneck in Large Audio Language Models

Add code
Jan 16, 2026
Viaarxiv icon

The Few Govern the Many:Unveiling Few-Layer Dominance for Time Series Models

Add code
Nov 10, 2025
Viaarxiv icon

PRISM: Preference Refinement via Implicit Scene Modeling for 3D Vision-Language Preference-Based Reinforcement Learning

Add code
Mar 13, 2025
Figure 1 for PRISM: Preference Refinement via Implicit Scene Modeling for 3D Vision-Language Preference-Based Reinforcement Learning
Figure 2 for PRISM: Preference Refinement via Implicit Scene Modeling for 3D Vision-Language Preference-Based Reinforcement Learning
Figure 3 for PRISM: Preference Refinement via Implicit Scene Modeling for 3D Vision-Language Preference-Based Reinforcement Learning
Figure 4 for PRISM: Preference Refinement via Implicit Scene Modeling for 3D Vision-Language Preference-Based Reinforcement Learning
Viaarxiv icon

Integrating Chain-of-Thought for Multimodal Alignment: A Study on 3D Vision-Language Learning

Add code
Mar 08, 2025
Figure 1 for Integrating Chain-of-Thought for Multimodal Alignment: A Study on 3D Vision-Language Learning
Figure 2 for Integrating Chain-of-Thought for Multimodal Alignment: A Study on 3D Vision-Language Learning
Figure 3 for Integrating Chain-of-Thought for Multimodal Alignment: A Study on 3D Vision-Language Learning
Figure 4 for Integrating Chain-of-Thought for Multimodal Alignment: A Study on 3D Vision-Language Learning
Viaarxiv icon

Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning

Add code
Feb 25, 2025
Figure 1 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 2 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 3 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 4 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Viaarxiv icon

Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye

Add code
Oct 29, 2024
Figure 1 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Figure 2 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Figure 3 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Figure 4 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Viaarxiv icon

The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models

Add code
Oct 09, 2024
Figure 1 for The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models
Figure 2 for The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models
Figure 3 for The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models
Figure 4 for The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models
Viaarxiv icon