Picture for Xinghao Chen

Xinghao Chen

and Other Contributors

Single Domain Generalization for Few-Shot Counting via Universal Representation Matching

Add code
May 22, 2025
Figure 1 for Single Domain Generalization for Few-Shot Counting via Universal Representation Matching
Figure 2 for Single Domain Generalization for Few-Shot Counting via Universal Representation Matching
Figure 3 for Single Domain Generalization for Few-Shot Counting via Universal Representation Matching
Figure 4 for Single Domain Generalization for Few-Shot Counting via Universal Representation Matching
Viaarxiv icon

MultiConIR: Towards multi-condition Information Retrieval

Add code
Mar 11, 2025
Figure 1 for MultiConIR: Towards multi-condition Information Retrieval
Figure 2 for MultiConIR: Towards multi-condition Information Retrieval
Figure 3 for MultiConIR: Towards multi-condition Information Retrieval
Figure 4 for MultiConIR: Towards multi-condition Information Retrieval
Viaarxiv icon

Integrating Chain-of-Thought for Multimodal Alignment: A Study on 3D Vision-Language Learning

Add code
Mar 08, 2025
Viaarxiv icon

Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning

Add code
Feb 25, 2025
Figure 1 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 2 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 3 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 4 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Viaarxiv icon

GenVidBench: A Challenging Benchmark for Detecting AI-Generated Video

Add code
Jan 20, 2025
Figure 1 for GenVidBench: A Challenging Benchmark for Detecting AI-Generated Video
Figure 2 for GenVidBench: A Challenging Benchmark for Detecting AI-Generated Video
Figure 3 for GenVidBench: A Challenging Benchmark for Detecting AI-Generated Video
Figure 4 for GenVidBench: A Challenging Benchmark for Detecting AI-Generated Video
Viaarxiv icon

Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language Model

Add code
Dec 02, 2024
Figure 1 for Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language Model
Figure 2 for Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language Model
Figure 3 for Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language Model
Figure 4 for Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language Model
Viaarxiv icon

TinyViM: Frequency Decoupling for Tiny Hybrid Vision Mamba

Add code
Nov 26, 2024
Figure 1 for TinyViM: Frequency Decoupling for Tiny Hybrid Vision Mamba
Figure 2 for TinyViM: Frequency Decoupling for Tiny Hybrid Vision Mamba
Figure 3 for TinyViM: Frequency Decoupling for Tiny Hybrid Vision Mamba
Figure 4 for TinyViM: Frequency Decoupling for Tiny Hybrid Vision Mamba
Viaarxiv icon

Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye

Add code
Oct 29, 2024
Figure 1 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Figure 2 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Figure 3 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Figure 4 for Instruction-Tuned LLMs Succeed in Document-Level MT Without Fine-Tuning -- But BLEU Turns a Blind Eye
Viaarxiv icon

The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models

Add code
Oct 09, 2024
Viaarxiv icon

Full-Stage Pseudo Label Quality Enhancement for Weakly-supervised Temporal Action Localization

Add code
Jul 12, 2024
Viaarxiv icon