Picture for Hui Su

Hui Su

Investigating and Enhancing the Robustness of Large Multimodal Models Against Temporal Inconsistency

Add code
May 20, 2025
Viaarxiv icon

Rethinking Visual Layer Selection in Multimodal LLMs

Add code
Apr 30, 2025
Viaarxiv icon

MultiConIR: Towards multi-condition Information Retrieval

Add code
Mar 11, 2025
Figure 1 for MultiConIR: Towards multi-condition Information Retrieval
Figure 2 for MultiConIR: Towards multi-condition Information Retrieval
Figure 3 for MultiConIR: Towards multi-condition Information Retrieval
Figure 4 for MultiConIR: Towards multi-condition Information Retrieval
Viaarxiv icon

Multi-Layer Visual Feature Fusion in Multimodal LLMs: Methods, Analysis, and Best Practices

Add code
Mar 08, 2025
Viaarxiv icon

Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning

Add code
Feb 25, 2025
Figure 1 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 2 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 3 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Figure 4 for Unveiling the Key Factors for Distilling Chain-of-Thought Reasoning
Viaarxiv icon

Detecting Errors through Ensembling Prompts (DEEP): An End-to-End LLM Framework for Detecting Factual Errors

Add code
Jun 18, 2024
Viaarxiv icon

Unraveling the Mystery of Scaling Laws: Part I

Add code
Mar 21, 2024
Figure 1 for Unraveling the Mystery of Scaling Laws: Part I
Figure 2 for Unraveling the Mystery of Scaling Laws: Part I
Figure 3 for Unraveling the Mystery of Scaling Laws: Part I
Viaarxiv icon

SASFormer: Transformers for Sparsely Annotated Semantic Segmentation

Add code
Dec 06, 2022
Figure 1 for SASFormer: Transformers for Sparsely Annotated Semantic Segmentation
Figure 2 for SASFormer: Transformers for Sparsely Annotated Semantic Segmentation
Figure 3 for SASFormer: Transformers for Sparsely Annotated Semantic Segmentation
Figure 4 for SASFormer: Transformers for Sparsely Annotated Semantic Segmentation
Viaarxiv icon

WeLM: A Well-Read Pre-trained Language Model for Chinese

Add code
Oct 12, 2022
Figure 1 for WeLM: A Well-Read Pre-trained Language Model for Chinese
Figure 2 for WeLM: A Well-Read Pre-trained Language Model for Chinese
Figure 3 for WeLM: A Well-Read Pre-trained Language Model for Chinese
Figure 4 for WeLM: A Well-Read Pre-trained Language Model for Chinese
Viaarxiv icon

Re-Attention Transformer for Weakly Supervised Object Localization

Add code
Aug 03, 2022
Viaarxiv icon