Picture for Dongyan Zhao

Dongyan Zhao

Understanding Multimodal Hallucination with Parameter-Free Representation Alignment

Add code
Sep 02, 2024
Figure 1 for Understanding Multimodal Hallucination with Parameter-Free Representation Alignment
Figure 2 for Understanding Multimodal Hallucination with Parameter-Free Representation Alignment
Figure 3 for Understanding Multimodal Hallucination with Parameter-Free Representation Alignment
Figure 4 for Understanding Multimodal Hallucination with Parameter-Free Representation Alignment
Viaarxiv icon

ReMamba: Equip Mamba with Effective Long-Sequence Modeling

Add code
Sep 01, 2024
Figure 1 for ReMamba: Equip Mamba with Effective Long-Sequence Modeling
Figure 2 for ReMamba: Equip Mamba with Effective Long-Sequence Modeling
Figure 3 for ReMamba: Equip Mamba with Effective Long-Sequence Modeling
Figure 4 for ReMamba: Equip Mamba with Effective Long-Sequence Modeling
Viaarxiv icon

Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering

Add code
Aug 27, 2024
Figure 1 for Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering
Figure 2 for Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering
Figure 3 for Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering
Figure 4 for Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering
Viaarxiv icon

Internal and External Knowledge Interactive Refinement Framework for Knowledge-Intensive Question Answering

Add code
Aug 23, 2024
Figure 1 for Internal and External Knowledge Interactive Refinement Framework for Knowledge-Intensive Question Answering
Figure 2 for Internal and External Knowledge Interactive Refinement Framework for Knowledge-Intensive Question Answering
Figure 3 for Internal and External Knowledge Interactive Refinement Framework for Knowledge-Intensive Question Answering
Figure 4 for Internal and External Knowledge Interactive Refinement Framework for Knowledge-Intensive Question Answering
Viaarxiv icon

In-Context Learning with Reinforcement Learning for Incomplete Utterance Rewriting

Add code
Aug 23, 2024
Figure 1 for In-Context Learning with Reinforcement Learning for Incomplete Utterance Rewriting
Figure 2 for In-Context Learning with Reinforcement Learning for Incomplete Utterance Rewriting
Figure 3 for In-Context Learning with Reinforcement Learning for Incomplete Utterance Rewriting
Figure 4 for In-Context Learning with Reinforcement Learning for Incomplete Utterance Rewriting
Viaarxiv icon

Graph-Structured Speculative Decoding

Add code
Jul 23, 2024
Figure 1 for Graph-Structured Speculative Decoding
Figure 2 for Graph-Structured Speculative Decoding
Figure 3 for Graph-Structured Speculative Decoding
Figure 4 for Graph-Structured Speculative Decoding
Viaarxiv icon

End-to-End Video Question Answering with Frame Scoring Mechanisms and Adaptive Sampling

Add code
Jul 23, 2024
Figure 1 for End-to-End Video Question Answering with Frame Scoring Mechanisms and Adaptive Sampling
Figure 2 for End-to-End Video Question Answering with Frame Scoring Mechanisms and Adaptive Sampling
Figure 3 for End-to-End Video Question Answering with Frame Scoring Mechanisms and Adaptive Sampling
Figure 4 for End-to-End Video Question Answering with Frame Scoring Mechanisms and Adaptive Sampling
Viaarxiv icon

Mixture-of-Modules: Reinventing Transformers as Dynamic Assemblies of Modules

Add code
Jul 09, 2024
Figure 1 for Mixture-of-Modules: Reinventing Transformers as Dynamic Assemblies of Modules
Figure 2 for Mixture-of-Modules: Reinventing Transformers as Dynamic Assemblies of Modules
Figure 3 for Mixture-of-Modules: Reinventing Transformers as Dynamic Assemblies of Modules
Figure 4 for Mixture-of-Modules: Reinventing Transformers as Dynamic Assemblies of Modules
Viaarxiv icon

Unlocking the Potential of Model Merging for Low-Resource Languages

Add code
Jul 04, 2024
Figure 1 for Unlocking the Potential of Model Merging for Low-Resource Languages
Figure 2 for Unlocking the Potential of Model Merging for Low-Resource Languages
Figure 3 for Unlocking the Potential of Model Merging for Low-Resource Languages
Figure 4 for Unlocking the Potential of Model Merging for Low-Resource Languages
Viaarxiv icon

VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models

Add code
Jun 24, 2024
Figure 1 for VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models
Figure 2 for VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models
Figure 3 for VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models
Figure 4 for VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models
Viaarxiv icon