Picture for Cheng-Yu Hsieh

Cheng-Yu Hsieh

Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps

Add code
Jul 09, 2024
Figure 1 for Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
Figure 2 for Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
Figure 3 for Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
Figure 4 for Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
Viaarxiv icon

Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions

Add code
Jul 09, 2024
Figure 1 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Figure 2 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Figure 3 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Figure 4 for Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions
Viaarxiv icon

Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization

Add code
Jun 23, 2024
Figure 1 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Figure 2 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Figure 3 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Figure 4 for Found in the Middle: Calibrating Positional Attention Bias Improves Long Context Utilization
Viaarxiv icon

DataComp-LM: In search of the next generation of training sets for language models

Add code
Jun 18, 2024
Figure 1 for DataComp-LM: In search of the next generation of training sets for language models
Figure 2 for DataComp-LM: In search of the next generation of training sets for language models
Figure 3 for DataComp-LM: In search of the next generation of training sets for language models
Figure 4 for DataComp-LM: In search of the next generation of training sets for language models
Viaarxiv icon

The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better

Add code
Jun 07, 2024
Figure 1 for The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better
Figure 2 for The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better
Figure 3 for The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better
Figure 4 for The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better
Viaarxiv icon

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

Add code
Oct 08, 2023
Figure 1 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Figure 2 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Figure 3 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Figure 4 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Viaarxiv icon

Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models

Add code
Aug 01, 2023
Figure 1 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Figure 2 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Figure 3 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Figure 4 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Viaarxiv icon

SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality

Add code
Jun 26, 2023
Figure 1 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Figure 2 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Figure 3 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Figure 4 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Viaarxiv icon

Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes

Add code
May 03, 2023
Figure 1 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Figure 2 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Figure 3 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Figure 4 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Viaarxiv icon

Understanding Programmatic Weak Supervision via Source-aware Influence Function

Add code
May 25, 2022
Figure 1 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Figure 2 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Figure 3 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Figure 4 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Viaarxiv icon