Picture for Yusen Zhang

Yusen Zhang

Coverage-based Fairness in Multi-document Summarization

Add code
Dec 11, 2024
Figure 1 for Coverage-based Fairness in Multi-document Summarization
Figure 2 for Coverage-based Fairness in Multi-document Summarization
Figure 3 for Coverage-based Fairness in Multi-document Summarization
Figure 4 for Coverage-based Fairness in Multi-document Summarization
Viaarxiv icon

VisOnlyQA: Large Vision Language Models Still Struggle with Visual Perception of Geometric Information

Add code
Dec 01, 2024
Figure 1 for VisOnlyQA: Large Vision Language Models Still Struggle with Visual Perception of Geometric Information
Figure 2 for VisOnlyQA: Large Vision Language Models Still Struggle with Visual Perception of Geometric Information
Figure 3 for VisOnlyQA: Large Vision Language Models Still Struggle with Visual Perception of Geometric Information
Figure 4 for VisOnlyQA: Large Vision Language Models Still Struggle with Visual Perception of Geometric Information
Viaarxiv icon

Verbosity $ eq$ Veracity: Demystify Verbosity Compensation Behavior of Large Language Models

Add code
Nov 12, 2024
Figure 1 for Verbosity $ eq$ Veracity: Demystify Verbosity Compensation Behavior of Large Language Models
Figure 2 for Verbosity $ eq$ Veracity: Demystify Verbosity Compensation Behavior of Large Language Models
Figure 3 for Verbosity $ eq$ Veracity: Demystify Verbosity Compensation Behavior of Large Language Models
Figure 4 for Verbosity $ eq$ Veracity: Demystify Verbosity Compensation Behavior of Large Language Models
Viaarxiv icon

AAAR-1.0: Assessing AI's Potential to Assist Research

Add code
Oct 29, 2024
Figure 1 for AAAR-1.0: Assessing AI's Potential to Assist Research
Figure 2 for AAAR-1.0: Assessing AI's Potential to Assist Research
Figure 3 for AAAR-1.0: Assessing AI's Potential to Assist Research
Figure 4 for AAAR-1.0: Assessing AI's Potential to Assist Research
Viaarxiv icon

Chain-of-Scrutiny: Detecting Backdoor Attacks for Large Language Models

Add code
Jun 10, 2024
Figure 1 for Chain-of-Scrutiny: Detecting Backdoor Attacks for Large Language Models
Figure 2 for Chain-of-Scrutiny: Detecting Backdoor Attacks for Large Language Models
Figure 3 for Chain-of-Scrutiny: Detecting Backdoor Attacks for Large Language Models
Figure 4 for Chain-of-Scrutiny: Detecting Backdoor Attacks for Large Language Models
Viaarxiv icon

Chain of Agents: Large Language Models Collaborating on Long-Context Tasks

Add code
Jun 04, 2024
Figure 1 for Chain of Agents: Large Language Models Collaborating on Long-Context Tasks
Figure 2 for Chain of Agents: Large Language Models Collaborating on Long-Context Tasks
Figure 3 for Chain of Agents: Large Language Models Collaborating on Long-Context Tasks
Figure 4 for Chain of Agents: Large Language Models Collaborating on Long-Context Tasks
Viaarxiv icon

When Can LLMs Actually Correct Their Own Mistakes? A Critical Survey of Self-Correction of LLMs

Add code
Jun 03, 2024
Viaarxiv icon

Evaluating LLMs at Detecting Errors in LLM Responses

Add code
Apr 04, 2024
Figure 1 for Evaluating LLMs at Detecting Errors in LLM Responses
Figure 2 for Evaluating LLMs at Detecting Errors in LLM Responses
Figure 3 for Evaluating LLMs at Detecting Errors in LLM Responses
Figure 4 for Evaluating LLMs at Detecting Errors in LLM Responses
Viaarxiv icon

A General Benchmark Framework is Dynamic Graph Neural Network Need

Add code
Jan 12, 2024
Viaarxiv icon

Fair Abstractive Summarization of Diverse Perspectives

Add code
Nov 14, 2023
Figure 1 for Fair Abstractive Summarization of Diverse Perspectives
Figure 2 for Fair Abstractive Summarization of Diverse Perspectives
Figure 3 for Fair Abstractive Summarization of Diverse Perspectives
Figure 4 for Fair Abstractive Summarization of Diverse Perspectives
Viaarxiv icon