Picture for Min-Yen Kan

Min-Yen Kan

Columbia University

V-DPO: Mitigating Hallucination in Large Vision Language Models via Vision-Guided Direct Preference Optimization

Add code
Nov 05, 2024
Figure 1 for V-DPO: Mitigating Hallucination in Large Vision Language Models via Vision-Guided Direct Preference Optimization
Figure 2 for V-DPO: Mitigating Hallucination in Large Vision Language Models via Vision-Guided Direct Preference Optimization
Figure 3 for V-DPO: Mitigating Hallucination in Large Vision Language Models via Vision-Guided Direct Preference Optimization
Figure 4 for V-DPO: Mitigating Hallucination in Large Vision Language Models via Vision-Guided Direct Preference Optimization
Viaarxiv icon

Multi-expert Prompting Improves Reliability, Safety, and Usefulness of Large Language Models

Add code
Nov 01, 2024
Figure 1 for Multi-expert Prompting Improves Reliability, Safety, and Usefulness of Large Language Models
Figure 2 for Multi-expert Prompting Improves Reliability, Safety, and Usefulness of Large Language Models
Figure 3 for Multi-expert Prompting Improves Reliability, Safety, and Usefulness of Large Language Models
Figure 4 for Multi-expert Prompting Improves Reliability, Safety, and Usefulness of Large Language Models
Viaarxiv icon

DataTales: A Benchmark for Real-World Intelligent Data Narration

Add code
Oct 23, 2024
Figure 1 for DataTales: A Benchmark for Real-World Intelligent Data Narration
Figure 2 for DataTales: A Benchmark for Real-World Intelligent Data Narration
Figure 3 for DataTales: A Benchmark for Real-World Intelligent Data Narration
Figure 4 for DataTales: A Benchmark for Real-World Intelligent Data Narration
Viaarxiv icon

CCSBench: Evaluating Compositional Controllability in LLMs for Scientific Document Summarization

Add code
Oct 16, 2024
Figure 1 for CCSBench: Evaluating Compositional Controllability in LLMs for Scientific Document Summarization
Figure 2 for CCSBench: Evaluating Compositional Controllability in LLMs for Scientific Document Summarization
Figure 3 for CCSBench: Evaluating Compositional Controllability in LLMs for Scientific Document Summarization
Figure 4 for CCSBench: Evaluating Compositional Controllability in LLMs for Scientific Document Summarization
Viaarxiv icon

COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement

Add code
Oct 12, 2024
Figure 1 for COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement
Figure 2 for COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement
Figure 3 for COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement
Figure 4 for COrAL: Order-Agnostic Language Modeling for Efficient Iterative Refinement
Viaarxiv icon

MVP-Bench: Can Large Vision--Language Models Conduct Multi-level Visual Perception Like Humans?

Add code
Oct 06, 2024
Figure 1 for MVP-Bench: Can Large Vision--Language Models Conduct Multi-level Visual Perception Like Humans?
Figure 2 for MVP-Bench: Can Large Vision--Language Models Conduct Multi-level Visual Perception Like Humans?
Figure 3 for MVP-Bench: Can Large Vision--Language Models Conduct Multi-level Visual Perception Like Humans?
Figure 4 for MVP-Bench: Can Large Vision--Language Models Conduct Multi-level Visual Perception Like Humans?
Viaarxiv icon

TART: An Open-Source Tool-Augmented Framework for Explainable Table-based Reasoning

Add code
Sep 18, 2024
Figure 1 for TART: An Open-Source Tool-Augmented Framework for Explainable Table-based Reasoning
Figure 2 for TART: An Open-Source Tool-Augmented Framework for Explainable Table-based Reasoning
Figure 3 for TART: An Open-Source Tool-Augmented Framework for Explainable Table-based Reasoning
Figure 4 for TART: An Open-Source Tool-Augmented Framework for Explainable Table-based Reasoning
Viaarxiv icon

LLMs Are Biased Towards Output Formats! Systematically Evaluating and Mitigating Output Format Bias of LLMs

Add code
Aug 16, 2024
Viaarxiv icon

The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models

Add code
Jun 14, 2024
Figure 1 for The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models
Figure 2 for The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models
Figure 3 for The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models
Figure 4 for The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models
Viaarxiv icon

Decompose and Aggregate: A Step-by-Step Interpretable Evaluation Framework

Add code
May 24, 2024
Figure 1 for Decompose and Aggregate: A Step-by-Step Interpretable Evaluation Framework
Figure 2 for Decompose and Aggregate: A Step-by-Step Interpretable Evaluation Framework
Figure 3 for Decompose and Aggregate: A Step-by-Step Interpretable Evaluation Framework
Figure 4 for Decompose and Aggregate: A Step-by-Step Interpretable Evaluation Framework
Viaarxiv icon