Picture for Ziyang Luo

Ziyang Luo

MFC-Bench: Benchmarking Multimodal Fact-Checking with Large Vision-Language Models

Add code
Jun 17, 2024
Figure 1 for MFC-Bench: Benchmarking Multimodal Fact-Checking with Large Vision-Language Models
Figure 2 for MFC-Bench: Benchmarking Multimodal Fact-Checking with Large Vision-Language Models
Figure 3 for MFC-Bench: Benchmarking Multimodal Fact-Checking with Large Vision-Language Models
Figure 4 for MFC-Bench: Benchmarking Multimodal Fact-Checking with Large Vision-Language Models
Viaarxiv icon

VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs

Add code
Jun 11, 2024
Viaarxiv icon

CofiPara: A Coarse-to-fine Paradigm for Multimodal Sarcasm Target Identification with Large Multimodal Models

Add code
May 01, 2024
Figure 1 for CofiPara: A Coarse-to-fine Paradigm for Multimodal Sarcasm Target Identification with Large Multimodal Models
Figure 2 for CofiPara: A Coarse-to-fine Paradigm for Multimodal Sarcasm Target Identification with Large Multimodal Models
Figure 3 for CofiPara: A Coarse-to-fine Paradigm for Multimodal Sarcasm Target Identification with Large Multimodal Models
Figure 4 for CofiPara: A Coarse-to-fine Paradigm for Multimodal Sarcasm Target Identification with Large Multimodal Models
Viaarxiv icon

CodeHalu: Code Hallucinations in LLMs Driven by Execution-based Verification

Add code
Apr 30, 2024
Figure 1 for CodeHalu: Code Hallucinations in LLMs Driven by Execution-based Verification
Figure 2 for CodeHalu: Code Hallucinations in LLMs Driven by Execution-based Verification
Figure 3 for CodeHalu: Code Hallucinations in LLMs Driven by Execution-based Verification
Figure 4 for CodeHalu: Code Hallucinations in LLMs Driven by Execution-based Verification
Viaarxiv icon

MMCode: Evaluating Multi-Modal Code Large Language Models with Visually Rich Programming Problems

Add code
Apr 15, 2024
Figure 1 for MMCode: Evaluating Multi-Modal Code Large Language Models with Visually Rich Programming Problems
Figure 2 for MMCode: Evaluating Multi-Modal Code Large Language Models with Visually Rich Programming Problems
Figure 3 for MMCode: Evaluating Multi-Modal Code Large Language Models with Visually Rich Programming Problems
Figure 4 for MMCode: Evaluating Multi-Modal Code Large Language Models with Visually Rich Programming Problems
Viaarxiv icon

Aurora-M: The First Open Source Multilingual Language Model Red-teamed according to the U.S. Executive Order

Add code
Mar 30, 2024
Viaarxiv icon

Towards Explainable Harmful Meme Detection through Multimodal Debate between Large Language Models

Add code
Jan 24, 2024
Figure 1 for Towards Explainable Harmful Meme Detection through Multimodal Debate between Large Language Models
Figure 2 for Towards Explainable Harmful Meme Detection through Multimodal Debate between Large Language Models
Figure 3 for Towards Explainable Harmful Meme Detection through Multimodal Debate between Large Language Models
Figure 4 for Towards Explainable Harmful Meme Detection through Multimodal Debate between Large Language Models
Viaarxiv icon

GOAT-Bench: Safety Insights to Large Multimodal Models through Meme-Based Social Abuse

Add code
Jan 07, 2024
Viaarxiv icon

Beneath the Surface: Unveiling Harmful Memes with Multimodal Reasoning Distilled from Large Language Models

Add code
Dec 09, 2023
Viaarxiv icon

VSCode: General Visual Salient and Camouflaged Object Detection with 2D Prompt Learning

Add code
Nov 25, 2023
Figure 1 for VSCode: General Visual Salient and Camouflaged Object Detection with 2D Prompt Learning
Figure 2 for VSCode: General Visual Salient and Camouflaged Object Detection with 2D Prompt Learning
Figure 3 for VSCode: General Visual Salient and Camouflaged Object Detection with 2D Prompt Learning
Figure 4 for VSCode: General Visual Salient and Camouflaged Object Detection with 2D Prompt Learning
Viaarxiv icon