Picture for Jieyu Zhao

Jieyu Zhao

Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective

Add code
Jul 03, 2024
Figure 1 for Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective
Figure 2 for Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective
Figure 3 for Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective
Figure 4 for Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective
Viaarxiv icon

Rethinking LLM-based Preference Evaluation

Add code
Jul 01, 2024
Viaarxiv icon

"You Gotta be a Doctor, Lin": An Investigation of Name-Based Bias of Large Language Models in Employment Recommendations

Add code
Jun 18, 2024
Viaarxiv icon

InterIntent: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context

Add code
Jun 18, 2024
Figure 1 for InterIntent: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context
Figure 2 for InterIntent: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context
Figure 3 for InterIntent: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context
Figure 4 for InterIntent: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context
Viaarxiv icon

TrustLLM: Trustworthiness in Large Language Models

Add code
Jan 25, 2024
Figure 1 for TrustLLM: Trustworthiness in Large Language Models
Figure 2 for TrustLLM: Trustworthiness in Large Language Models
Figure 3 for TrustLLM: Trustworthiness in Large Language Models
Figure 4 for TrustLLM: Trustworthiness in Large Language Models
Viaarxiv icon

Multilingual large language models leak human stereotypes across language boundaries

Add code
Dec 12, 2023
Figure 1 for Multilingual large language models leak human stereotypes across language boundaries
Figure 2 for Multilingual large language models leak human stereotypes across language boundaries
Figure 3 for Multilingual large language models leak human stereotypes across language boundaries
Figure 4 for Multilingual large language models leak human stereotypes across language boundaries
Viaarxiv icon

SCORE: A framework for Self-Contradictory Reasoning Evaluation

Add code
Nov 16, 2023
Figure 1 for SCORE: A framework for Self-Contradictory Reasoning Evaluation
Figure 2 for SCORE: A framework for Self-Contradictory Reasoning Evaluation
Figure 3 for SCORE: A framework for Self-Contradictory Reasoning Evaluation
Figure 4 for SCORE: A framework for Self-Contradictory Reasoning Evaluation
Viaarxiv icon

Safer-Instruct: Aligning Language Models with Automated Preference Data

Add code
Nov 15, 2023
Viaarxiv icon

Fair Abstractive Summarization of Diverse Perspectives

Add code
Nov 14, 2023
Figure 1 for Fair Abstractive Summarization of Diverse Perspectives
Figure 2 for Fair Abstractive Summarization of Diverse Perspectives
Figure 3 for Fair Abstractive Summarization of Diverse Perspectives
Figure 4 for Fair Abstractive Summarization of Diverse Perspectives
Viaarxiv icon

Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems

Add code
Oct 23, 2023
Figure 1 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Figure 2 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Figure 3 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Figure 4 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Viaarxiv icon