Alert button
Picture for Jieyu Zhao

Jieyu Zhao

Alert button

TrustLLM: Trustworthiness in Large Language Models

Jan 25, 2024
Lichao Sun, Yue Huang, Haoran Wang, Siyuan Wu, Qihui Zhang, Chujie Gao, Yixin Huang, Wenhan Lyu, Yixuan Zhang, Xiner Li, Zhengliang Liu, Yixin Liu, Yijue Wang, Zhikun Zhang, Bhavya Kailkhura, Caiming Xiong, Chaowei Xiao, Chunyuan Li, Eric Xing, Furong Huang, Hao Liu, Heng Ji, Hongyi Wang, Huan Zhang, Huaxiu Yao, Manolis Kellis, Marinka Zitnik, Meng Jiang, Mohit Bansal, James Zou, Jian Pei, Jian Liu, Jianfeng Gao, Jiawei Han, Jieyu Zhao, Jiliang Tang, Jindong Wang, John Mitchell, Kai Shu, Kaidi Xu, Kai-Wei Chang, Lifang He, Lifu Huang, Michael Backes, Neil Zhenqiang Gong, Philip S. Yu, Pin-Yu Chen, Quanquan Gu, Ran Xu, Rex Ying, Shuiwang Ji, Suman Jana, Tianlong Chen, Tianming Liu, Tianyi Zhou, William Wang, Xiang Li, Xiangliang Zhang, Xiao Wang, Xing Xie, Xun Chen, Xuyu Wang, Yan Liu, Yanfang Ye, Yinzhi Cao, Yong Chen, Yue Zhao

Figure 1 for TrustLLM: Trustworthiness in Large Language Models
Figure 2 for TrustLLM: Trustworthiness in Large Language Models
Figure 3 for TrustLLM: Trustworthiness in Large Language Models
Figure 4 for TrustLLM: Trustworthiness in Large Language Models
Viaarxiv icon

Multilingual large language models leak human stereotypes across language boundaries

Dec 12, 2023
Yang Trista Cao, Anna Sotnikova, Jieyu Zhao, Linda X. Zou, Rachel Rudinger, Hal Daume III

Viaarxiv icon

SCORE: A framework for Self-Contradictory Reasoning Evaluation

Nov 16, 2023
Ziyi Liu, Isabelle Lee, Yongkang Du, Soumya Sanyal, Jieyu Zhao

Viaarxiv icon

Safer-Instruct: Aligning Language Models with Automated Preference Data

Nov 15, 2023
Taiwei Shi, Kai Chen, Jieyu Zhao

Viaarxiv icon

Fair Abstractive Summarization of Diverse Perspectives

Nov 14, 2023
Yusen Zhang, Nan Zhang, Yixin Liu, Alexander Fabbri, Junru Liu, Ryo Kamoi, Xiaoxin Lu, Caiming Xiong, Jieyu Zhao, Dragomir Radev, Kathleen McKeown, Rui Zhang

Viaarxiv icon

Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems

Oct 23, 2023
Yixin Wan, Jieyu Zhao, Aman Chadha, Nanyun Peng, Kai-Wei Chang

Figure 1 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Figure 2 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Figure 3 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Figure 4 for Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona Biases in Dialogue Systems
Viaarxiv icon

Equal Long-term Benefit Rate: Adapting Static Fairness Notions to Sequential Decision Making

Sep 07, 2023
Yuancheng Xu, Chenghao Deng, Yanchao Sun, Ruijie Zheng, Xiyao Wang, Jieyu Zhao, Furong Huang

Figure 1 for Equal Long-term Benefit Rate: Adapting Static Fairness Notions to Sequential Decision Making
Figure 2 for Equal Long-term Benefit Rate: Adapting Static Fairness Notions to Sequential Decision Making
Figure 3 for Equal Long-term Benefit Rate: Adapting Static Fairness Notions to Sequential Decision Making
Figure 4 for Equal Long-term Benefit Rate: Adapting Static Fairness Notions to Sequential Decision Making
Viaarxiv icon