Alert button
Picture for Zeyi Liao

Zeyi Liao

Alert button

Introducing v0.5 of the AI Safety Benchmark from MLCommons

Add code
Bookmark button
Alert button
Apr 18, 2024
Bertie Vidgen, Adarsh Agrawal, Ahmed M. Ahmed, Victor Akinwande, Namir Al-Nuaimi, Najla Alfaraj, Elie Alhajjar, Lora Aroyo, Trupti Bavalatti, Borhane Blili-Hamelin, Kurt Bollacker, Rishi Bomassani, Marisa Ferrara Boston, Siméon Campos, Kal Chakra, Canyu Chen, Cody Coleman, Zacharie Delpierre Coudert, Leon Derczynski, Debojyoti Dutta, Ian Eisenberg, James Ezick, Heather Frase, Brian Fuller, Ram Gandikota, Agasthya Gangavarapu, Ananya Gangavarapu, James Gealy, Rajat Ghosh, James Goel, Usman Gohar, Sujata Goswami, Scott A. Hale, Wiebke Hutiri, Joseph Marvin Imperial, Surgan Jandial, Nick Judd, Felix Juefei-Xu, Foutse Khomh, Bhavya Kailkhura, Hannah Rose Kirk, Kevin Klyman, Chris Knotz, Michael Kuchnik, Shachi H. Kumar, Chris Lengerich, Bo Li, Zeyi Liao, Eileen Peters Long, Victor Lu, Yifan Mai, Priyanka Mary Mammen, Kelvin Manyeki, Sean McGregor, Virendra Mehta, Shafee Mohammed, Emanuel Moss, Lama Nachman, Dinesh Jinenhally Naganna, Amin Nikanjam, Besmira Nushi, Luis Oala, Iftach Orr, Alicia Parrish, Cigdem Patlak, William Pietri, Forough Poursabzi-Sangdeh, Eleonora Presani, Fabrizio Puletti, Paul Röttger, Saurav Sahay, Tim Santos, Nino Scherrer, Alice Schoenauer Sebag, Patrick Schramowski, Abolfazl Shahbazi, Vin Sharma, Xudong Shen, Vamsi Sistla, Leonard Tang, Davide Testuggine, Vithursan Thangarasa, Elizabeth Anne Watkins, Rebecca Weiss, Chris Welty, Tyler Wilbers, Adina Williams, Carole-Jean Wu, Poonam Yadav, Xianjun Yang, Yi Zeng, Wenhui Zhang, Fedor Zhdanov, Jiacheng Zhu, Percy Liang, Peter Mattson, Joaquin Vanschoren

Viaarxiv icon

AmpleGCG: Learning a Universal and Transferable Generative Model of Adversarial Suffixes for Jailbreaking Both Open and Closed LLMs

Add code
Bookmark button
Alert button
Apr 11, 2024
Zeyi Liao, Huan Sun

Viaarxiv icon

AttributionBench: How Hard is Automatic Attribution Evaluation?

Add code
Bookmark button
Alert button
Feb 23, 2024
Yifei Li, Xiang Yue, Zeyi Liao, Huan Sun

Viaarxiv icon

A Trembling House of Cards? Mapping Adversarial Attacks against Language Agents

Add code
Bookmark button
Alert button
Feb 15, 2024
Lingbo Mo, Zeyi Liao, Boyuan Zheng, Yu Su, Chaowei Xiao, Huan Sun

Viaarxiv icon

In Search of the Long-Tail: Systematic Generation of Long-Tail Knowledge via Logical Rule Guided Search

Add code
Bookmark button
Alert button
Nov 13, 2023
Huihan Li, Yuting Ning, Zeyi Liao, Siyuan Wang, Xiang Lorraine Li, Ximing Lu, Faeze Brahman, Wenting Zhao, Yejin Choi, Xiang Ren

Viaarxiv icon

ChatCounselor: A Large Language Models for Mental Health Support

Add code
Bookmark button
Alert button
Sep 27, 2023
June M. Liu, Donghao Li, He Cao, Tianhe Ren, Zeyi Liao, Jiamin Wu

Figure 1 for ChatCounselor: A Large Language Models for Mental Health Support
Figure 2 for ChatCounselor: A Large Language Models for Mental Health Support
Figure 3 for ChatCounselor: A Large Language Models for Mental Health Support
Figure 4 for ChatCounselor: A Large Language Models for Mental Health Support
Viaarxiv icon

RobustLR: Evaluating Robustness to Logical Perturbation in Deductive Reasoning

Add code
Bookmark button
Alert button
May 25, 2022
Soumya Sanyal, Zeyi Liao, Xiang Ren

Figure 1 for RobustLR: Evaluating Robustness to Logical Perturbation in Deductive Reasoning
Figure 2 for RobustLR: Evaluating Robustness to Logical Perturbation in Deductive Reasoning
Figure 3 for RobustLR: Evaluating Robustness to Logical Perturbation in Deductive Reasoning
Figure 4 for RobustLR: Evaluating Robustness to Logical Perturbation in Deductive Reasoning
Viaarxiv icon