Alert button
Picture for Yuxia Wang

Yuxia Wang

Alert button

Against The Achilles' Heel: A Survey on Red Teaming for Generative Models

Add code
Bookmark button
Alert button
Mar 31, 2024
Lizhi Lin, Honglin Mu, Zenan Zhai, Minghan Wang, Yuxia Wang, Renxi Wang, Junjie Gao, Yixuan Zhang, Wanxiang Che, Timothy Baldwin, Xudong Han, Haonan Li

Viaarxiv icon

A Chinese Dataset for Evaluating the Safeguards in Large Language Models

Add code
Bookmark button
Alert button
Feb 19, 2024
Yuxia Wang, Zenan Zhai, Haonan Li, Xudong Han, Lizhi Lin, Zhenxuan Zhang, Jingru Zhao, Preslav Nakov, Timothy Baldwin

Viaarxiv icon

M4GT-Bench: Evaluation Benchmark for Black-Box Machine-Generated Text Detection

Add code
Bookmark button
Alert button
Feb 17, 2024
Yuxia Wang, Jonibek Mansurov, Petar Ivanov, Jinyan Su, Artem Shelmanov, Akim Tsvigun, Osama Mohanned Afzal, Tarek Mahmoud, Giovanni Puccetti, Thomas Arnold, Alham Fikri Aji, Nizar Habash, Iryna Gurevych, Preslav Nakov

Viaarxiv icon

Factuality of Large Language Models in the Year 2024

Add code
Bookmark button
Alert button
Feb 09, 2024
Yuxia Wang, Minghan Wang, Muhammad Arslan Manzoor, Fei Liu, Georgi Georgiev, Rocktim Jyoti Das, Preslav Nakov

Viaarxiv icon

Understanding the Instruction Mixture for Large Language Model Fine-tuning

Add code
Bookmark button
Alert button
Dec 19, 2023
Renxi Wang, Minghao Wu, Yuxia Wang, Xudong Han, Chiyu Zhang, Haonan Li

Viaarxiv icon

Factcheck-GPT: End-to-End Fine-Grained Document-Level Fact-Checking and Correction of LLM Output

Add code
Bookmark button
Alert button
Nov 16, 2023
Yuxia Wang, Revanth Gangi Reddy, Zain Muhammad Mujahid, Arnav Arora, Aleksandr Rubashevskii, Jiahui Geng, Osama Mohammed Afzal, Liangming Pan, Nadav Borenstein, Aditya Pillai, Isabelle Augenstein, Iryna Gurevych, Preslav Nakov

Viaarxiv icon

A Survey of Language Model Confidence Estimation and Calibration

Add code
Bookmark button
Alert button
Nov 14, 2023
Jiahui Geng, Fengyu Cai, Yuxia Wang, Heinz Koeppl, Preslav Nakov, Iryna Gurevych

Viaarxiv icon

Rethinking STS and NLI in Large Language Models

Add code
Bookmark button
Alert button
Sep 16, 2023
Yuxia Wang, Minghan Wang, Preslav Nakov

Figure 1 for Rethinking STS and NLI in Large Language Models
Figure 2 for Rethinking STS and NLI in Large Language Models
Figure 3 for Rethinking STS and NLI in Large Language Models
Figure 4 for Rethinking STS and NLI in Large Language Models
Viaarxiv icon

Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs

Add code
Bookmark button
Alert button
Sep 04, 2023
Yuxia Wang, Haonan Li, Xudong Han, Preslav Nakov, Timothy Baldwin

Figure 1 for Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs
Figure 2 for Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs
Figure 3 for Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs
Figure 4 for Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs
Viaarxiv icon