Alert button
Picture for Haonan Li

Haonan Li

Alert button

Against The Achilles' Heel: A Survey on Red Teaming for Generative Models

Add code
Bookmark button
Alert button
Mar 31, 2024
Lizhi Lin, Honglin Mu, Zenan Zhai, Minghan Wang, Yuxia Wang, Renxi Wang, Junjie Gao, Yixuan Zhang, Wanxiang Che, Timothy Baldwin, Xudong Han, Haonan Li

Viaarxiv icon

EXAMS-V: A Multi-Discipline Multilingual Multimodal Exam Benchmark for Evaluating Vision Language Models

Add code
Bookmark button
Alert button
Mar 15, 2024
Rocktim Jyoti Das, Simeon Emilov Hristov, Haonan Li, Dimitar Iliyanov Dimitrov, Ivan Koychev, Preslav Nakov

Figure 1 for EXAMS-V: A Multi-Discipline Multilingual Multimodal Exam Benchmark for Evaluating Vision Language Models
Figure 2 for EXAMS-V: A Multi-Discipline Multilingual Multimodal Exam Benchmark for Evaluating Vision Language Models
Figure 3 for EXAMS-V: A Multi-Discipline Multilingual Multimodal Exam Benchmark for Evaluating Vision Language Models
Figure 4 for EXAMS-V: A Multi-Discipline Multilingual Multimodal Exam Benchmark for Evaluating Vision Language Models
Viaarxiv icon

Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification

Add code
Bookmark button
Alert button
Mar 07, 2024
Ekaterina Fadeeva, Aleksandr Rubashevskii, Artem Shelmanov, Sergey Petrakov, Haonan Li, Hamdy Mubarak, Evgenii Tsymbalov, Gleb Kuzmin, Alexander Panchenko, Timothy Baldwin, Preslav Nakov, Maxim Panov

Figure 1 for Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification
Figure 2 for Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification
Figure 3 for Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification
Figure 4 for Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification
Viaarxiv icon

ArabicMMLU: Assessing Massive Multitask Language Understanding in Arabic

Add code
Bookmark button
Alert button
Feb 20, 2024
Fajri Koto, Haonan Li, Sara Shatnawi, Jad Doughman, Abdelrahman Boda Sadallah, Aisha Alraeesi, Khalid Almubarak, Zaid Alyafeai, Neha Sengupta, Shady Shehata, Nizar Habash, Preslav Nakov, Timothy Baldwin

Viaarxiv icon

A Chinese Dataset for Evaluating the Safeguards in Large Language Models

Add code
Bookmark button
Alert button
Feb 19, 2024
Yuxia Wang, Zenan Zhai, Haonan Li, Xudong Han, Lizhi Lin, Zhenxuan Zhang, Jingru Zhao, Preslav Nakov, Timothy Baldwin

Viaarxiv icon

Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agents

Add code
Bookmark button
Alert button
Feb 18, 2024
Renxi Wang, Haonan Li, Xudong Han, Yixuan Zhang, Timothy Baldwin

Viaarxiv icon

Location Aware Modular Biencoder for Tourism Question Answering

Add code
Bookmark button
Alert button
Jan 04, 2024
Haonan Li, Martin Tomko, Timothy Baldwin

Viaarxiv icon

Understanding the Instruction Mixture for Large Language Model Fine-tuning

Add code
Bookmark button
Alert button
Dec 19, 2023
Renxi Wang, Minghao Wu, Yuxia Wang, Xudong Han, Chiyu Zhang, Haonan Li

Viaarxiv icon

LLM360: Towards Fully Transparent Open-Source LLMs

Add code
Bookmark button
Alert button
Dec 11, 2023
Zhengzhong Liu, Aurick Qiao, Willie Neiswanger, Hongyi Wang, Bowen Tan, Tianhua Tao, Junbo Li, Yuqi Wang, Suqi Sun, Omkar Pangarkar, Richard Fan, Yi Gu, Victor Miller, Yonghao Zhuang, Guowei He, Haonan Li, Fajri Koto, Liping Tang, Nikhil Ranjan, Zhiqiang Shen, Xuguang Ren, Roberto Iriondo, Cun Mu, Zhiting Hu, Mark Schulze, Preslav Nakov, Tim Baldwin, Eric P. Xing

Figure 1 for LLM360: Towards Fully Transparent Open-Source LLMs
Figure 2 for LLM360: Towards Fully Transparent Open-Source LLMs
Figure 3 for LLM360: Towards Fully Transparent Open-Source LLMs
Figure 4 for LLM360: Towards Fully Transparent Open-Source LLMs
Viaarxiv icon

Can Large Language Model Comprehend Ancient Chinese? A Preliminary Test on ACLUE

Add code
Bookmark button
Alert button
Oct 14, 2023
Yixuan Zhang, Haonan Li

Viaarxiv icon