Alert button
Picture for Wenlin Yao

Wenlin Yao

Alert button

Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era

Add code
Bookmark button
Alert button
Mar 13, 2024
Xuansheng Wu, Haiyan Zhao, Yaochen Zhu, Yucheng Shi, Fan Yang, Tianming Liu, Xiaoming Zhai, Wenlin Yao, Jundong Li, Mengnan Du, Ninghao Liu

Figure 1 for Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era
Figure 2 for Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era
Figure 3 for Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era
Figure 4 for Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era
Viaarxiv icon

Fact-and-Reflection (FaR) Improves Confidence Calibration of Large Language Models

Add code
Bookmark button
Alert button
Feb 27, 2024
Xinran Zhao, Hongming Zhang, Xiaoman Pan, Wenlin Yao, Dong Yu, Tongshuang Wu, Jianshu Chen

Viaarxiv icon

WebVoyager: Building an End-to-End Web Agent with Large Multimodal Models

Add code
Bookmark button
Alert button
Jan 28, 2024
Hongliang He, Wenlin Yao, Kaixin Ma, Wenhao Yu, Yong Dai, Hongming Zhang, Zhenzhong Lan, Dong Yu

Viaarxiv icon

InFoBench: Evaluating Instruction Following Ability in Large Language Models

Add code
Bookmark button
Alert button
Jan 07, 2024
Yiwei Qin, Kaiqiang Song, Yebowen Hu, Wenlin Yao, Sangwoo Cho, Xiaoyang Wang, Xuansheng Wu, Fei Liu, Pengfei Liu, Dong Yu

Viaarxiv icon

MMC: Advancing Multimodal Chart Understanding with Large-scale Instruction Tuning

Add code
Bookmark button
Alert button
Nov 15, 2023
Fuxiao Liu, Xiaoyang Wang, Wenlin Yao, Jianshu Chen, Kaiqiang Song, Sangwoo Cho, Yaser Yacoob, Dong Yu

Viaarxiv icon

TencentLLMEval: A Hierarchical Evaluation of Real-World Capabilities for Human-Aligned LLMs

Add code
Bookmark button
Alert button
Nov 09, 2023
Shuyi Xie, Wenlin Yao, Yong Dai, Shaobo Wang, Donlin Zhou, Lifeng Jin, Xinhua Feng, Pengzhi Wei, Yujie Lin, Zhichao Hu, Dong Yu, Zhengyou Zhang, Jing Nie, Yuhong Liu

Viaarxiv icon

From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning

Add code
Bookmark button
Alert button
Sep 30, 2023
Xuansheng Wu, Wenlin Yao, Jianshu Chen, Xiaoman Pan, Xiaoyang Wang, Ninghao Liu, Dong Yu

Figure 1 for From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning
Figure 2 for From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning
Figure 3 for From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning
Figure 4 for From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning
Viaarxiv icon

Thrust: Adaptively Propels Large Language Models with External Knowledge

Add code
Bookmark button
Alert button
Jul 19, 2023
Xinran Zhao, Hongming Zhang, Xiaoman Pan, Wenlin Yao, Dong Yu, Jianshu Chen

Figure 1 for Thrust: Adaptively Propels Large Language Models with External Knowledge
Figure 2 for Thrust: Adaptively Propels Large Language Models with External Knowledge
Figure 3 for Thrust: Adaptively Propels Large Language Models with External Knowledge
Figure 4 for Thrust: Adaptively Propels Large Language Models with External Knowledge
Viaarxiv icon

MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning

Add code
Bookmark button
Alert button
Jul 16, 2023
Zhenwen Liang, Dian Yu, Xiaoman Pan, Wenlin Yao, Qingkai Zeng, Xiangliang Zhang, Dong Yu

Figure 1 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Figure 2 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Figure 3 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Figure 4 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Viaarxiv icon

A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation

Add code
Bookmark button
Alert button
Jul 08, 2023
Neeraj Varshney, Wenlin Yao, Hongming Zhang, Jianshu Chen, Dong Yu

Figure 1 for A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation
Figure 2 for A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation
Figure 3 for A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation
Figure 4 for A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation
Viaarxiv icon