Alert button
Picture for Lifan Yuan

Lifan Yuan

Alert button

Advancing LLM Reasoning Generalists with Preference Trees

Add code
Bookmark button
Alert button
Apr 02, 2024
Lifan Yuan, Ganqu Cui, Hanbin Wang, Ning Ding, Xingyao Wang, Jia Deng, Boji Shan, Huimin Chen, Ruobing Xie, Yankai Lin, Zhenghao Liu, Bowen Zhou, Hao Peng, Zhiyuan Liu, Maosong Sun

Viaarxiv icon

Controllable Preference Optimization: Toward Controllable Multi-Objective Alignment

Add code
Bookmark button
Alert button
Feb 29, 2024
Yiju Guo, Ganqu Cui, Lifan Yuan, Ning Ding, Jiexin Wang, Huimin Chen, Bowen Sun, Ruobing Xie, Jie Zhou, Yankai Lin, Zhiyuan Liu, Maosong Sun

Viaarxiv icon

Executable Code Actions Elicit Better LLM Agents

Add code
Bookmark button
Alert button
Feb 01, 2024
Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji

Viaarxiv icon

Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown

Add code
Bookmark button
Alert button
Nov 16, 2023
Genglin Liu, Xingyao Wang, Lifan Yuan, Yangyi Chen, Hao Peng

Figure 1 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Figure 2 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Figure 3 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Figure 4 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Viaarxiv icon

UltraFeedback: Boosting Language Models with High-quality Feedback

Add code
Bookmark button
Alert button
Oct 02, 2023
Ganqu Cui, Lifan Yuan, Ning Ding, Guanming Yao, Wei Zhu, Yuan Ni, Guotong Xie, Zhiyuan Liu, Maosong Sun

Figure 1 for UltraFeedback: Boosting Language Models with High-quality Feedback
Figure 2 for UltraFeedback: Boosting Language Models with High-quality Feedback
Figure 3 for UltraFeedback: Boosting Language Models with High-quality Feedback
Figure 4 for UltraFeedback: Boosting Language Models with High-quality Feedback
Viaarxiv icon

CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets

Add code
Bookmark button
Alert button
Sep 29, 2023
Lifan Yuan, Yangyi Chen, Xingyao Wang, Yi R. Fung, Hao Peng, Heng Ji

Figure 1 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Figure 2 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Figure 3 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Figure 4 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Viaarxiv icon

MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback

Add code
Bookmark button
Alert button
Sep 19, 2023
Xingyao Wang, Zihan Wang, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng, Heng Ji

Figure 1 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Figure 2 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Figure 3 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Figure 4 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Viaarxiv icon

Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations

Add code
Bookmark button
Alert button
Jun 07, 2023
Lifan Yuan, Yangyi Chen, Ganqu Cui, Hongcheng Gao, Fangyuan Zou, Xingyi Cheng, Heng Ji, Zhiyuan Liu, Maosong Sun

Figure 1 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Figure 2 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Figure 3 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Figure 4 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Viaarxiv icon

From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework

Add code
Bookmark button
Alert button
May 29, 2023
Yangyi Chen, Hongcheng Gao, Ganqu Cui, Lifan Yuan, Dehan Kong, Hanlu Wu, Ning Shi, Bo Yuan, Longtao Huang, Hui Xue, Zhiyuan Liu, Maosong Sun, Heng Ji

Figure 1 for From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework
Figure 2 for From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework
Figure 3 for From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework
Figure 4 for From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework
Viaarxiv icon

A Close Look into the Calibration of Pre-trained Language Models

Add code
Bookmark button
Alert button
Oct 31, 2022
Yangyi Chen, Lifan Yuan, Ganqu Cui, Zhiyuan Liu, Heng Ji

Figure 1 for A Close Look into the Calibration of Pre-trained Language Models
Figure 2 for A Close Look into the Calibration of Pre-trained Language Models
Figure 3 for A Close Look into the Calibration of Pre-trained Language Models
Figure 4 for A Close Look into the Calibration of Pre-trained Language Models
Viaarxiv icon