Alert button
Picture for Yangyi Chen

Yangyi Chen

Alert button

Executable Code Actions Elicit Better LLM Agents

Add code
Bookmark button
Alert button
Feb 01, 2024
Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji

Viaarxiv icon

ViStruct: Visual Structural Knowledge Extraction via Curriculum Guided Code-Vision Representation

Add code
Bookmark button
Alert button
Nov 22, 2023
Yangyi Chen, Xingyao Wang, Manling Li, Derek Hoiem, Heng Ji

Figure 1 for ViStruct: Visual Structural Knowledge Extraction via Curriculum Guided Code-Vision Representation
Figure 2 for ViStruct: Visual Structural Knowledge Extraction via Curriculum Guided Code-Vision Representation
Figure 3 for ViStruct: Visual Structural Knowledge Extraction via Curriculum Guided Code-Vision Representation
Figure 4 for ViStruct: Visual Structural Knowledge Extraction via Curriculum Guided Code-Vision Representation
Viaarxiv icon

DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback

Add code
Bookmark button
Alert button
Nov 16, 2023
Yangyi Chen, Karan Sikka, Michael Cogswell, Heng Ji, Ajay Divakaran

Figure 1 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Figure 2 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Figure 3 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Figure 4 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Viaarxiv icon

Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown

Add code
Bookmark button
Alert button
Nov 16, 2023
Genglin Liu, Xingyao Wang, Lifan Yuan, Yangyi Chen, Hao Peng

Figure 1 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Figure 2 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Figure 3 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Figure 4 for Prudent Silence or Foolish Babble? Examining Large Language Models' Responses to the Unknown
Viaarxiv icon

R-Tuning: Teaching Large Language Models to Refuse Unknown Questions

Add code
Bookmark button
Alert button
Nov 16, 2023
Hanning Zhang, Shizhe Diao, Yong Lin, Yi R. Fung, Qing Lian, Xingyao Wang, Yangyi Chen, Heng Ji, Tong Zhang

Figure 1 for R-Tuning: Teaching Large Language Models to Refuse Unknown Questions
Figure 2 for R-Tuning: Teaching Large Language Models to Refuse Unknown Questions
Figure 3 for R-Tuning: Teaching Large Language Models to Refuse Unknown Questions
Figure 4 for R-Tuning: Teaching Large Language Models to Refuse Unknown Questions
Viaarxiv icon

CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets

Add code
Bookmark button
Alert button
Sep 29, 2023
Lifan Yuan, Yangyi Chen, Xingyao Wang, Yi R. Fung, Hao Peng, Heng Ji

Figure 1 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Figure 2 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Figure 3 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Figure 4 for CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets
Viaarxiv icon

MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback

Add code
Bookmark button
Alert button
Sep 19, 2023
Xingyao Wang, Zihan Wang, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng, Heng Ji

Figure 1 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Figure 2 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Figure 3 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Figure 4 for MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback
Viaarxiv icon

Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models

Add code
Bookmark button
Alert button
Sep 08, 2023
Yangyi Chen, Karan Sikka, Michael Cogswell, Heng Ji, Ajay Divakaran

Figure 1 for Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models
Figure 2 for Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models
Figure 3 for Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models
Figure 4 for Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models
Viaarxiv icon

Making Pre-trained Language Models both Task-solvers and Self-calibrators

Add code
Bookmark button
Alert button
Jul 21, 2023
Yangyi Chen, Xingyao Wang, Heng Ji

Viaarxiv icon

Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations

Add code
Bookmark button
Alert button
Jun 07, 2023
Lifan Yuan, Yangyi Chen, Ganqu Cui, Hongcheng Gao, Fangyuan Zou, Xingyi Cheng, Heng Ji, Zhiyuan Liu, Maosong Sun

Figure 1 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Figure 2 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Figure 3 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Figure 4 for Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Viaarxiv icon