Alert button
Picture for Aaron Chan

Aaron Chan

Alert button

Copilot Evaluation Harness: Evaluating LLM-Guided Software Programming

Add code
Bookmark button
Alert button
Feb 22, 2024
Anisha Agarwal, Aaron Chan, Shubham Chandel, Jinu Jang, Shaun Miller, Roshanak Zilouchian Moghaddam, Yevhen Mohylevskyy, Neel Sundaresan, Michele Tufano

Viaarxiv icon

Tailoring Self-Rationalizers with Multi-Reward Distillation

Add code
Bookmark button
Alert button
Nov 06, 2023
Sahana Ramnath, Brihi Joshi, Skyler Hallinan, Ximing Lu, Liunian Harold Li, Aaron Chan, Jack Hessel, Yejin Choi, Xiang Ren

Viaarxiv icon

Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models

Add code
Bookmark button
Alert button
Oct 07, 2023
Song Jiang, Zahra Shakeri, Aaron Chan, Maziar Sanjabi, Hamed Firooz, Yinglong Xia, Bugra Akyildiz, Yizhou Sun, Jinchao Li, Qifan Wang, Asli Celikyilmaz

Figure 1 for Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models
Figure 2 for Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models
Figure 3 for Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models
Figure 4 for Resprompt: Residual Connection Prompting Advances Multi-Step Reasoning in Large Language Models
Viaarxiv icon

Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales

Add code
Bookmark button
Alert button
May 11, 2023
Brihi Joshi, Ziyi Liu, Sahana Ramnath, Aaron Chan, Zhewei Tong, Shaoliang Nie, Qifan Wang, Yejin Choi, Xiang Ren

Figure 1 for Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales
Figure 2 for Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales
Figure 3 for Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales
Figure 4 for Are Machine Rationales (Not) Useful to Humans? Measuring and Improving Human Utility of Free-Text Rationales
Viaarxiv icon

KNIFE: Knowledge Distillation with Free-Text Rationales

Add code
Bookmark button
Alert button
Dec 19, 2022
Aaron Chan, Zhiyuan Zeng, Wyatt Lake, Brihi Joshi, Hanjie Chen, Xiang Ren

Figure 1 for KNIFE: Knowledge Distillation with Free-Text Rationales
Figure 2 for KNIFE: Knowledge Distillation with Free-Text Rationales
Figure 3 for KNIFE: Knowledge Distillation with Free-Text Rationales
Figure 4 for KNIFE: Knowledge Distillation with Free-Text Rationales
Viaarxiv icon

PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales

Add code
Bookmark button
Alert button
Nov 03, 2022
Peifeng Wang, Aaron Chan, Filip Ilievski, Muhao Chen, Xiang Ren

Figure 1 for PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales
Figure 2 for PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales
Figure 3 for PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales
Figure 4 for PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales
Viaarxiv icon

XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models

Add code
Bookmark button
Alert button
Oct 30, 2022
Dong-Ho Lee, Akshen Kadakia, Brihi Joshi, Aaron Chan, Ziyi Liu, Kiran Narahari, Takashi Shibuya, Ryosuke Mitani, Toshiyuki Sekiya, Jay Pujara, Xiang Ren

Figure 1 for XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models
Figure 2 for XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models
Figure 3 for XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models
Figure 4 for XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models
Viaarxiv icon

FRAME: Evaluating Simulatability Metrics for Free-Text Rationales

Add code
Bookmark button
Alert button
Jul 02, 2022
Aaron Chan, Shaoliang Nie, Liang Tan, Xiaochang Peng, Hamed Firooz, Maziar Sanjabi, Xiang Ren

Figure 1 for FRAME: Evaluating Simulatability Metrics for Free-Text Rationales
Figure 2 for FRAME: Evaluating Simulatability Metrics for Free-Text Rationales
Figure 3 for FRAME: Evaluating Simulatability Metrics for Free-Text Rationales
Figure 4 for FRAME: Evaluating Simulatability Metrics for Free-Text Rationales
Viaarxiv icon

ER-TEST: Evaluating Explanation Regularization Methods for NLP Models

Add code
Bookmark button
Alert button
May 25, 2022
Brihi Joshi, Aaron Chan, Ziyi Liu, Shaoliang Nie, Maziar Sanjabi, Hamed Firooz, Xiang Ren

Figure 1 for ER-TEST: Evaluating Explanation Regularization Methods for NLP Models
Figure 2 for ER-TEST: Evaluating Explanation Regularization Methods for NLP Models
Figure 3 for ER-TEST: Evaluating Explanation Regularization Methods for NLP Models
Figure 4 for ER-TEST: Evaluating Explanation Regularization Methods for NLP Models
Viaarxiv icon

UniREx: A Unified Learning Framework for Language Model Rationale Extraction

Add code
Bookmark button
Alert button
Dec 16, 2021
Aaron Chan, Maziar Sanjabi, Lambert Mathias, Liang Tan, Shaoliang Nie, Xiaochang Peng, Xiang Ren, Hamed Firooz

Figure 1 for UniREx: A Unified Learning Framework for Language Model Rationale Extraction
Figure 2 for UniREx: A Unified Learning Framework for Language Model Rationale Extraction
Figure 3 for UniREx: A Unified Learning Framework for Language Model Rationale Extraction
Figure 4 for UniREx: A Unified Learning Framework for Language Model Rationale Extraction
Viaarxiv icon