Alert button
Picture for Cheng-Yu Hsieh

Cheng-Yu Hsieh

Alert button

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

Add code
Bookmark button
Alert button
Oct 08, 2023
Lu Yin, You Wu, Zhenyu Zhang, Cheng-Yu Hsieh, Yaqing Wang, Yiling Jia, Mykola Pechenizkiy, Yi Liang, Zhangyang Wang, Shiwei Liu

Figure 1 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Figure 2 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Figure 3 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Figure 4 for Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Viaarxiv icon

Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models

Add code
Bookmark button
Alert button
Aug 01, 2023
Cheng-Yu Hsieh, Si-An Chen, Chun-Liang Li, Yasuhisa Fujii, Alexander Ratner, Chen-Yu Lee, Ranjay Krishna, Tomas Pfister

Figure 1 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Figure 2 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Figure 3 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Figure 4 for Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models
Viaarxiv icon

SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality

Add code
Bookmark button
Alert button
Jun 26, 2023
Cheng-Yu Hsieh, Jieyu Zhang, Zixian Ma, Aniruddha Kembhavi, Ranjay Krishna

Figure 1 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Figure 2 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Figure 3 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Figure 4 for SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality
Viaarxiv icon

Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes

Add code
Bookmark button
Alert button
May 03, 2023
Cheng-Yu Hsieh, Chun-Liang Li, Chih-Kuan Yeh, Hootan Nakhost, Yasuhisa Fujii, Alexander Ratner, Ranjay Krishna, Chen-Yu Lee, Tomas Pfister

Figure 1 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Figure 2 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Figure 3 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Figure 4 for Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes
Viaarxiv icon

Understanding Programmatic Weak Supervision via Source-aware Influence Function

Add code
Bookmark button
Alert button
May 25, 2022
Jieyu Zhang, Haonan Wang, Cheng-Yu Hsieh, Alexander Ratner

Figure 1 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Figure 2 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Figure 3 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Figure 4 for Understanding Programmatic Weak Supervision via Source-aware Influence Function
Viaarxiv icon

Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data Programming

Add code
Bookmark button
Alert button
Mar 23, 2022
Cheng-Yu Hsieh, Jieyu Zhang, Alexander Ratner

Figure 1 for Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data Programming
Figure 2 for Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data Programming
Figure 3 for Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data Programming
Figure 4 for Nemo: Guiding and Contextualizing Weak Supervision for Interactive Data Programming
Viaarxiv icon

A Survey on Programmatic Weak Supervision

Add code
Bookmark button
Alert button
Feb 14, 2022
Jieyu Zhang, Cheng-Yu Hsieh, Yue Yu, Chao Zhang, Alexander Ratner

Figure 1 for A Survey on Programmatic Weak Supervision
Figure 2 for A Survey on Programmatic Weak Supervision
Viaarxiv icon

Active Refinement for Multi-Label Learning: A Pseudo-Label Approach

Add code
Bookmark button
Alert button
Sep 29, 2021
Cheng-Yu Hsieh, Wei-I Lin, Miao Xu, Gang Niu, Hsuan-Tien Lin, Masashi Sugiyama

Figure 1 for Active Refinement for Multi-Label Learning: A Pseudo-Label Approach
Figure 2 for Active Refinement for Multi-Label Learning: A Pseudo-Label Approach
Figure 3 for Active Refinement for Multi-Label Learning: A Pseudo-Label Approach
Figure 4 for Active Refinement for Multi-Label Learning: A Pseudo-Label Approach
Viaarxiv icon

Evaluations and Methods for Explanation through Robustness Analysis

Add code
Bookmark button
Alert button
May 31, 2020
Cheng-Yu Hsieh, Chih-Kuan Yeh, Xuanqing Liu, Pradeep Ravikumar, Seungyeon Kim, Sanjiv Kumar, Cho-Jui Hsieh

Figure 1 for Evaluations and Methods for Explanation through Robustness Analysis
Figure 2 for Evaluations and Methods for Explanation through Robustness Analysis
Figure 3 for Evaluations and Methods for Explanation through Robustness Analysis
Figure 4 for Evaluations and Methods for Explanation through Robustness Analysis
Viaarxiv icon