Alert button
Picture for Sihwa Lee

Sihwa Lee

Alert button

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Add code
Bookmark button
Alert button
Aug 13, 2023
Minsoo Kim, Sihwa Lee, Janghwan Lee, Sukjin Hong, Du-Seong Chang, Wonyong Sung, Jungwook Choi

Figure 1 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 2 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 3 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 4 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Viaarxiv icon

Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders

Add code
Bookmark button
Alert button
Nov 20, 2022
Minsoo Kim, Sihwa Lee, Sukjin Hong, Du-Seong Chang, Jungwook Choi

Figure 1 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 2 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 3 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 4 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Viaarxiv icon

NN-LUT: Neural Approximation of Non-Linear Operations for Efficient Transformer Inference

Add code
Bookmark button
Alert button
Dec 03, 2021
Joonsang Yu, Junki Park, Seongmin Park, Minsoo Kim, Sihwa Lee, Dong Hyun Lee, Jungwook Choi

Figure 1 for NN-LUT: Neural Approximation of Non-Linear Operations for Efficient Transformer Inference
Figure 2 for NN-LUT: Neural Approximation of Non-Linear Operations for Efficient Transformer Inference
Figure 3 for NN-LUT: Neural Approximation of Non-Linear Operations for Efficient Transformer Inference
Figure 4 for NN-LUT: Neural Approximation of Non-Linear Operations for Efficient Transformer Inference
Viaarxiv icon