Alert button
Picture for Jungwook Choi

Jungwook Choi

Alert button

Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization

Nov 09, 2023
Jangwhan Lee, Minsoo Kim, Seungcheol Baek, Seok Joong Hwang, Wonyong Sung, Jungwook Choi

Viaarxiv icon

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Aug 13, 2023
Minsoo Kim, Sihwa Lee, Janghwan Lee, Sukjin Hong, Du-Seong Chang, Wonyong Sung, Jungwook Choi

Figure 1 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 2 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 3 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 4 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Viaarxiv icon

PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices

May 15, 2023
Minjae Lee, Hyungmin Kim, Seongmin Park, Minyong Yoon, Janghwan Lee, Junwon Choi, Mingu Kang, Jungwook Choi

Figure 1 for PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices
Figure 2 for PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices
Figure 3 for PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices
Figure 4 for PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices
Viaarxiv icon

Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers

Feb 23, 2023
Minsoo Kim, Kyuhong Shim, Seongmin Park, Wonyong Sung, Jungwook Choi

Figure 1 for Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers
Figure 2 for Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers
Figure 3 for Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers
Figure 4 for Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers
Viaarxiv icon

Exploring Attention Map Reuse for Efficient Transformer Neural Networks

Jan 29, 2023
Kyuhong Shim, Jungwook Choi, Wonyong Sung

Figure 1 for Exploring Attention Map Reuse for Efficient Transformer Neural Networks
Figure 2 for Exploring Attention Map Reuse for Efficient Transformer Neural Networks
Figure 3 for Exploring Attention Map Reuse for Efficient Transformer Neural Networks
Figure 4 for Exploring Attention Map Reuse for Efficient Transformer Neural Networks
Viaarxiv icon

Automatic Network Adaptation for Ultra-Low Uniform-Precision Quantization

Jan 04, 2023
Seongmin Park, Beomseok Kwon, Jieun Lim, Kyuyoung Sim, Tae-Ho Kim, Jungwook Choi

Figure 1 for Automatic Network Adaptation for Ultra-Low Uniform-Precision Quantization
Figure 2 for Automatic Network Adaptation for Ultra-Low Uniform-Precision Quantization
Figure 3 for Automatic Network Adaptation for Ultra-Low Uniform-Precision Quantization
Figure 4 for Automatic Network Adaptation for Ultra-Low Uniform-Precision Quantization
Viaarxiv icon

Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders

Nov 20, 2022
Minsoo Kim, Sihwa Lee, Sukjin Hong, Du-Seong Chang, Jungwook Choi

Figure 1 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 2 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 3 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 4 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Viaarxiv icon

Learning from distinctive candidates to optimize reduced-precision convolution program on tensor cores

Feb 24, 2022
Junkyeong Choi, Hyucksung Kwon, Woongkyu Lee, Jungwook Choi, Jieun Lim

Figure 1 for Learning from distinctive candidates to optimize reduced-precision convolution program on tensor cores
Figure 2 for Learning from distinctive candidates to optimize reduced-precision convolution program on tensor cores
Figure 3 for Learning from distinctive candidates to optimize reduced-precision convolution program on tensor cores
Figure 4 for Learning from distinctive candidates to optimize reduced-precision convolution program on tensor cores
Viaarxiv icon