Alert button
Picture for Wonyong Sung

Wonyong Sung

Alert button

Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling

Add code
Bookmark button
Alert button
Oct 07, 2021
Kyuhong Shim, Iksoo Choi, Wonyong Sung, Jungwook Choi

Figure 1 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Figure 2 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Figure 3 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Figure 4 for Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling
Viaarxiv icon

Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks

Add code
Bookmark button
Alert button
Sep 30, 2020
Yoonho Boo, Sungho Shin, Jungwook Choi, Wonyong Sung

Figure 1 for Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks
Figure 2 for Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks
Figure 3 for Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks
Figure 4 for Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks
Viaarxiv icon

S-SGD: Symmetrical Stochastic Gradient Descent with Weight Noise Injection for Reaching Flat Minima

Add code
Bookmark button
Alert button
Sep 05, 2020
Wonyong Sung, Iksoo Choi, Jinhwan Park, Seokhyun Choi, Sungho Shin

Figure 1 for S-SGD: Symmetrical Stochastic Gradient Descent with Weight Noise Injection for Reaching Flat Minima
Figure 2 for S-SGD: Symmetrical Stochastic Gradient Descent with Weight Noise Injection for Reaching Flat Minima
Figure 3 for S-SGD: Symmetrical Stochastic Gradient Descent with Weight Noise Injection for Reaching Flat Minima
Figure 4 for S-SGD: Symmetrical Stochastic Gradient Descent with Weight Noise Injection for Reaching Flat Minima
Viaarxiv icon

Quantized Neural Networks: Characterization and Holistic Optimization

Add code
Bookmark button
Alert button
May 31, 2020
Yoonho Boo, Sungho Shin, Wonyong Sung

Figure 1 for Quantized Neural Networks: Characterization and Holistic Optimization
Figure 2 for Quantized Neural Networks: Characterization and Holistic Optimization
Figure 3 for Quantized Neural Networks: Characterization and Holistic Optimization
Figure 4 for Quantized Neural Networks: Characterization and Holistic Optimization
Viaarxiv icon

SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks

Add code
Bookmark button
Alert button
Feb 02, 2020
Sungho Shin, Yoonho Boo, Wonyong Sung

Figure 1 for SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks
Figure 2 for SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks
Figure 3 for SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks
Figure 4 for SQWA: Stochastic Quantized Weight Averaging for Improving the Generalization Capability of Low-Precision Deep Neural Networks
Viaarxiv icon

Empirical Analysis of Knowledge Distillation Technique for Optimization of Quantized Deep Neural Networks

Add code
Bookmark button
Alert button
Oct 05, 2019
Sungho Shin, Yoonho Boo, Wonyong Sung

Figure 1 for Empirical Analysis of Knowledge Distillation Technique for Optimization of Quantized Deep Neural Networks
Figure 2 for Empirical Analysis of Knowledge Distillation Technique for Optimization of Quantized Deep Neural Networks
Figure 3 for Empirical Analysis of Knowledge Distillation Technique for Optimization of Quantized Deep Neural Networks
Figure 4 for Empirical Analysis of Knowledge Distillation Technique for Optimization of Quantized Deep Neural Networks
Viaarxiv icon

Single Stream Parallelization of Recurrent Neural Networks for Low Power and Fast Inference

Add code
Bookmark button
Alert button
Mar 30, 2018
Wonyong Sung, Jinhwan Park

Figure 1 for Single Stream Parallelization of Recurrent Neural Networks for Low Power and Fast Inference
Figure 2 for Single Stream Parallelization of Recurrent Neural Networks for Low Power and Fast Inference
Figure 3 for Single Stream Parallelization of Recurrent Neural Networks for Low Power and Fast Inference
Figure 4 for Single Stream Parallelization of Recurrent Neural Networks for Low Power and Fast Inference
Viaarxiv icon

Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations

Add code
Bookmark button
Alert button
Jul 01, 2017
Yoonho Boo, Wonyong Sung

Figure 1 for Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations
Figure 2 for Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations
Figure 3 for Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations
Figure 4 for Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations
Viaarxiv icon

Generative Knowledge Transfer for Neural Language Models

Add code
Bookmark button
Alert button
Feb 28, 2017
Sungho Shin, Kyuyeon Hwang, Wonyong Sung

Figure 1 for Generative Knowledge Transfer for Neural Language Models
Figure 2 for Generative Knowledge Transfer for Neural Language Models
Figure 3 for Generative Knowledge Transfer for Neural Language Models
Figure 4 for Generative Knowledge Transfer for Neural Language Models
Viaarxiv icon