Alert button
Picture for Du-Seong Chang

Du-Seong Chang

Alert button

PSYDIAL: Personality-based Synthetic Dialogue Generation using Large Language Models

Add code
Bookmark button
Alert button
Apr 01, 2024
Ji-Eun Han, Jun-Seok Koh, Hyeon-Tae Seo, Du-Seong Chang, Kyung-Ah Sohn

Viaarxiv icon

NASH: A Simple Unified Framework of Structured Pruning for Accelerating Encoder-Decoder Language Models

Add code
Bookmark button
Alert button
Oct 16, 2023
Jongwoo Ko, Seungjoon Park, Yujin Kim, Sumyeong Ahn, Du-Seong Chang, Euijai Ahn, Se-Young Yun

Viaarxiv icon

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Add code
Bookmark button
Alert button
Aug 13, 2023
Minsoo Kim, Sihwa Lee, Janghwan Lee, Sukjin Hong, Du-Seong Chang, Wonyong Sung, Jungwook Choi

Figure 1 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 2 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 3 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Figure 4 for Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Viaarxiv icon

Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective

Add code
Bookmark button
Alert button
Feb 03, 2023
Jongwoo Ko, Seungjoon Park, Minchan Jeong, Sukjin Hong, Euijai Ahn, Du-Seong Chang, Se-Young Yun

Figure 1 for Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective
Figure 2 for Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective
Figure 3 for Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective
Figure 4 for Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective
Viaarxiv icon

Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders

Add code
Bookmark button
Alert button
Nov 20, 2022
Minsoo Kim, Sihwa Lee, Sukjin Hong, Du-Seong Chang, Jungwook Choi

Figure 1 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 2 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 3 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Figure 4 for Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
Viaarxiv icon