Picture for Hyungjun Kim

Hyungjun Kim

QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference

Add code
Feb 15, 2024
Viaarxiv icon

SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks

Add code
Feb 14, 2024
Figure 1 for SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Figure 2 for SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Figure 3 for SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Figure 4 for SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Viaarxiv icon

Squeezing Large-Scale Diffusion Models for Mobile

Add code
Jul 03, 2023
Figure 1 for Squeezing Large-Scale Diffusion Models for Mobile
Figure 2 for Squeezing Large-Scale Diffusion Models for Mobile
Figure 3 for Squeezing Large-Scale Diffusion Models for Mobile
Figure 4 for Squeezing Large-Scale Diffusion Models for Mobile
Viaarxiv icon

OWQ: Lessons learned from activation outliers for weight quantization in large language models

Add code
Jun 13, 2023
Figure 1 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Figure 2 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Figure 3 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Figure 4 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Viaarxiv icon

Temporal Dynamic Quantization for Diffusion Models

Add code
Jun 04, 2023
Figure 1 for Temporal Dynamic Quantization for Diffusion Models
Figure 2 for Temporal Dynamic Quantization for Diffusion Models
Figure 3 for Temporal Dynamic Quantization for Diffusion Models
Figure 4 for Temporal Dynamic Quantization for Diffusion Models
Viaarxiv icon

INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold

Add code
Apr 18, 2022
Figure 1 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 2 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 3 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 4 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Viaarxiv icon

Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution

Add code
Dec 02, 2020
Figure 1 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 2 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 3 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 4 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Viaarxiv icon

Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression

Add code
Sep 08, 2020
Figure 1 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 2 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 3 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 4 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Viaarxiv icon

BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations

Add code
Feb 16, 2020
Figure 1 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 2 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 3 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 4 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Viaarxiv icon

Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays

Add code
Aug 02, 2019
Figure 1 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Figure 2 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Figure 3 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Figure 4 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Viaarxiv icon