Picture for Jungwook Choi

Jungwook Choi

Enhancing Generalization in Data-free Quantization via Mixup-class Prompting

Add code
Jul 29, 2025
Viaarxiv icon

Saliency-Aware Quantized Imitation Learning for Efficient Robotic Control

Add code
May 21, 2025
Viaarxiv icon

LoL-PIM: Long-Context LLM Decoding with Scalable DRAM-PIM System

Add code
Dec 28, 2024
Figure 1 for LoL-PIM: Long-Context LLM Decoding with Scalable DRAM-PIM System
Figure 2 for LoL-PIM: Long-Context LLM Decoding with Scalable DRAM-PIM System
Figure 3 for LoL-PIM: Long-Context LLM Decoding with Scalable DRAM-PIM System
Figure 4 for LoL-PIM: Long-Context LLM Decoding with Scalable DRAM-PIM System
Viaarxiv icon

RILQ: Rank-Insensitive LoRA-based Quantization Error Compensation for Boosting 2-bit Large Language Model Accuracy

Add code
Dec 02, 2024
Viaarxiv icon

Quantization-Aware Imitation-Learning for Resource-Efficient Robotic Control

Add code
Dec 02, 2024
Figure 1 for Quantization-Aware Imitation-Learning for Resource-Efficient Robotic Control
Figure 2 for Quantization-Aware Imitation-Learning for Resource-Efficient Robotic Control
Figure 3 for Quantization-Aware Imitation-Learning for Resource-Efficient Robotic Control
Figure 4 for Quantization-Aware Imitation-Learning for Resource-Efficient Robotic Control
Viaarxiv icon

AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference

Add code
Nov 15, 2024
Figure 1 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 2 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 3 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 4 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Viaarxiv icon

InfiniPot: Infinite Context Processing on Memory-Constrained LLMs

Add code
Oct 02, 2024
Figure 1 for InfiniPot: Infinite Context Processing on Memory-Constrained LLMs
Figure 2 for InfiniPot: Infinite Context Processing on Memory-Constrained LLMs
Figure 3 for InfiniPot: Infinite Context Processing on Memory-Constrained LLMs
Figure 4 for InfiniPot: Infinite Context Processing on Memory-Constrained LLMs
Viaarxiv icon

Selectively Dilated Convolution for Accuracy-Preserving Sparse Pillar-based Embedded 3D Object Detection

Add code
Aug 25, 2024
Figure 1 for Selectively Dilated Convolution for Accuracy-Preserving Sparse Pillar-based Embedded 3D Object Detection
Figure 2 for Selectively Dilated Convolution for Accuracy-Preserving Sparse Pillar-based Embedded 3D Object Detection
Figure 3 for Selectively Dilated Convolution for Accuracy-Preserving Sparse Pillar-based Embedded 3D Object Detection
Figure 4 for Selectively Dilated Convolution for Accuracy-Preserving Sparse Pillar-based Embedded 3D Object Detection
Viaarxiv icon

Improving Conversational Abilities of Quantized Large Language Models via Direct Preference Alignment

Add code
Jul 03, 2024
Viaarxiv icon

Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization

Add code
Nov 09, 2023
Figure 1 for Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization
Figure 2 for Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization
Figure 3 for Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization
Figure 4 for Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization
Viaarxiv icon