Picture for Ji Lin

Ji Lin

Tiny Machine Learning: Progress and Futures

Add code
Mar 29, 2024
Figure 1 for Tiny Machine Learning: Progress and Futures
Figure 2 for Tiny Machine Learning: Progress and Futures
Figure 3 for Tiny Machine Learning: Progress and Futures
Figure 4 for Tiny Machine Learning: Progress and Futures
Viaarxiv icon

VILA: On Pre-training for Visual Language Models

Add code
Dec 14, 2023
Figure 1 for VILA: On Pre-training for Visual Language Models
Figure 2 for VILA: On Pre-training for Visual Language Models
Figure 3 for VILA: On Pre-training for Visual Language Models
Figure 4 for VILA: On Pre-training for Visual Language Models
Viaarxiv icon

PockEngine: Sparse and Efficient Fine-tuning in a Pocket

Add code
Oct 26, 2023
Figure 1 for PockEngine: Sparse and Efficient Fine-tuning in a Pocket
Figure 2 for PockEngine: Sparse and Efficient Fine-tuning in a Pocket
Figure 3 for PockEngine: Sparse and Efficient Fine-tuning in a Pocket
Figure 4 for PockEngine: Sparse and Efficient Fine-tuning in a Pocket
Viaarxiv icon

AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration

Add code
Jun 01, 2023
Figure 1 for AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Figure 2 for AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Figure 3 for AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Figure 4 for AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Viaarxiv icon

Offsite-Tuning: Transfer Learning without Full Model

Add code
Feb 09, 2023
Figure 1 for Offsite-Tuning: Transfer Learning without Full Model
Figure 2 for Offsite-Tuning: Transfer Learning without Full Model
Figure 3 for Offsite-Tuning: Transfer Learning without Full Model
Figure 4 for Offsite-Tuning: Transfer Learning without Full Model
Viaarxiv icon

SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Add code
Nov 28, 2022
Figure 1 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Figure 2 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Figure 3 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Figure 4 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Viaarxiv icon

Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models

Add code
Nov 15, 2022
Figure 1 for Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
Figure 2 for Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
Figure 3 for Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
Figure 4 for Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
Viaarxiv icon

On-Device Training Under 256KB Memory

Add code
Jul 14, 2022
Figure 1 for On-Device Training Under 256KB Memory
Figure 2 for On-Device Training Under 256KB Memory
Figure 3 for On-Device Training Under 256KB Memory
Figure 4 for On-Device Training Under 256KB Memory
Viaarxiv icon

Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications

Apr 25, 2022
Figure 1 for Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Figure 2 for Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Figure 3 for Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Figure 4 for Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Viaarxiv icon

MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning

Oct 28, 2021
Figure 1 for MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Figure 2 for MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Figure 3 for MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Figure 4 for MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
Viaarxiv icon