Picture for Xiuying Wei

Xiuying Wei

Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis

Add code
Jul 13, 2024
Viaarxiv icon

Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers

Add code
Jun 24, 2024
Figure 1 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Figure 2 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Figure 3 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Figure 4 for Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Viaarxiv icon

Selective Focus: Investigating Semantics Sensitivity in Post-training Quantization for Lane Detection

Add code
May 10, 2024
Viaarxiv icon

Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in Minutes

Add code
May 09, 2024
Figure 1 for Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in Minutes
Figure 2 for Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in Minutes
Figure 3 for Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in Minutes
Figure 4 for Fast and Controllable Post-training Sparsity: Learning Optimal Sparsity Allocation with Global Constraint in Minutes
Viaarxiv icon

QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models

Add code
Oct 12, 2023
Figure 1 for QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models
Figure 2 for QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models
Figure 3 for QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models
Figure 4 for QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models
Viaarxiv icon

Lossy and Lossless Post-training Model Size Compression

Add code
Aug 08, 2023
Figure 1 for Lossy and Lossless  Post-training Model Size Compression
Figure 2 for Lossy and Lossless  Post-training Model Size Compression
Figure 3 for Lossy and Lossless  Post-training Model Size Compression
Figure 4 for Lossy and Lossless  Post-training Model Size Compression
Viaarxiv icon

Outlier Suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling

Add code
Apr 18, 2023
Figure 1 for Outlier Suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling
Figure 2 for Outlier Suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling
Figure 3 for Outlier Suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling
Figure 4 for Outlier Suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling
Viaarxiv icon

Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models

Add code
Sep 27, 2022
Figure 1 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Figure 2 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Figure 3 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Figure 4 for Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Viaarxiv icon

QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization

Add code
Mar 11, 2022
Figure 1 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Figure 2 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Figure 3 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Figure 4 for QDrop: Randomly Dropping Quantization for Extremely Low-bit Post-Training Quantization
Viaarxiv icon