Picture for Xingrun Xing

Xingrun Xing

Position-Aware Depth Decay Decoding ($D^3$): Boosting Large Language Model Inference Efficiency

Add code
Mar 11, 2025
Viaarxiv icon

EfficientLLM: Scalable Pruning-Aware Pretraining for Architecture-Agnostic Edge Language Models

Add code
Feb 10, 2025
Viaarxiv icon

Enhancing Generalization via Sharpness-Aware Trajectory Matching for Dataset Condensation

Add code
Feb 03, 2025
Viaarxiv icon

From Pixels to Tokens: Byte-Pair Encoding on Quantized Visual Modalities

Add code
Oct 03, 2024
Figure 1 for From Pixels to Tokens: Byte-Pair Encoding on Quantized Visual Modalities
Figure 2 for From Pixels to Tokens: Byte-Pair Encoding on Quantized Visual Modalities
Figure 3 for From Pixels to Tokens: Byte-Pair Encoding on Quantized Visual Modalities
Figure 4 for From Pixels to Tokens: Byte-Pair Encoding on Quantized Visual Modalities
Viaarxiv icon

OmniGen: Unified Image Generation

Add code
Sep 17, 2024
Figure 1 for OmniGen: Unified Image Generation
Figure 2 for OmniGen: Unified Image Generation
Figure 3 for OmniGen: Unified Image Generation
Figure 4 for OmniGen: Unified Image Generation
Viaarxiv icon

SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking

Add code
Jul 05, 2024
Figure 1 for SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking
Figure 2 for SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking
Figure 3 for SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking
Figure 4 for SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking
Viaarxiv icon

SpikeLM: Towards General Spike-Driven Language Modeling via Elastic Bi-Spiking Mechanisms

Add code
Jun 05, 2024
Viaarxiv icon

BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials

Add code
Dec 14, 2023
Figure 1 for BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials
Figure 2 for BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials
Figure 3 for BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials
Figure 4 for BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials
Viaarxiv icon

LM-Cocktail: Resilient Tuning of Language Models via Model Merging

Add code
Dec 08, 2023
Figure 1 for LM-Cocktail: Resilient Tuning of Language Models via Model Merging
Figure 2 for LM-Cocktail: Resilient Tuning of Language Models via Model Merging
Figure 3 for LM-Cocktail: Resilient Tuning of Language Models via Model Merging
Figure 4 for LM-Cocktail: Resilient Tuning of Language Models via Model Merging
Viaarxiv icon

Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis

Add code
Sep 11, 2023
Figure 1 for Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis
Figure 2 for Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis
Figure 3 for Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis
Figure 4 for Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis
Viaarxiv icon