Picture for Peiyan Dong

Peiyan Dong

Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation

Add code
May 28, 2025
Viaarxiv icon

Structured Agent Distillation for Large Language Model

Add code
May 20, 2025
Viaarxiv icon

RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation

Add code
Jan 08, 2025
Figure 1 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Figure 2 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Figure 3 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Figure 4 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Viaarxiv icon

Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers

Add code
Jul 25, 2024
Figure 1 for Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers
Figure 2 for Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers
Figure 3 for Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers
Figure 4 for Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers
Viaarxiv icon

EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge

Add code
Feb 16, 2024
Figure 1 for EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge
Figure 2 for EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge
Figure 3 for EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge
Figure 4 for EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge
Viaarxiv icon

Agile-Quant: Activation-Guided Quantization for Faster Inference of LLMs on the Edge

Add code
Dec 09, 2023
Viaarxiv icon

SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices

Add code
Sep 21, 2023
Viaarxiv icon

Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training

Add code
Nov 19, 2022
Figure 1 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 2 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 3 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 4 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Viaarxiv icon

HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers

Add code
Nov 15, 2022
Viaarxiv icon

The Lottery Ticket Hypothesis for Vision Transformers

Add code
Nov 02, 2022
Viaarxiv icon