Picture for Zhongzhi Yu

Zhongzhi Yu

Celine

Model Tells You Where to Merge: Adaptive KV Cache Merging for LLMs on Long-Context Tasks

Add code
Jul 11, 2024
Viaarxiv icon

MG-Verilog: Multi-grained Dataset Towards Enhanced LLM-assisted Verilog Generation

Add code
Jul 02, 2024
Viaarxiv icon

EDGE-LLM: Enabling Efficient Large Language Model Adaptation on Edge Devices via Layerwise Unified Compression and Adaptive Layer Tuning and Voting

Add code
Jun 22, 2024
Viaarxiv icon

Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Add code
Jun 22, 2024
Viaarxiv icon

GPT4AIGChip: Towards Next-Generation AI Accelerator Design Automation via Large Language Models

Add code
Sep 19, 2023
Viaarxiv icon

Master-ASR: Achieving Multilingual Scalability and Low-Resource Adaptation in ASR with Modular Learning

Add code
Jun 23, 2023
Viaarxiv icon

NetBooster: Empowering Tiny Deep Learning By Standing on the Shoulders of Deep Giants

Add code
Jun 23, 2023
Viaarxiv icon

Hint-Aug: Drawing Hints from Foundation Vision Transformers Towards Boosted Few-Shot Parameter-Efficient Tuning

Add code
Apr 26, 2023
Viaarxiv icon

Losses Can Be Blessings: Routing Self-Supervised Speech Representations Towards Efficient Multilingual and Multitask Speech Processing

Add code
Nov 02, 2022
Viaarxiv icon

ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design

Add code
Oct 18, 2022
Figure 1 for ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design
Figure 2 for ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design
Figure 3 for ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design
Figure 4 for ViTCoD: Vision Transformer Acceleration via Dedicated Algorithm and Accelerator Co-Design
Viaarxiv icon