Picture for Zhengyan Zhang

Zhengyan Zhang

Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parameters

Add code
Jun 11, 2024
Figure 1 for Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parameters
Figure 2 for Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parameters
Figure 3 for Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parameters
Figure 4 for Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parameters
Viaarxiv icon

Robust and Scalable Model Editing for Large Language Models

Add code
Mar 26, 2024
Figure 1 for Robust and Scalable Model Editing for Large Language Models
Figure 2 for Robust and Scalable Model Editing for Large Language Models
Figure 3 for Robust and Scalable Model Editing for Large Language Models
Figure 4 for Robust and Scalable Model Editing for Large Language Models
Viaarxiv icon

ProSparse: Introducing and Enhancing Intrinsic Activation Sparsity within Large Language Models

Add code
Feb 27, 2024
Viaarxiv icon

InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory

Add code
Feb 07, 2024
Viaarxiv icon

ReLU$^2$ Wins: Discovering Efficient Activation Functions for Sparse LLMs

Add code
Feb 06, 2024
Viaarxiv icon

Variator: Accelerating Pre-trained Models with Plug-and-Play Compression Modules

Add code
Oct 24, 2023
Figure 1 for Variator: Accelerating Pre-trained Models with Plug-and-Play Compression Modules
Figure 2 for Variator: Accelerating Pre-trained Models with Plug-and-Play Compression Modules
Figure 3 for Variator: Accelerating Pre-trained Models with Plug-and-Play Compression Modules
Figure 4 for Variator: Accelerating Pre-trained Models with Plug-and-Play Compression Modules
Viaarxiv icon

CPET: Effective Parameter-Efficient Tuning for Compressed Large Language Models

Add code
Jul 15, 2023
Figure 1 for CPET: Effective Parameter-Efficient Tuning for Compressed Large Language Models
Figure 2 for CPET: Effective Parameter-Efficient Tuning for Compressed Large Language Models
Figure 3 for CPET: Effective Parameter-Efficient Tuning for Compressed Large Language Models
Figure 4 for CPET: Effective Parameter-Efficient Tuning for Compressed Large Language Models
Viaarxiv icon

Plug-and-Play Knowledge Injection for Pre-trained Language Models

Add code
May 28, 2023
Figure 1 for Plug-and-Play Knowledge Injection for Pre-trained Language Models
Figure 2 for Plug-and-Play Knowledge Injection for Pre-trained Language Models
Figure 3 for Plug-and-Play Knowledge Injection for Pre-trained Language Models
Figure 4 for Plug-and-Play Knowledge Injection for Pre-trained Language Models
Viaarxiv icon

Plug-and-Play Document Modules for Pre-trained Models

Add code
May 28, 2023
Figure 1 for Plug-and-Play Document Modules for Pre-trained Models
Figure 2 for Plug-and-Play Document Modules for Pre-trained Models
Figure 3 for Plug-and-Play Document Modules for Pre-trained Models
Figure 4 for Plug-and-Play Document Modules for Pre-trained Models
Viaarxiv icon

Emergent Modularity in Pre-trained Transformers

Add code
May 28, 2023
Figure 1 for Emergent Modularity in Pre-trained Transformers
Figure 2 for Emergent Modularity in Pre-trained Transformers
Figure 3 for Emergent Modularity in Pre-trained Transformers
Figure 4 for Emergent Modularity in Pre-trained Transformers
Viaarxiv icon