Picture for Boxing Chen

Boxing Chen

Huawei Noah's Ark Lab

QDyLoRA: Quantized Dynamic Low-Rank Adaptation for Efficient Large Language Model Tuning

Add code
Feb 16, 2024
Figure 1 for QDyLoRA: Quantized Dynamic Low-Rank Adaptation for Efficient Large Language Model Tuning
Figure 2 for QDyLoRA: Quantized Dynamic Low-Rank Adaptation for Efficient Large Language Model Tuning
Figure 3 for QDyLoRA: Quantized Dynamic Low-Rank Adaptation for Efficient Large Language Model Tuning
Figure 4 for QDyLoRA: Quantized Dynamic Low-Rank Adaptation for Efficient Large Language Model Tuning
Viaarxiv icon

On the importance of Data Scale in Pretraining Arabic Language Models

Add code
Jan 15, 2024
Figure 1 for On the importance of Data Scale in Pretraining Arabic Language Models
Figure 2 for On the importance of Data Scale in Pretraining Arabic Language Models
Figure 3 for On the importance of Data Scale in Pretraining Arabic Language Models
Figure 4 for On the importance of Data Scale in Pretraining Arabic Language Models
Viaarxiv icon

NoMIRACL: Knowing When You Don't Know for Robust Multilingual Retrieval-Augmented Generation

Add code
Dec 18, 2023
Figure 1 for NoMIRACL: Knowing When You Don't Know for Robust Multilingual Retrieval-Augmented Generation
Figure 2 for NoMIRACL: Knowing When You Don't Know for Robust Multilingual Retrieval-Augmented Generation
Figure 3 for NoMIRACL: Knowing When You Don't Know for Robust Multilingual Retrieval-Augmented Generation
Figure 4 for NoMIRACL: Knowing When You Don't Know for Robust Multilingual Retrieval-Augmented Generation
Viaarxiv icon

Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models

Add code
Dec 15, 2023
Figure 1 for Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models
Figure 2 for Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models
Figure 3 for Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models
Figure 4 for Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models
Viaarxiv icon

Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference Using Sorted Fine-Tuning (SoFT)

Add code
Sep 16, 2023
Figure 1 for Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference Using Sorted Fine-Tuning (SoFT)
Figure 2 for Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference Using Sorted Fine-Tuning (SoFT)
Figure 3 for Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference Using Sorted Fine-Tuning (SoFT)
Figure 4 for Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference Using Sorted Fine-Tuning (SoFT)
Viaarxiv icon

SortedNet, a Place for Every Network and Every Network in its Place: Towards a Generalized Solution for Training Many-in-One Neural Networks

Add code
Sep 01, 2023
Viaarxiv icon

On the Transferability of Whisper-based Representations for "In-the-Wild" Cross-Task Downstream Speech Applications

Add code
May 23, 2023
Viaarxiv icon

One Adapter for All Programming Languages? Adapter Tuning for Code Search and Summarization

Add code
Mar 28, 2023
Viaarxiv icon

Translate the Beauty in Songs: Jointly Learning to Align Melody and Translate Lyrics

Add code
Mar 28, 2023
Viaarxiv icon

Mathematical Challenges in Deep Learning

Add code
Mar 24, 2023
Figure 1 for Mathematical Challenges in Deep Learning
Figure 2 for Mathematical Challenges in Deep Learning
Figure 3 for Mathematical Challenges in Deep Learning
Figure 4 for Mathematical Challenges in Deep Learning
Viaarxiv icon