Picture for Rhui Dih Lee

Rhui Dih Lee

IBM Research

Collaboratively adding new knowledge to an LLM

Add code
Oct 18, 2024
Figure 1 for Collaboratively adding new knowledge to an LLM
Figure 2 for Collaboratively adding new knowledge to an LLM
Figure 3 for Collaboratively adding new knowledge to an LLM
Figure 4 for Collaboratively adding new knowledge to an LLM
Viaarxiv icon

Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts

Add code
Aug 30, 2024
Figure 1 for Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts
Figure 2 for Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts
Figure 3 for Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts
Figure 4 for Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts
Viaarxiv icon

Enhancing Training Efficiency Using Packing with Flash Attention

Add code
Jul 12, 2024
Figure 1 for Enhancing Training Efficiency Using Packing with Flash Attention
Figure 2 for Enhancing Training Efficiency Using Packing with Flash Attention
Figure 3 for Enhancing Training Efficiency Using Packing with Flash Attention
Figure 4 for Enhancing Training Efficiency Using Packing with Flash Attention
Viaarxiv icon

Efficiently Distilling LLMs for Edge Applications

Add code
Apr 01, 2024
Figure 1 for Efficiently Distilling LLMs for Edge Applications
Figure 2 for Efficiently Distilling LLMs for Edge Applications
Figure 3 for Efficiently Distilling LLMs for Edge Applications
Figure 4 for Efficiently Distilling LLMs for Edge Applications
Viaarxiv icon

TOFA: Transfer-Once-for-All

Add code
Mar 27, 2023
Figure 1 for TOFA: Transfer-Once-for-All
Figure 2 for TOFA: Transfer-Once-for-All
Figure 3 for TOFA: Transfer-Once-for-All
Figure 4 for TOFA: Transfer-Once-for-All
Viaarxiv icon