Picture for Young Jin Kim

Young Jin Kim

Division of Cardiovascular Medicine, Radcliffe Department of Medicine, University of Oxford, Department of Radiology, Severance Hospital, South Korea

FineQuant: Unlocking Efficiency with Fine-Grained Weight-Only Quantization for LLMs

Add code
Aug 16, 2023
Viaarxiv icon

How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation

Add code
Feb 18, 2023
Figure 1 for How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation
Figure 2 for How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation
Figure 3 for How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation
Figure 4 for How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation
Viaarxiv icon

Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production

Add code
Nov 18, 2022
Figure 1 for Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production
Figure 2 for Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production
Figure 3 for Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production
Figure 4 for Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production
Viaarxiv icon

AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers

Add code
Oct 14, 2022
Figure 1 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Figure 2 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Figure 3 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Figure 4 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Viaarxiv icon

Fast Vocabulary Projection Method via Clustering for Multilingual Machine Translation on GPU

Add code
Aug 14, 2022
Figure 1 for Fast Vocabulary Projection Method via Clustering for Multilingual Machine Translation on GPU
Figure 2 for Fast Vocabulary Projection Method via Clustering for Multilingual Machine Translation on GPU
Figure 3 for Fast Vocabulary Projection Method via Clustering for Multilingual Machine Translation on GPU
Figure 4 for Fast Vocabulary Projection Method via Clustering for Multilingual Machine Translation on GPU
Viaarxiv icon

Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers

Add code
May 28, 2022
Figure 1 for Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers
Figure 2 for Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers
Figure 3 for Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers
Figure 4 for Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers
Viaarxiv icon

Taming Sparsely Activated Transformer with Stochastic Experts

Add code
Oct 12, 2021
Figure 1 for Taming Sparsely Activated Transformer with Stochastic Experts
Figure 2 for Taming Sparsely Activated Transformer with Stochastic Experts
Figure 3 for Taming Sparsely Activated Transformer with Stochastic Experts
Figure 4 for Taming Sparsely Activated Transformer with Stochastic Experts
Viaarxiv icon

Scalable and Efficient MoE Training for Multitask Multilingual Models

Add code
Sep 22, 2021
Figure 1 for Scalable and Efficient MoE Training for Multitask Multilingual Models
Figure 2 for Scalable and Efficient MoE Training for Multitask Multilingual Models
Figure 3 for Scalable and Efficient MoE Training for Multitask Multilingual Models
Figure 4 for Scalable and Efficient MoE Training for Multitask Multilingual Models
Viaarxiv icon

FastFormers: Highly Efficient Transformer Models for Natural Language Understanding

Add code
Oct 26, 2020
Figure 1 for FastFormers: Highly Efficient Transformer Models for Natural Language Understanding
Figure 2 for FastFormers: Highly Efficient Transformer Models for Natural Language Understanding
Figure 3 for FastFormers: Highly Efficient Transformer Models for Natural Language Understanding
Figure 4 for FastFormers: Highly Efficient Transformer Models for Natural Language Understanding
Viaarxiv icon

Automated Quality Control in Image Segmentation: Application to the UK Biobank Cardiac MR Imaging Study

Add code
Jan 27, 2019
Figure 1 for Automated Quality Control in Image Segmentation: Application to the UK Biobank Cardiac MR Imaging Study
Figure 2 for Automated Quality Control in Image Segmentation: Application to the UK Biobank Cardiac MR Imaging Study
Figure 3 for Automated Quality Control in Image Segmentation: Application to the UK Biobank Cardiac MR Imaging Study
Figure 4 for Automated Quality Control in Image Segmentation: Application to the UK Biobank Cardiac MR Imaging Study
Viaarxiv icon