Picture for Shaohan Huang

Shaohan Huang

GanLM: Encoder-Decoder Pre-training with an Auxiliary Discriminator

Add code
Dec 20, 2022
Viaarxiv icon

TorchScale: Transformers at Scale

Add code
Nov 23, 2022
Viaarxiv icon

Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning

Add code
Oct 26, 2022
Figure 1 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Figure 2 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Figure 3 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Figure 4 for Beyond English-Centric Bitexts for Better Multilingual Language Representation Learning
Viaarxiv icon

Foundation Transformers

Add code
Oct 19, 2022
Figure 1 for Foundation Transformers
Figure 2 for Foundation Transformers
Figure 3 for Foundation Transformers
Figure 4 for Foundation Transformers
Viaarxiv icon

CROP: Zero-shot Cross-lingual Named Entity Recognition with Multilingual Labeled Sequence Translation

Add code
Oct 13, 2022
Figure 1 for CROP: Zero-shot Cross-lingual Named Entity Recognition with Multilingual Labeled Sequence Translation
Figure 2 for CROP: Zero-shot Cross-lingual Named Entity Recognition with Multilingual Labeled Sequence Translation
Figure 3 for CROP: Zero-shot Cross-lingual Named Entity Recognition with Multilingual Labeled Sequence Translation
Figure 4 for CROP: Zero-shot Cross-lingual Named Entity Recognition with Multilingual Labeled Sequence Translation
Viaarxiv icon

MoEC: Mixture of Expert Clusters

Add code
Jul 19, 2022
Figure 1 for MoEC: Mixture of Expert Clusters
Figure 2 for MoEC: Mixture of Expert Clusters
Figure 3 for MoEC: Mixture of Expert Clusters
Figure 4 for MoEC: Mixture of Expert Clusters
Viaarxiv icon

Language Models are General-Purpose Interfaces

Add code
Jun 13, 2022
Figure 1 for Language Models are General-Purpose Interfaces
Figure 2 for Language Models are General-Purpose Interfaces
Figure 3 for Language Models are General-Purpose Interfaces
Figure 4 for Language Models are General-Purpose Interfaces
Viaarxiv icon

Task-Specific Expert Pruning for Sparse Mixture-of-Experts

Add code
Jun 02, 2022
Figure 1 for Task-Specific Expert Pruning for Sparse Mixture-of-Experts
Figure 2 for Task-Specific Expert Pruning for Sparse Mixture-of-Experts
Figure 3 for Task-Specific Expert Pruning for Sparse Mixture-of-Experts
Figure 4 for Task-Specific Expert Pruning for Sparse Mixture-of-Experts
Viaarxiv icon

THE-X: Privacy-Preserving Transformer Inference with Homomorphic Encryption

Add code
Jun 02, 2022
Viaarxiv icon

On the Representation Collapse of Sparse Mixture of Experts

Add code
Apr 20, 2022
Figure 1 for On the Representation Collapse of Sparse Mixture of Experts
Figure 2 for On the Representation Collapse of Sparse Mixture of Experts
Figure 3 for On the Representation Collapse of Sparse Mixture of Experts
Figure 4 for On the Representation Collapse of Sparse Mixture of Experts
Viaarxiv icon