Picture for Qun Liu

Qun Liu

LightMBERT: A Simple Yet Effective Method for Multilingual BERT Distillation

Add code
Mar 11, 2021
Figure 1 for LightMBERT: A Simple Yet Effective Method for Multilingual BERT Distillation
Figure 2 for LightMBERT: A Simple Yet Effective Method for Multilingual BERT Distillation
Figure 3 for LightMBERT: A Simple Yet Effective Method for Multilingual BERT Distillation
Figure 4 for LightMBERT: A Simple Yet Effective Method for Multilingual BERT Distillation
Viaarxiv icon

Training Multilingual Pre-trained Language Model with Byte-level Subwords

Add code
Jan 23, 2021
Figure 1 for Training Multilingual Pre-trained Language Model with Byte-level Subwords
Figure 2 for Training Multilingual Pre-trained Language Model with Byte-level Subwords
Figure 3 for Training Multilingual Pre-trained Language Model with Byte-level Subwords
Figure 4 for Training Multilingual Pre-trained Language Model with Byte-level Subwords
Viaarxiv icon

Revisiting Robust Neural Machine Translation: A Transformer Case Study

Add code
Dec 31, 2020
Figure 1 for Revisiting Robust Neural Machine Translation: A Transformer Case Study
Figure 2 for Revisiting Robust Neural Machine Translation: A Transformer Case Study
Figure 3 for Revisiting Robust Neural Machine Translation: A Transformer Case Study
Figure 4 for Revisiting Robust Neural Machine Translation: A Transformer Case Study
Viaarxiv icon

BinaryBERT: Pushing the Limit of BERT Quantization

Add code
Dec 31, 2020
Figure 1 for BinaryBERT: Pushing the Limit of BERT Quantization
Figure 2 for BinaryBERT: Pushing the Limit of BERT Quantization
Figure 3 for BinaryBERT: Pushing the Limit of BERT Quantization
Figure 4 for BinaryBERT: Pushing the Limit of BERT Quantization
Viaarxiv icon

Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuning

Add code
Dec 31, 2020
Figure 1 for Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuning
Figure 2 for Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuning
Figure 3 for Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuning
Figure 4 for Better Robustness by More Coverage: Adversarial Training with Mixup Augmentation for Robust Fine-tuning
Viaarxiv icon

HopRetriever: Retrieve Hops over Wikipedia to Answer Complex Questions

Add code
Dec 31, 2020
Figure 1 for HopRetriever: Retrieve Hops over Wikipedia to Answer Complex Questions
Figure 2 for HopRetriever: Retrieve Hops over Wikipedia to Answer Complex Questions
Figure 3 for HopRetriever: Retrieve Hops over Wikipedia to Answer Complex Questions
Figure 4 for HopRetriever: Retrieve Hops over Wikipedia to Answer Complex Questions
Viaarxiv icon

ALP-KD: Attention-Based Layer Projection for Knowledge Distillation

Add code
Dec 27, 2020
Figure 1 for ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Figure 2 for ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Figure 3 for ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Figure 4 for ALP-KD: Attention-Based Layer Projection for Knowledge Distillation
Viaarxiv icon

Improving Task-Agnostic BERT Distillation with Layer Mapping Search

Add code
Dec 11, 2020
Figure 1 for Improving Task-Agnostic BERT Distillation with Layer Mapping Search
Figure 2 for Improving Task-Agnostic BERT Distillation with Layer Mapping Search
Figure 3 for Improving Task-Agnostic BERT Distillation with Layer Mapping Search
Figure 4 for Improving Task-Agnostic BERT Distillation with Layer Mapping Search
Viaarxiv icon

Document Graph for Neural Machine Translation

Add code
Dec 08, 2020
Figure 1 for Document Graph for Neural Machine Translation
Figure 2 for Document Graph for Neural Machine Translation
Figure 3 for Document Graph for Neural Machine Translation
Figure 4 for Document Graph for Neural Machine Translation
Viaarxiv icon

PPKE: Knowledge Representation Learning by Path-based Pre-training

Add code
Dec 07, 2020
Figure 1 for PPKE: Knowledge Representation Learning by Path-based Pre-training
Figure 2 for PPKE: Knowledge Representation Learning by Path-based Pre-training
Figure 3 for PPKE: Knowledge Representation Learning by Path-based Pre-training
Figure 4 for PPKE: Knowledge Representation Learning by Path-based Pre-training
Viaarxiv icon