Picture for Davis Liang

Davis Liang

RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training

Add code
Dec 07, 2023
Figure 1 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 2 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 3 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 4 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Viaarxiv icon

Co-training and Co-distillation for Quality Improvement and Compression of Language Models

Add code
Nov 07, 2023
Figure 1 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Figure 2 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Figure 3 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Figure 4 for Co-training and Co-distillation for Quality Improvement and Compression of Language Models
Viaarxiv icon

The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants

Add code
Aug 31, 2023
Figure 1 for The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants
Figure 2 for The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants
Figure 3 for The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants
Figure 4 for The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants
Viaarxiv icon

A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models

Add code
May 26, 2023
Figure 1 for A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models
Figure 2 for A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models
Figure 3 for A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models
Figure 4 for A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models
Viaarxiv icon

XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models

Add code
Jan 25, 2023
Figure 1 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Figure 2 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Figure 3 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Figure 4 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Viaarxiv icon

Adaptable Claim Rewriting with Offline Reinforcement Learning for Effective Misinformation Discovery

Add code
Oct 14, 2022
Figure 1 for Adaptable Claim Rewriting with Offline Reinforcement Learning for Effective Misinformation Discovery
Figure 2 for Adaptable Claim Rewriting with Offline Reinforcement Learning for Effective Misinformation Discovery
Figure 3 for Adaptable Claim Rewriting with Offline Reinforcement Learning for Effective Misinformation Discovery
Figure 4 for Adaptable Claim Rewriting with Offline Reinforcement Learning for Effective Misinformation Discovery
Viaarxiv icon

Attention-guided Generative Models for Extractive Question Answering

Add code
Oct 12, 2021
Figure 1 for Attention-guided Generative Models for Extractive Question Answering
Figure 2 for Attention-guided Generative Models for Extractive Question Answering
Figure 3 for Attention-guided Generative Models for Extractive Question Answering
Figure 4 for Attention-guided Generative Models for Extractive Question Answering
Viaarxiv icon

Multiplicative Position-aware Transformer Models for Language Understanding

Add code
Sep 27, 2021
Figure 1 for Multiplicative Position-aware Transformer Models for Language Understanding
Figure 2 for Multiplicative Position-aware Transformer Models for Language Understanding
Figure 3 for Multiplicative Position-aware Transformer Models for Language Understanding
Figure 4 for Multiplicative Position-aware Transformer Models for Language Understanding
Viaarxiv icon

Decoding and Diversity in Machine Translation

Add code
Nov 26, 2020
Figure 1 for Decoding and Diversity in Machine Translation
Figure 2 for Decoding and Diversity in Machine Translation
Figure 3 for Decoding and Diversity in Machine Translation
Figure 4 for Decoding and Diversity in Machine Translation
Viaarxiv icon

Improve Transformer Models with Better Relative Position Embeddings

Add code
Sep 28, 2020
Figure 1 for Improve Transformer Models with Better Relative Position Embeddings
Figure 2 for Improve Transformer Models with Better Relative Position Embeddings
Figure 3 for Improve Transformer Models with Better Relative Position Embeddings
Figure 4 for Improve Transformer Models with Better Relative Position Embeddings
Viaarxiv icon