Alert button
Picture for Weizhu Chen

Weizhu Chen

Alert button

Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization

Add code
Bookmark button
Alert button
Jun 08, 2021
Chen Liang, Simiao Zuo, Minshuo Chen, Haoming Jiang, Xiaodong Liu, Pengcheng He, Tuo Zhao, Weizhu Chen

Figure 1 for Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization
Figure 2 for Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization
Figure 3 for Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization
Figure 4 for Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization
Viaarxiv icon

HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization

Add code
Bookmark button
Alert button
May 31, 2021
Jiaao Chen, Dinghan Shen, Weizhu Chen, Diyi Yang

Figure 1 for HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization
Figure 2 for HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization
Figure 3 for HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization
Figure 4 for HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization
Viaarxiv icon

Memory-Efficient Differentiable Transformer Architecture Search

Add code
Bookmark button
Alert button
May 31, 2021
Yuekai Zhao, Li Dong, Yelong Shen, Zhihua Zhang, Furu Wei, Weizhu Chen

Figure 1 for Memory-Efficient Differentiable Transformer Architecture Search
Figure 2 for Memory-Efficient Differentiable Transformer Architecture Search
Figure 3 for Memory-Efficient Differentiable Transformer Architecture Search
Figure 4 for Memory-Efficient Differentiable Transformer Architecture Search
Viaarxiv icon

Poolingformer: Long Document Modeling with Pooling Attention

Add code
Bookmark button
Alert button
May 10, 2021
Hang Zhang, Yeyun Gong, Yelong Shen, Weisheng Li, Jiancheng Lv, Nan Duan, Weizhu Chen

Figure 1 for Poolingformer: Long Document Modeling with Pooling Attention
Figure 2 for Poolingformer: Long Document Modeling with Pooling Attention
Figure 3 for Poolingformer: Long Document Modeling with Pooling Attention
Figure 4 for Poolingformer: Long Document Modeling with Pooling Attention
Viaarxiv icon

A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation

Add code
Bookmark button
Alert button
Apr 18, 2021
Tianyu Liu, Yizhe Zhang, Chris Brockett, Yi Mao, Zhifang Sui, Weizhu Chen, Bill Dolan

Figure 1 for A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation
Figure 2 for A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation
Figure 3 for A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation
Figure 4 for A Token-level Reference-free Hallucination Detection Benchmark for Free-form Text Generation
Viaarxiv icon

Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach

Add code
Bookmark button
Alert button
Apr 11, 2021
Simiao Zuo, Chen Liang, Haoming Jiang, Xiaodong Liu, Pengcheng He, Jianfeng Gao, Weizhu Chen, Tuo Zhao

Figure 1 for Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach
Figure 2 for Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach
Figure 3 for Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach
Figure 4 for Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach
Viaarxiv icon

Finetuning Pretrained Transformers into RNNs

Add code
Bookmark button
Alert button
Mar 24, 2021
Jungo Kasai, Hao Peng, Yizhe Zhang, Dani Yogatama, Gabriel Ilharco, Nikolaos Pappas, Yi Mao, Weizhu Chen, Noah A. Smith

Figure 1 for Finetuning Pretrained Transformers into RNNs
Figure 2 for Finetuning Pretrained Transformers into RNNs
Figure 3 for Finetuning Pretrained Transformers into RNNs
Figure 4 for Finetuning Pretrained Transformers into RNNs
Viaarxiv icon

Token-wise Curriculum Learning for Neural Machine Translation

Add code
Bookmark button
Alert button
Mar 20, 2021
Chen Liang, Haoming Jiang, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Tuo Zhao

Figure 1 for Token-wise Curriculum Learning for Neural Machine Translation
Figure 2 for Token-wise Curriculum Learning for Neural Machine Translation
Figure 3 for Token-wise Curriculum Learning for Neural Machine Translation
Figure 4 for Token-wise Curriculum Learning for Neural Machine Translation
Viaarxiv icon

What Makes Good In-Context Examples for GPT-$3$?

Add code
Bookmark button
Alert button
Jan 17, 2021
Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, Weizhu Chen

Figure 1 for What Makes Good In-Context Examples for GPT-$3$?
Figure 2 for What Makes Good In-Context Examples for GPT-$3$?
Figure 3 for What Makes Good In-Context Examples for GPT-$3$?
Figure 4 for What Makes Good In-Context Examples for GPT-$3$?
Viaarxiv icon