Picture for Weizhu Chen

Weizhu Chen

Mixing and Shifting: Exploiting Global and Local Dependencies in Vision MLPs

Add code
Feb 14, 2022
Viaarxiv icon

Reasoning Like Program Executors

Add code
Jan 27, 2022
Figure 1 for Reasoning Like Program Executors
Figure 2 for Reasoning Like Program Executors
Figure 3 for Reasoning Like Program Executors
Figure 4 for Reasoning Like Program Executors
Viaarxiv icon

CodeRetriever: Unimodal and Bimodal Contrastive Learning

Add code
Jan 26, 2022
Figure 1 for CodeRetriever: Unimodal and Bimodal Contrastive Learning
Figure 2 for CodeRetriever: Unimodal and Bimodal Contrastive Learning
Figure 3 for CodeRetriever: Unimodal and Bimodal Contrastive Learning
Figure 4 for CodeRetriever: Unimodal and Bimodal Contrastive Learning
Viaarxiv icon

Contextual Bandit Applications in Customer Support Bot

Add code
Dec 06, 2021
Figure 1 for Contextual Bandit Applications in Customer Support Bot
Figure 2 for Contextual Bandit Applications in Customer Support Bot
Figure 3 for Contextual Bandit Applications in Customer Support Bot
Figure 4 for Contextual Bandit Applications in Customer Support Bot
Viaarxiv icon

DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing

Add code
Nov 18, 2021
Figure 1 for DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing
Figure 2 for DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing
Figure 3 for DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing
Figure 4 for DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing
Viaarxiv icon

DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models

Add code
Oct 30, 2021
Figure 1 for DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models
Figure 2 for DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models
Figure 3 for DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models
Figure 4 for DSEE: Dually Sparsity-embedded Efficient Tuning of Pre-trained Language Models
Viaarxiv icon

Adversarial Retriever-Ranker for dense text retrieval

Add code
Oct 29, 2021
Figure 1 for Adversarial Retriever-Ranker for dense text retrieval
Figure 2 for Adversarial Retriever-Ranker for dense text retrieval
Figure 3 for Adversarial Retriever-Ranker for dense text retrieval
Figure 4 for Adversarial Retriever-Ranker for dense text retrieval
Viaarxiv icon

A Good Prompt Is Worth Millions of Parameters? Low-resource Prompt-based Learning for Vision-Language Models

Add code
Oct 16, 2021
Figure 1 for A Good Prompt Is Worth Millions of Parameters? Low-resource Prompt-based Learning for Vision-Language Models
Figure 2 for A Good Prompt Is Worth Millions of Parameters? Low-resource Prompt-based Learning for Vision-Language Models
Figure 3 for A Good Prompt Is Worth Millions of Parameters? Low-resource Prompt-based Learning for Vision-Language Models
Figure 4 for A Good Prompt Is Worth Millions of Parameters? Low-resource Prompt-based Learning for Vision-Language Models
Viaarxiv icon

XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge

Add code
Sep 26, 2021
Figure 1 for XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge
Figure 2 for XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge
Figure 3 for XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge
Figure 4 for XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge
Viaarxiv icon

ARCH: Efficient Adversarial Regularized Training with Caching

Add code
Sep 15, 2021
Figure 1 for ARCH: Efficient Adversarial Regularized Training with Caching
Figure 2 for ARCH: Efficient Adversarial Regularized Training with Caching
Figure 3 for ARCH: Efficient Adversarial Regularized Training with Caching
Figure 4 for ARCH: Efficient Adversarial Regularized Training with Caching
Viaarxiv icon