Picture for Shuohuan Wang

Shuohuan Wang

DHA: Learning Decoupled-Head Attention from Transformer Checkpoints via Adaptive Heads Fusion

Add code
Jun 03, 2024
Viaarxiv icon

HFT: Half Fine-Tuning for Large Language Models

Add code
Apr 29, 2024
Figure 1 for HFT: Half Fine-Tuning for Large Language Models
Figure 2 for HFT: Half Fine-Tuning for Large Language Models
Figure 3 for HFT: Half Fine-Tuning for Large Language Models
Figure 4 for HFT: Half Fine-Tuning for Large Language Models
Viaarxiv icon

Dual Modalities of Text: Visual and Textual Generative Pre-training

Add code
Apr 17, 2024
Viaarxiv icon

On Training Data Influence of GPT Models

Add code
Apr 11, 2024
Figure 1 for On Training Data Influence of GPT Models
Figure 2 for On Training Data Influence of GPT Models
Figure 3 for On Training Data Influence of GPT Models
Figure 4 for On Training Data Influence of GPT Models
Viaarxiv icon

Tool-Augmented Reward Modeling

Add code
Oct 02, 2023
Figure 1 for Tool-Augmented Reward Modeling
Figure 2 for Tool-Augmented Reward Modeling
Figure 3 for Tool-Augmented Reward Modeling
Figure 4 for Tool-Augmented Reward Modeling
Viaarxiv icon

ERNIE-Music: Text-to-Waveform Music Generation with Diffusion Models

Add code
Feb 09, 2023
Figure 1 for ERNIE-Music: Text-to-Waveform Music Generation with Diffusion Models
Figure 2 for ERNIE-Music: Text-to-Waveform Music Generation with Diffusion Models
Figure 3 for ERNIE-Music: Text-to-Waveform Music Generation with Diffusion Models
Figure 4 for ERNIE-Music: Text-to-Waveform Music Generation with Diffusion Models
Viaarxiv icon

ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages

Add code
Dec 13, 2022
Figure 1 for ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages
Figure 2 for ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages
Figure 3 for ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages
Figure 4 for ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for Programming Languages
Viaarxiv icon

X-PuDu at SemEval-2022 Task 6: Multilingual Learning for English and Arabic Sarcasm Detection

Add code
Nov 30, 2022
Figure 1 for X-PuDu at SemEval-2022 Task 6: Multilingual Learning for English and Arabic Sarcasm Detection
Figure 2 for X-PuDu at SemEval-2022 Task 6: Multilingual Learning for English and Arabic Sarcasm Detection
Figure 3 for X-PuDu at SemEval-2022 Task 6: Multilingual Learning for English and Arabic Sarcasm Detection
Figure 4 for X-PuDu at SemEval-2022 Task 6: Multilingual Learning for English and Arabic Sarcasm Detection
Viaarxiv icon

X-PuDu at SemEval-2022 Task 7: A Replaced Token Detection Task Pre-trained Model with Pattern-aware Ensembling for Identifying Plausible Clarifications

Add code
Nov 27, 2022
Figure 1 for X-PuDu at SemEval-2022 Task 7: A Replaced Token Detection Task Pre-trained Model with Pattern-aware Ensembling for Identifying Plausible Clarifications
Figure 2 for X-PuDu at SemEval-2022 Task 7: A Replaced Token Detection Task Pre-trained Model with Pattern-aware Ensembling for Identifying Plausible Clarifications
Figure 3 for X-PuDu at SemEval-2022 Task 7: A Replaced Token Detection Task Pre-trained Model with Pattern-aware Ensembling for Identifying Plausible Clarifications
Figure 4 for X-PuDu at SemEval-2022 Task 7: A Replaced Token Detection Task Pre-trained Model with Pattern-aware Ensembling for Identifying Plausible Clarifications
Viaarxiv icon

ERNIE-UniX2: A Unified Cross-lingual Cross-modal Framework for Understanding and Generation

Add code
Nov 09, 2022
Figure 1 for ERNIE-UniX2: A Unified Cross-lingual Cross-modal Framework for Understanding and Generation
Figure 2 for ERNIE-UniX2: A Unified Cross-lingual Cross-modal Framework for Understanding and Generation
Figure 3 for ERNIE-UniX2: A Unified Cross-lingual Cross-modal Framework for Understanding and Generation
Figure 4 for ERNIE-UniX2: A Unified Cross-lingual Cross-modal Framework for Understanding and Generation
Viaarxiv icon