Alert button
Picture for Mehdi Rezagholizadeh

Mehdi Rezagholizadeh

Alert button

Pro-KD: Progressive Distillation by Following the Footsteps of the Teacher

Add code
Bookmark button
Alert button
Oct 16, 2021
Mehdi Rezagholizadeh, Aref Jafari, Puneeth Salad, Pranav Sharma, Ali Saheb Pasand, Ali Ghodsi

Figure 1 for Pro-KD: Progressive Distillation by Following the Footsteps of the Teacher
Figure 2 for Pro-KD: Progressive Distillation by Following the Footsteps of the Teacher
Figure 3 for Pro-KD: Progressive Distillation by Following the Footsteps of the Teacher
Figure 4 for Pro-KD: Progressive Distillation by Following the Footsteps of the Teacher
Viaarxiv icon

A Short Study on Compressing Decoder-Based Language Models

Add code
Bookmark button
Alert button
Oct 16, 2021
Tianda Li, Yassir El Mesbahi, Ivan Kobyzev, Ahmad Rashid, Atif Mahmud, Nithin Anchuri, Habib Hajimolahoseini, Yang Liu, Mehdi Rezagholizadeh

Figure 1 for A Short Study on Compressing Decoder-Based Language Models
Figure 2 for A Short Study on Compressing Decoder-Based Language Models
Figure 3 for A Short Study on Compressing Decoder-Based Language Models
Figure 4 for A Short Study on Compressing Decoder-Based Language Models
Viaarxiv icon

Kronecker Decomposition for GPT Compression

Add code
Bookmark button
Alert button
Oct 15, 2021
Ali Edalati, Marzieh Tahaei, Ahmad Rashid, Vahid Partovi Nia, James J. Clark, Mehdi Rezagholizadeh

Figure 1 for Kronecker Decomposition for GPT Compression
Figure 2 for Kronecker Decomposition for GPT Compression
Figure 3 for Kronecker Decomposition for GPT Compression
Figure 4 for Kronecker Decomposition for GPT Compression
Viaarxiv icon

RAIL-KD: RAndom Intermediate Layer Mapping for Knowledge Distillation

Add code
Bookmark button
Alert button
Oct 01, 2021
Md Akmal Haidar, Nithin Anchuri, Mehdi Rezagholizadeh, Abbas Ghaddar, Philippe Langlais, Pascal Poupart

Figure 1 for RAIL-KD: RAndom Intermediate Layer Mapping for Knowledge Distillation
Figure 2 for RAIL-KD: RAndom Intermediate Layer Mapping for Knowledge Distillation
Figure 3 for RAIL-KD: RAndom Intermediate Layer Mapping for Knowledge Distillation
Figure 4 for RAIL-KD: RAndom Intermediate Layer Mapping for Knowledge Distillation
Viaarxiv icon

Knowledge Distillation with Noisy Labels for Natural Language Understanding

Add code
Bookmark button
Alert button
Sep 21, 2021
Shivendra Bhardwaj, Abbas Ghaddar, Ahmad Rashid, Khalil Bibi, Chengyang Li, Ali Ghodsi, Philippe Langlais, Mehdi Rezagholizadeh

Figure 1 for Knowledge Distillation with Noisy Labels for Natural Language Understanding
Figure 2 for Knowledge Distillation with Noisy Labels for Natural Language Understanding
Figure 3 for Knowledge Distillation with Noisy Labels for Natural Language Understanding
Viaarxiv icon

How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding

Add code
Bookmark button
Alert button
Sep 20, 2021
Tianda Li, Ahmad Rashid, Aref Jafari, Pranav Sharma, Ali Ghodsi, Mehdi Rezagholizadeh

Figure 1 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Figure 2 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Figure 3 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Figure 4 for How to Select One Among All? An Extensive Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding
Viaarxiv icon

KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation

Add code
Bookmark button
Alert button
Sep 13, 2021
Marzieh S. Tahaei, Ella Charlaix, Vahid Partovi Nia, Ali Ghodsi, Mehdi Rezagholizadeh

Figure 1 for KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation
Figure 2 for KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation
Figure 3 for KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation
Figure 4 for KroneckerBERT: Learning Kronecker Decomposition for Pre-trained Language Models via Knowledge Distillation
Viaarxiv icon

End-to-End Self-Debiasing Framework for Robust NLU Training

Add code
Bookmark button
Alert button
Sep 05, 2021
Abbas Ghaddar, Philippe Langlais, Mehdi Rezagholizadeh, Ahmad Rashid

Figure 1 for End-to-End Self-Debiasing Framework for Robust NLU Training
Figure 2 for End-to-End Self-Debiasing Framework for Robust NLU Training
Figure 3 for End-to-End Self-Debiasing Framework for Robust NLU Training
Viaarxiv icon

Context-aware Adversarial Training for Name Regularity Bias in Named Entity Recognition

Add code
Bookmark button
Alert button
Jul 24, 2021
Abbas Ghaddar, Philippe Langlais, Ahmad Rashid, Mehdi Rezagholizadeh

Viaarxiv icon