Alert button
Picture for Subhabrata Mukherjee

Subhabrata Mukherjee

Alert button

AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning

Add code
Bookmark button
Alert button
Nov 02, 2022
Yaqing Wang, Sahaj Agarwal, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao

Figure 1 for AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning
Figure 2 for AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning
Figure 3 for AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning
Figure 4 for AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning
Viaarxiv icon

AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers

Add code
Bookmark button
Alert button
Oct 14, 2022
Ganesh Jawahar, Subhabrata Mukherjee, Xiaodong Liu, Young Jin Kim, Muhammad Abdul-Mageed, Laks V. S. Lakshmanan, Ahmed Hassan Awadallah, Sebastien Bubeck, Jianfeng Gao

Figure 1 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Figure 2 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Figure 3 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Figure 4 for AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers
Viaarxiv icon

Small Character Models Match Large Word Models for Autocomplete Under Memory Constraints

Add code
Bookmark button
Alert button
Oct 06, 2022
Ganesh Jawahar, Subhabrata Mukherjee, Debadeepta Dey, Muhammad Abdul-Mageed, Laks V. S. Lakshmanan, Caio Cesar Teodoro Mendes, Gustavo Henrique de Rosa, Shital Shah

Figure 1 for Small Character Models Match Large Word Models for Autocomplete Under Memory Constraints
Figure 2 for Small Character Models Match Large Word Models for Autocomplete Under Memory Constraints
Figure 3 for Small Character Models Match Large Word Models for Autocomplete Under Memory Constraints
Figure 4 for Small Character Models Match Large Word Models for Autocomplete Under Memory Constraints
Viaarxiv icon

ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels

Add code
Bookmark button
Alert button
Aug 24, 2022
Yue Zhao, Guoqing Zheng, Subhabrata Mukherjee, Robert McCann, Ahmed Awadallah

Figure 1 for ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels
Figure 2 for ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels
Figure 3 for ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels
Figure 4 for ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels
Viaarxiv icon

AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models

Add code
Bookmark button
Alert button
May 24, 2022
Yaqing Wang, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao

Figure 1 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Figure 2 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Figure 3 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Figure 4 for AdaMix: Mixture-of-Adapter for Parameter-efficient Tuning of Large Language Models
Viaarxiv icon

Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners

Add code
Bookmark button
Alert button
Apr 16, 2022
Shashank Gupta, Subhabrata Mukherjee, Krishan Subudhi, Eduardo Gonzalez, Damien Jose, Ahmed H. Awadallah, Jianfeng Gao

Figure 1 for Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Figure 2 for Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Figure 3 for Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Figure 4 for Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners
Viaarxiv icon

LiteTransformerSearch: Training-free On-device Search for Efficient Autoregressive Language Models

Add code
Bookmark button
Alert button
Mar 04, 2022
Mojan Javaheripi, Shital Shah, Subhabrata Mukherjee, Tomasz L. Religa, Caio C. T. Mendes, Gustavo H. de Rosa, Sebastien Bubeck, Farinaz Koushanfar, Debadeepta Dey

Figure 1 for LiteTransformerSearch: Training-free On-device Search for Efficient Autoregressive Language Models
Figure 2 for LiteTransformerSearch: Training-free On-device Search for Efficient Autoregressive Language Models
Figure 3 for LiteTransformerSearch: Training-free On-device Search for Efficient Autoregressive Language Models
Figure 4 for LiteTransformerSearch: Training-free On-device Search for Efficient Autoregressive Language Models
Viaarxiv icon

AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models

Add code
Bookmark button
Alert button
Jan 29, 2022
Dongkuan Xu, Subhabrata Mukherjee, Xiaodong Liu, Debadeepta Dey, Wenhui Wang, Xiang Zhang, Ahmed Hassan Awadallah, Jianfeng Gao

Figure 1 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Figure 2 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Figure 3 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Figure 4 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Viaarxiv icon

CLUES: Few-Shot Learning Evaluation in Natural Language Understanding

Add code
Bookmark button
Alert button
Nov 04, 2021
Subhabrata Mukherjee, Xiaodong Liu, Guoqing Zheng, Saghar Hosseini, Hao Cheng, Greg Yang, Christopher Meek, Ahmed Hassan Awadallah, Jianfeng Gao

Figure 1 for CLUES: Few-Shot Learning Evaluation in Natural Language Understanding
Figure 2 for CLUES: Few-Shot Learning Evaluation in Natural Language Understanding
Figure 3 for CLUES: Few-Shot Learning Evaluation in Natural Language Understanding
Figure 4 for CLUES: Few-Shot Learning Evaluation in Natural Language Understanding
Viaarxiv icon

What do Compressed Large Language Models Forget? Robustness Challenges in Model Compression

Add code
Bookmark button
Alert button
Oct 16, 2021
Mengnan Du, Subhabrata Mukherjee, Yu Cheng, Milad Shokouhi, Xia Hu, Ahmed Hassan Awadallah

Figure 1 for What do Compressed Large Language Models Forget? Robustness Challenges in Model Compression
Figure 2 for What do Compressed Large Language Models Forget? Robustness Challenges in Model Compression
Figure 3 for What do Compressed Large Language Models Forget? Robustness Challenges in Model Compression
Figure 4 for What do Compressed Large Language Models Forget? Robustness Challenges in Model Compression
Viaarxiv icon