Alert button
Picture for Srinadh Bhojanapalli

Srinadh Bhojanapalli

Alert button

Efficient Language Model Architectures for Differentially Private Federated Learning

Add code
Bookmark button
Alert button
Mar 12, 2024
Jae Hun Ro, Srinadh Bhojanapalli, Zheng Xu, Yanxiang Zhang, Ananda Theertha Suresh

Figure 1 for Efficient Language Model Architectures for Differentially Private Federated Learning
Figure 2 for Efficient Language Model Architectures for Differentially Private Federated Learning
Figure 3 for Efficient Language Model Architectures for Differentially Private Federated Learning
Figure 4 for Efficient Language Model Architectures for Differentially Private Federated Learning
Viaarxiv icon

HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM Inference

Add code
Bookmark button
Alert button
Feb 14, 2024
Yashas Samaga B L, Varun Yerram, Chong You, Srinadh Bhojanapalli, Sanjiv Kumar, Prateek Jain, Praneeth Netrapalli

Viaarxiv icon

Efficacy of Dual-Encoders for Extreme Multi-Label Classification

Add code
Bookmark button
Alert button
Oct 16, 2023
Nilesh Gupta, Devvrit Khatri, Ankit S Rawat, Srinadh Bhojanapalli, Prateek Jain, Inderjit S Dhillon

Figure 1 for Efficacy of Dual-Encoders for Extreme Multi-Label Classification
Figure 2 for Efficacy of Dual-Encoders for Extreme Multi-Label Classification
Figure 3 for Efficacy of Dual-Encoders for Extreme Multi-Label Classification
Figure 4 for Efficacy of Dual-Encoders for Extreme Multi-Label Classification
Viaarxiv icon

Functional Interpolation for Relative Positions Improves Long Context Transformers

Add code
Bookmark button
Alert button
Oct 06, 2023
Shanda Li, Chong You, Guru Guruganesh, Joshua Ainslie, Santiago Ontanon, Manzil Zaheer, Sumit Sanghai, Yiming Yang, Sanjiv Kumar, Srinadh Bhojanapalli

Viaarxiv icon

Depth Dependence of $μ$P Learning Rates in ReLU MLPs

Add code
Bookmark button
Alert button
May 13, 2023
Samy Jelassi, Boris Hanin, Ziwei Ji, Sashank J. Reddi, Srinadh Bhojanapalli, Sanjiv Kumar

Viaarxiv icon

On student-teacher deviations in distillation: does it pay to disobey?

Add code
Bookmark button
Alert button
Jan 30, 2023
Vaishnavh Nagarajan, Aditya Krishna Menon, Srinadh Bhojanapalli, Hossein Mobahi, Sanjiv Kumar

Figure 1 for On student-teacher deviations in distillation: does it pay to disobey?
Figure 2 for On student-teacher deviations in distillation: does it pay to disobey?
Figure 3 for On student-teacher deviations in distillation: does it pay to disobey?
Figure 4 for On student-teacher deviations in distillation: does it pay to disobey?
Viaarxiv icon

On the Adversarial Robustness of Mixture of Experts

Add code
Bookmark button
Alert button
Oct 19, 2022
Joan Puigcerver, Rodolphe Jenatton, Carlos Riquelme, Pranjal Awasthi, Srinadh Bhojanapalli

Figure 1 for On the Adversarial Robustness of Mixture of Experts
Figure 2 for On the Adversarial Robustness of Mixture of Experts
Figure 3 for On the Adversarial Robustness of Mixture of Experts
Figure 4 for On the Adversarial Robustness of Mixture of Experts
Viaarxiv icon

Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers

Add code
Bookmark button
Alert button
Oct 12, 2022
Zonglin Li, Chong You, Srinadh Bhojanapalli, Daliang Li, Ankit Singh Rawat, Sashank J. Reddi, Ke Ye, Felix Chern, Felix Yu, Ruiqi Guo, Sanjiv Kumar

Figure 1 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Figure 2 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Figure 3 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Figure 4 for Large Models are Parsimonious Learners: Activation Sparsity in Trained Transformers
Viaarxiv icon

Treeformer: Dense Gradient Trees for Efficient Attention Computation

Add code
Bookmark button
Alert button
Aug 18, 2022
Lovish Madaan, Srinadh Bhojanapalli, Himanshu Jain, Prateek Jain

Figure 1 for Treeformer: Dense Gradient Trees for Efficient Attention Computation
Figure 2 for Treeformer: Dense Gradient Trees for Efficient Attention Computation
Figure 3 for Treeformer: Dense Gradient Trees for Efficient Attention Computation
Figure 4 for Treeformer: Dense Gradient Trees for Efficient Attention Computation
Viaarxiv icon

Robust Training of Neural Networks using Scale Invariant Architectures

Add code
Bookmark button
Alert button
Feb 02, 2022
Zhiyuan Li, Srinadh Bhojanapalli, Manzil Zaheer, Sashank J. Reddi, Sanjiv Kumar

Figure 1 for Robust Training of Neural Networks using Scale Invariant Architectures
Figure 2 for Robust Training of Neural Networks using Scale Invariant Architectures
Figure 3 for Robust Training of Neural Networks using Scale Invariant Architectures
Figure 4 for Robust Training of Neural Networks using Scale Invariant Architectures
Viaarxiv icon