Alert button
Picture for Michal Lukasik

Michal Lukasik

Alert button

Metric-aware LLM inference

Mar 07, 2024
Michal Lukasik, Harikrishna Narasimhan, Aditya Krishna Menon, Felix Yu, Sanjiv Kumar

Figure 1 for Metric-aware LLM inference
Figure 2 for Metric-aware LLM inference
Figure 3 for Metric-aware LLM inference
Figure 4 for Metric-aware LLM inference
Viaarxiv icon

It's an Alignment, Not a Trade-off: Revisiting Bias and Variance in Deep Models

Oct 13, 2023
Lin Chen, Michal Lukasik, Wittawat Jitkrittum, Chong You, Sanjiv Kumar

Viaarxiv icon

What do larger image classifiers memorise?

Oct 09, 2023
Michal Lukasik, Vaishnavh Nagarajan, Ankit Singh Rawat, Aditya Krishna Menon, Sanjiv Kumar

Figure 1 for What do larger image classifiers memorise?
Figure 2 for What do larger image classifiers memorise?
Figure 3 for What do larger image classifiers memorise?
Figure 4 for What do larger image classifiers memorise?
Viaarxiv icon

ResMem: Learn what you can and memorize the rest

Feb 03, 2023
Zitong Yang, Michal Lukasik, Vaishnavh Nagarajan, Zonglin Li, Ankit Singh Rawat, Manzil Zaheer, Aditya Krishna Menon, Sanjiv Kumar

Figure 1 for ResMem: Learn what you can and memorize the rest
Figure 2 for ResMem: Learn what you can and memorize the rest
Figure 3 for ResMem: Learn what you can and memorize the rest
Figure 4 for ResMem: Learn what you can and memorize the rest
Viaarxiv icon

Large Language Models with Controllable Working Memory

Nov 09, 2022
Daliang Li, Ankit Singh Rawat, Manzil Zaheer, Xin Wang, Michal Lukasik, Andreas Veit, Felix Yu, Sanjiv Kumar

Figure 1 for Large Language Models with Controllable Working Memory
Figure 2 for Large Language Models with Controllable Working Memory
Figure 3 for Large Language Models with Controllable Working Memory
Figure 4 for Large Language Models with Controllable Working Memory
Viaarxiv icon

Preserving In-Context Learning ability in Large Language Model Fine-tuning

Nov 01, 2022
Yihan Wang, Si Si, Daliang Li, Michal Lukasik, Felix Yu, Cho-Jui Hsieh, Inderjit S Dhillon, Sanjiv Kumar

Figure 1 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Figure 2 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Figure 3 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Figure 4 for Preserving In-Context Learning ability in Large Language Model Fine-tuning
Viaarxiv icon

Robust Distillation for Worst-class Performance

Jun 13, 2022
Serena Wang, Harikrishna Narasimhan, Yichen Zhou, Sara Hooker, Michal Lukasik, Aditya Krishna Menon

Figure 1 for Robust Distillation for Worst-class Performance
Figure 2 for Robust Distillation for Worst-class Performance
Figure 3 for Robust Distillation for Worst-class Performance
Figure 4 for Robust Distillation for Worst-class Performance
Viaarxiv icon

HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation

Oct 28, 2021
Wittawat Jitkrittum, Michal Lukasik, Ananda Theertha Suresh, Felix Yu, Gang Wang

Figure 1 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Figure 2 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Figure 3 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Figure 4 for HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party Computation
Viaarxiv icon

Leveraging redundancy in attention with Reuse Transformers

Oct 13, 2021
Srinadh Bhojanapalli, Ayan Chakrabarti, Andreas Veit, Michal Lukasik, Himanshu Jain, Frederick Liu, Yin-Wen Chang, Sanjiv Kumar

Figure 1 for Leveraging redundancy in attention with Reuse Transformers
Figure 2 for Leveraging redundancy in attention with Reuse Transformers
Figure 3 for Leveraging redundancy in attention with Reuse Transformers
Figure 4 for Leveraging redundancy in attention with Reuse Transformers
Viaarxiv icon

Teacher's pet: understanding and mitigating biases in distillation

Jul 08, 2021
Michal Lukasik, Srinadh Bhojanapalli, Aditya Krishna Menon, Sanjiv Kumar

Figure 1 for Teacher's pet: understanding and mitigating biases in distillation
Figure 2 for Teacher's pet: understanding and mitigating biases in distillation
Figure 3 for Teacher's pet: understanding and mitigating biases in distillation
Figure 4 for Teacher's pet: understanding and mitigating biases in distillation
Viaarxiv icon