Alert button
Picture for Leshem Choshen

Leshem Choshen

Alert button

Fuse to Forget: Bias Reduction and Selective Memorization through Model Fusion

Add code
Bookmark button
Alert button
Nov 13, 2023
Kerem Zaman, Leshem Choshen, Shashank Srivastava

Viaarxiv icon

Efficient Benchmarking (of Language Models)

Add code
Bookmark button
Alert button
Aug 31, 2023
Yotam Perlitz, Elron Bandel, Ariel Gera, Ofir Arviv, Liat Ein-Dor, Eyal Shnarch, Noam Slonim, Michal Shmueli-Scheuer, Leshem Choshen

Figure 1 for Efficient Benchmarking (of Language Models)
Figure 2 for Efficient Benchmarking (of Language Models)
Figure 3 for Efficient Benchmarking (of Language Models)
Figure 4 for Efficient Benchmarking (of Language Models)
Viaarxiv icon

Resolving Interference When Merging Models

Add code
Bookmark button
Alert button
Jun 02, 2023
Prateek Yadav, Derek Tam, Leshem Choshen, Colin Raffel, Mohit Bansal

Figure 1 for Resolving Interference When Merging Models
Figure 2 for Resolving Interference When Merging Models
Figure 3 for Resolving Interference When Merging Models
Figure 4 for Resolving Interference When Merging Models
Viaarxiv icon

MuLER: Detailed and Scalable Reference-based Evaluation

Add code
Bookmark button
Alert button
May 24, 2023
Taelin Karidi, Leshem Choshen, Gal Patel, Omri Abend

Figure 1 for MuLER: Detailed and Scalable Reference-based Evaluation
Figure 2 for MuLER: Detailed and Scalable Reference-based Evaluation
Figure 3 for MuLER: Detailed and Scalable Reference-based Evaluation
Figure 4 for MuLER: Detailed and Scalable Reference-based Evaluation
Viaarxiv icon

Jump to Conclusions: Short-Cutting Transformers With Linear Transformations

Add code
Bookmark button
Alert button
Mar 16, 2023
Alexander Yom Din, Taelin Karidi, Leshem Choshen, Mor Geva

Figure 1 for Jump to Conclusions: Short-Cutting Transformers With Linear Transformations
Figure 2 for Jump to Conclusions: Short-Cutting Transformers With Linear Transformations
Figure 3 for Jump to Conclusions: Short-Cutting Transformers With Linear Transformations
Figure 4 for Jump to Conclusions: Short-Cutting Transformers With Linear Transformations
Viaarxiv icon

Knowledge is a Region in Weight Space for Fine-tuned Language Models

Add code
Bookmark button
Alert button
Feb 12, 2023
Almog Gueta, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen

Figure 1 for Knowledge is a Region in Weight Space for Fine-tuned Language Models
Figure 2 for Knowledge is a Region in Weight Space for Fine-tuned Language Models
Figure 3 for Knowledge is a Region in Weight Space for Fine-tuned Language Models
Figure 4 for Knowledge is a Region in Weight Space for Fine-tuned Language Models
Viaarxiv icon

Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

Add code
Bookmark button
Alert button
Jan 27, 2023
Alex Warstadt, Leshem Choshen, Aaron Mueller, Adina Williams, Ethan Wilcox, Chengxu Zhuang

Figure 1 for Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus
Figure 2 for Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus
Viaarxiv icon

ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning

Add code
Bookmark button
Alert button
Dec 02, 2022
Shachar Don-Yehiya, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen

Figure 1 for ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning
Figure 2 for ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning
Figure 3 for ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning
Figure 4 for ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning
Viaarxiv icon