Alert button
Picture for Moshe Wasserblat

Moshe Wasserblat

Alert button

CoTAR: Chain-of-Thought Attribution Reasoning with Multi-level Granularity

Add code
Bookmark button
Alert button
Apr 16, 2024
Moshe Berchansky, Daniel Fleischer, Moshe Wasserblat, Peter Izsak

Viaarxiv icon

Optimizing Retrieval-augmented Reader Models via Token Elimination

Add code
Bookmark button
Alert button
Oct 20, 2023
Moshe Berchansky, Peter Izsak, Avi Caciularu, Ido Dagan, Moshe Wasserblat

Viaarxiv icon

An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs

Add code
Bookmark button
Alert button
Jun 28, 2023
Haihao Shen, Hengyu Meng, Bo Dong, Zhe Wang, Ofir Zafrir, Yi Ding, Yu Luo, Hanwen Chang, Qun Gao, Ziheng Wang, Guy Boudoukh, Moshe Wasserblat

Figure 1 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Figure 2 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Figure 3 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Figure 4 for An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs
Viaarxiv icon

QuaLA-MiniLM: a Quantized Length Adaptive MiniLM

Add code
Bookmark button
Alert button
Oct 31, 2022
Shira Guskin, Moshe Wasserblat, Chang Wang, Haihao Shen

Figure 1 for QuaLA-MiniLM: a Quantized Length Adaptive MiniLM
Figure 2 for QuaLA-MiniLM: a Quantized Length Adaptive MiniLM
Figure 3 for QuaLA-MiniLM: a Quantized Length Adaptive MiniLM
Viaarxiv icon

Fast DistilBERT on CPUs

Add code
Bookmark button
Alert button
Oct 27, 2022
Haihao Shen, Ofir Zafrir, Bo Dong, Hengyu Meng, Xinyu Ye, Zhe Wang, Yi Ding, Hanwen Chang, Guy Boudoukh, Moshe Wasserblat

Figure 1 for Fast DistilBERT on CPUs
Figure 2 for Fast DistilBERT on CPUs
Figure 3 for Fast DistilBERT on CPUs
Figure 4 for Fast DistilBERT on CPUs
Viaarxiv icon

Cross-Domain Aspect Extraction using Transformers Augmented with Knowledge Graphs

Add code
Bookmark button
Alert button
Oct 18, 2022
Phillip Howard, Arden Ma, Vasudev Lal, Ana Paula Simoes, Daniel Korat, Oren Pereg, Moshe Wasserblat, Gadi Singer

Figure 1 for Cross-Domain Aspect Extraction using Transformers Augmented with Knowledge Graphs
Figure 2 for Cross-Domain Aspect Extraction using Transformers Augmented with Knowledge Graphs
Figure 3 for Cross-Domain Aspect Extraction using Transformers Augmented with Knowledge Graphs
Figure 4 for Cross-Domain Aspect Extraction using Transformers Augmented with Knowledge Graphs
Viaarxiv icon

Efficient Few-Shot Learning Without Prompts

Add code
Bookmark button
Alert button
Sep 22, 2022
Lewis Tunstall, Nils Reimers, Unso Eun Seo Jo, Luke Bates, Daniel Korat, Moshe Wasserblat, Oren Pereg

Figure 1 for Efficient Few-Shot Learning Without Prompts
Figure 2 for Efficient Few-Shot Learning Without Prompts
Figure 3 for Efficient Few-Shot Learning Without Prompts
Figure 4 for Efficient Few-Shot Learning Without Prompts
Viaarxiv icon

TangoBERT: Reducing Inference Cost by using Cascaded Architecture

Add code
Bookmark button
Alert button
Apr 13, 2022
Jonathan Mamou, Oren Pereg, Moshe Wasserblat, Roy Schwartz

Figure 1 for TangoBERT: Reducing Inference Cost by using Cascaded Architecture
Figure 2 for TangoBERT: Reducing Inference Cost by using Cascaded Architecture
Figure 3 for TangoBERT: Reducing Inference Cost by using Cascaded Architecture
Figure 4 for TangoBERT: Reducing Inference Cost by using Cascaded Architecture
Viaarxiv icon

Dynamic-TinyBERT: Boost TinyBERT's Inference Efficiency by Dynamic Sequence Length

Add code
Bookmark button
Alert button
Nov 18, 2021
Shira Guskin, Moshe Wasserblat, Ke Ding, Gyuwan Kim

Figure 1 for Dynamic-TinyBERT: Boost TinyBERT's Inference Efficiency by Dynamic Sequence Length
Figure 2 for Dynamic-TinyBERT: Boost TinyBERT's Inference Efficiency by Dynamic Sequence Length
Figure 3 for Dynamic-TinyBERT: Boost TinyBERT's Inference Efficiency by Dynamic Sequence Length
Figure 4 for Dynamic-TinyBERT: Boost TinyBERT's Inference Efficiency by Dynamic Sequence Length
Viaarxiv icon

Prune Once for All: Sparse Pre-Trained Language Models

Add code
Bookmark button
Alert button
Nov 10, 2021
Ofir Zafrir, Ariel Larey, Guy Boudoukh, Haihao Shen, Moshe Wasserblat

Figure 1 for Prune Once for All: Sparse Pre-Trained Language Models
Figure 2 for Prune Once for All: Sparse Pre-Trained Language Models
Figure 3 for Prune Once for All: Sparse Pre-Trained Language Models
Figure 4 for Prune Once for All: Sparse Pre-Trained Language Models
Viaarxiv icon