Picture for Shivani Agrawal

Shivani Agrawal

Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers

Add code
Feb 07, 2024
Figure 1 for Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers
Figure 2 for Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers
Figure 3 for Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers
Figure 4 for Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers
Viaarxiv icon

USM-Lite: Quantization and Sparsity Aware Fine-tuning for Speech Recognition with Universal Speech Models

Add code
Jan 03, 2024
Figure 1 for USM-Lite: Quantization and Sparsity Aware Fine-tuning for Speech Recognition with Universal Speech Models
Figure 2 for USM-Lite: Quantization and Sparsity Aware Fine-tuning for Speech Recognition with Universal Speech Models
Figure 3 for USM-Lite: Quantization and Sparsity Aware Fine-tuning for Speech Recognition with Universal Speech Models
Figure 4 for USM-Lite: Quantization and Sparsity Aware Fine-tuning for Speech Recognition with Universal Speech Models
Viaarxiv icon

Gemini: A Family of Highly Capable Multimodal Models

Add code
Dec 19, 2023
Viaarxiv icon

JaxPruner: A concise library for sparsity research

Add code
May 02, 2023
Figure 1 for JaxPruner: A concise library for sparsity research
Figure 2 for JaxPruner: A concise library for sparsity research
Viaarxiv icon

STEP: Learning N:M Structured Sparsity Masks from Scratch with Precondition

Add code
Feb 02, 2023
Figure 1 for STEP: Learning N:M Structured Sparsity Masks from Scratch with Precondition
Figure 2 for STEP: Learning N:M Structured Sparsity Masks from Scratch with Precondition
Figure 3 for STEP: Learning N:M Structured Sparsity Masks from Scratch with Precondition
Figure 4 for STEP: Learning N:M Structured Sparsity Masks from Scratch with Precondition
Viaarxiv icon

Efficiently Scaling Transformer Inference

Add code
Nov 09, 2022
Figure 1 for Efficiently Scaling Transformer Inference
Figure 2 for Efficiently Scaling Transformer Inference
Figure 3 for Efficiently Scaling Transformer Inference
Figure 4 for Efficiently Scaling Transformer Inference
Viaarxiv icon

Streaming Parrotron for on-device speech-to-speech conversion

Add code
Oct 25, 2022
Figure 1 for Streaming Parrotron for on-device speech-to-speech conversion
Figure 2 for Streaming Parrotron for on-device speech-to-speech conversion
Figure 3 for Streaming Parrotron for on-device speech-to-speech conversion
Figure 4 for Streaming Parrotron for on-device speech-to-speech conversion
Viaarxiv icon

Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask

Add code
Sep 15, 2022
Figure 1 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Figure 2 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Figure 3 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Figure 4 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Viaarxiv icon

PaLM: Scaling Language Modeling with Pathways

Add code
Apr 19, 2022
Figure 1 for PaLM: Scaling Language Modeling with Pathways
Figure 2 for PaLM: Scaling Language Modeling with Pathways
Figure 3 for PaLM: Scaling Language Modeling with Pathways
Figure 4 for PaLM: Scaling Language Modeling with Pathways
Viaarxiv icon

4-bit Conformer with Native Quantization Aware Training for Speech Recognition

Add code
Mar 29, 2022
Figure 1 for 4-bit Conformer with Native Quantization Aware Training for Speech Recognition
Figure 2 for 4-bit Conformer with Native Quantization Aware Training for Speech Recognition
Figure 3 for 4-bit Conformer with Native Quantization Aware Training for Speech Recognition
Figure 4 for 4-bit Conformer with Native Quantization Aware Training for Speech Recognition
Viaarxiv icon