Alert button
Picture for S. Umesh

S. Umesh

Alert button

FusDom: Combining In-Domain and Out-of-Domain Knowledge for Continuous Self-Supervised Learning

Add code
Bookmark button
Alert button
Dec 20, 2023
Ashish Seth, Sreyan Ghosh, S. Umesh, Dinesh Manocha

Viaarxiv icon

Stable Distillation: Regularizing Continued Pre-training for Low-Resource Automatic Speech Recognition

Add code
Bookmark button
Alert button
Dec 20, 2023
Ashish Seth, Sreyan Ghosh, S. Umesh, Dinesh Manocha

Viaarxiv icon

The Tag-Team Approach: Leveraging CLS and Language Tagging for Enhancing Multilingual ASR

Add code
Bookmark button
Alert button
May 31, 2023
Kaousheik Jayakumar, Vrunda N. Sukhadia, A Arunkumar, S. Umesh

Figure 1 for The Tag-Team Approach: Leveraging CLS and Language Tagging for Enhancing Multilingual ASR
Figure 2 for The Tag-Team Approach: Leveraging CLS and Language Tagging for Enhancing Multilingual ASR
Figure 3 for The Tag-Team Approach: Leveraging CLS and Language Tagging for Enhancing Multilingual ASR
Figure 4 for The Tag-Team Approach: Leveraging CLS and Language Tagging for Enhancing Multilingual ASR
Viaarxiv icon

UNFUSED: UNsupervised Finetuning Using SElf supervised Distillation

Add code
Bookmark button
Alert button
Mar 10, 2023
Ashish Seth, Sreyan Ghosh, S. Umesh, Dinesh Manocha

Figure 1 for UNFUSED: UNsupervised Finetuning Using SElf supervised Distillation
Figure 2 for UNFUSED: UNsupervised Finetuning Using SElf supervised Distillation
Viaarxiv icon

Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR

Add code
Bookmark button
Alert button
Nov 03, 2022
Vrunda N. Sukhadia, A. Arunkumar, S. Umesh

Figure 1 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Figure 2 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Figure 3 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Figure 4 for Channel-Aware Pretraining of Joint Encoder-Decoder Self-Supervised Model for Telephonic-Speech ASR
Viaarxiv icon

SLICER: Learning universal audio representations using low-resource self-supervised pre-training

Add code
Bookmark button
Alert button
Nov 02, 2022
Ashish Seth, Sreyan Ghosh, S. Umesh, Dinesh Manocha

Figure 1 for SLICER: Learning universal audio representations using low-resource self-supervised pre-training
Figure 2 for SLICER: Learning universal audio representations using low-resource self-supervised pre-training
Figure 3 for SLICER: Learning universal audio representations using low-resource self-supervised pre-training
Figure 4 for SLICER: Learning universal audio representations using low-resource self-supervised pre-training
Viaarxiv icon

MAST: Multiscale Audio Spectrogram Transformers

Add code
Bookmark button
Alert button
Nov 02, 2022
Sreyan Ghosh, Ashish Seth, S. Umesh, Dinesh Manocha

Figure 1 for MAST: Multiscale Audio Spectrogram Transformers
Figure 2 for MAST: Multiscale Audio Spectrogram Transformers
Figure 3 for MAST: Multiscale Audio Spectrogram Transformers
Viaarxiv icon

data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup

Add code
Bookmark button
Alert button
Nov 02, 2022
Vasista Sai Lodagala, Sreyan Ghosh, S. Umesh

Figure 1 for data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup
Figure 2 for data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup
Figure 3 for data2vec-aqc: Search for the right Teaching Assistant in the Teacher-Student training setup
Viaarxiv icon

CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations

Add code
Bookmark button
Alert button
Oct 05, 2022
Vasista Sai Lodagala, Sreyan Ghosh, S. Umesh

Figure 1 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Figure 2 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Figure 3 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Figure 4 for CCC-wav2vec 2.0: Clustering aided Cross Contrastive Self-supervised learning of speech representations
Viaarxiv icon

Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition

Add code
Bookmark button
Alert button
Jun 11, 2022
A Arunkumar, Vrunda N Sukhadia, S. Umesh

Figure 1 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Figure 2 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Figure 3 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Figure 4 for Investigation of Ensemble features of Self-Supervised Pretrained Models for Automatic Speech Recognition
Viaarxiv icon