Alert button
Picture for Andrei Andrusenko

Andrei Andrusenko

Alert button

SALM: Speech-augmented Language Model with In-context Learning for Speech Recognition and Translation

Add code
Bookmark button
Alert button
Oct 13, 2023
Zhehuai Chen, He Huang, Andrei Andrusenko, Oleksii Hrinchuk, Krishna C. Puvvada, Jason Li, Subhankar Ghosh, Jagadeesh Balam, Boris Ginsburg

Viaarxiv icon

Uconv-Conformer: High Reduction of Input Sequence Length for End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Aug 16, 2022
Andrei Andrusenko, Rauf Nasretdinov, Aleksei Romanenko

Figure 1 for Uconv-Conformer: High Reduction of Input Sequence Length for End-to-End Speech Recognition
Figure 2 for Uconv-Conformer: High Reduction of Input Sequence Length for End-to-End Speech Recognition
Figure 3 for Uconv-Conformer: High Reduction of Input Sequence Length for End-to-End Speech Recognition
Figure 4 for Uconv-Conformer: High Reduction of Input Sequence Length for End-to-End Speech Recognition
Viaarxiv icon

LT-LM: a novel non-autoregressive language model for single-shot lattice rescoring

Add code
Bookmark button
Alert button
Apr 06, 2021
Anton Mitrofanov, Mariya Korenevskaya, Ivan Podluzhny, Yuri Khokhlov, Aleksandr Laptev, Andrei Andrusenko, Aleksei Ilin, Maxim Korenevsky, Ivan Medennikov, Aleksei Romanenko

Figure 1 for LT-LM: a novel non-autoregressive language model for single-shot lattice rescoring
Figure 2 for LT-LM: a novel non-autoregressive language model for single-shot lattice rescoring
Figure 3 for LT-LM: a novel non-autoregressive language model for single-shot lattice rescoring
Figure 4 for LT-LM: a novel non-autoregressive language model for single-shot lattice rescoring
Viaarxiv icon

Dynamic Acoustic Unit Augmentation With BPE-Dropout for Low-Resource End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Mar 12, 2021
Aleksandr Laptev, Andrei Andrusenko, Ivan Podluzhny, Anton Mitrofanov, Ivan Medennikov, Yuri Matveev

Figure 1 for Dynamic Acoustic Unit Augmentation With BPE-Dropout for Low-Resource End-to-End Speech Recognition
Figure 2 for Dynamic Acoustic Unit Augmentation With BPE-Dropout for Low-Resource End-to-End Speech Recognition
Figure 3 for Dynamic Acoustic Unit Augmentation With BPE-Dropout for Low-Resource End-to-End Speech Recognition
Figure 4 for Dynamic Acoustic Unit Augmentation With BPE-Dropout for Low-Resource End-to-End Speech Recognition
Viaarxiv icon

Exploration of End-to-End ASR for OpenSTT -- Russian Open Speech-to-Text Dataset

Add code
Bookmark button
Alert button
Jun 15, 2020
Andrei Andrusenko, Aleksandr Laptev, Ivan Medennikov

Figure 1 for Exploration of End-to-End ASR for OpenSTT -- Russian Open Speech-to-Text Dataset
Figure 2 for Exploration of End-to-End ASR for OpenSTT -- Russian Open Speech-to-Text Dataset
Figure 3 for Exploration of End-to-End ASR for OpenSTT -- Russian Open Speech-to-Text Dataset
Viaarxiv icon

Target-Speaker Voice Activity Detection: a Novel Approach for Multi-Speaker Diarization in a Dinner Party Scenario

Add code
Bookmark button
Alert button
May 14, 2020
Ivan Medennikov, Maxim Korenevsky, Tatiana Prisyach, Yuri Khokhlov, Mariya Korenevskaya, Ivan Sorokin, Tatiana Timofeeva, Anton Mitrofanov, Andrei Andrusenko, Ivan Podluzhny, Aleksandr Laptev, Aleksei Romanenko

Figure 1 for Target-Speaker Voice Activity Detection: a Novel Approach for Multi-Speaker Diarization in a Dinner Party Scenario
Figure 2 for Target-Speaker Voice Activity Detection: a Novel Approach for Multi-Speaker Diarization in a Dinner Party Scenario
Figure 3 for Target-Speaker Voice Activity Detection: a Novel Approach for Multi-Speaker Diarization in a Dinner Party Scenario
Figure 4 for Target-Speaker Voice Activity Detection: a Novel Approach for Multi-Speaker Diarization in a Dinner Party Scenario
Viaarxiv icon

You Do Not Need More Data: Improving End-To-End Speech Recognition by Text-To-Speech Data Augmentation

Add code
Bookmark button
Alert button
May 14, 2020
Aleksandr Laptev, Roman Korostik, Aleksey Svischev, Andrei Andrusenko, Ivan Medennikov, Sergey Rybin

Figure 1 for You Do Not Need More Data: Improving End-To-End Speech Recognition by Text-To-Speech Data Augmentation
Figure 2 for You Do Not Need More Data: Improving End-To-End Speech Recognition by Text-To-Speech Data Augmentation
Figure 3 for You Do Not Need More Data: Improving End-To-End Speech Recognition by Text-To-Speech Data Augmentation
Figure 4 for You Do Not Need More Data: Improving End-To-End Speech Recognition by Text-To-Speech Data Augmentation
Viaarxiv icon

Towards a Competitive End-to-End Speech Recognition for CHiME-6 Dinner Party Transcription

Add code
Bookmark button
Alert button
Apr 24, 2020
Andrei Andrusenko, Aleksandr Laptev, Ivan Medennikov

Figure 1 for Towards a Competitive End-to-End Speech Recognition for CHiME-6 Dinner Party Transcription
Figure 2 for Towards a Competitive End-to-End Speech Recognition for CHiME-6 Dinner Party Transcription
Figure 3 for Towards a Competitive End-to-End Speech Recognition for CHiME-6 Dinner Party Transcription
Figure 4 for Towards a Competitive End-to-End Speech Recognition for CHiME-6 Dinner Party Transcription
Viaarxiv icon