Alert button
Picture for Mateusz Lajszczak

Mateusz Lajszczak

Alert button

Enhancing the Stability of LLM-based Speech Generation Systems through Self-Supervised Representations

Add code
Bookmark button
Alert button
Feb 05, 2024
Álvaro Martín-Cortinas, Daniel Sáez-Trigueros, Iván Vallés-Pérez, Biel Tura-Vecino, Piotr Biliński, Mateusz Lajszczak, Grzegorz Beringer, Roberto Barra-Chicote, Jaime Lorenzo-Trueba

Viaarxiv icon

Controllable Emphasis with zero data for text-to-speech

Add code
Bookmark button
Alert button
Jul 13, 2023
Arnaud Joly, Marco Nicolis, Ekaterina Peterova, Alessandro Lombardi, Ammar Abbas, Arent van Korlaar, Aman Hussain, Parul Sharma, Alexis Moinet, Mateusz Lajszczak, Penny Karanasou, Antonio Bonafonte, Thomas Drugman, Elena Sokolova

Figure 1 for Controllable Emphasis with zero data for text-to-speech
Figure 2 for Controllable Emphasis with zero data for text-to-speech
Figure 3 for Controllable Emphasis with zero data for text-to-speech
Figure 4 for Controllable Emphasis with zero data for text-to-speech
Viaarxiv icon

CopyCat2: A Single Model for Multi-Speaker TTS and Many-to-Many Fine-Grained Prosody Transfer

Add code
Bookmark button
Alert button
Jun 27, 2022
Sri Karlapati, Penny Karanasou, Mateusz Lajszczak, Ammar Abbas, Alexis Moinet, Peter Makarov, Ray Li, Arent van Korlaar, Simon Slangen, Thomas Drugman

Figure 1 for CopyCat2: A Single Model for Multi-Speaker TTS and Many-to-Many Fine-Grained Prosody Transfer
Figure 2 for CopyCat2: A Single Model for Multi-Speaker TTS and Many-to-Many Fine-Grained Prosody Transfer
Viaarxiv icon

Distribution augmentation for low-resource expressive text-to-speech

Add code
Bookmark button
Alert button
Feb 19, 2022
Mateusz Lajszczak, Animesh Prasad, Arent van Korlaar, Bajibabu Bollepalli, Antonio Bonafonte, Arnaud Joly, Marco Nicolis, Alexis Moinet, Thomas Drugman, Trevor Wood, Elena Sokolova

Figure 1 for Distribution augmentation for low-resource expressive text-to-speech
Figure 2 for Distribution augmentation for low-resource expressive text-to-speech
Figure 3 for Distribution augmentation for low-resource expressive text-to-speech
Figure 4 for Distribution augmentation for low-resource expressive text-to-speech
Viaarxiv icon

Interpretable Deep Learning Model for the Detection and Reconstruction of Dysarthric Speech

Add code
Bookmark button
Alert button
Jul 10, 2019
Daniel Korzekwa, Roberto Barra-Chicote, Bozena Kostek, Thomas Drugman, Mateusz Lajszczak

Figure 1 for Interpretable Deep Learning Model for the Detection and Reconstruction of Dysarthric Speech
Figure 2 for Interpretable Deep Learning Model for the Detection and Reconstruction of Dysarthric Speech
Figure 3 for Interpretable Deep Learning Model for the Detection and Reconstruction of Dysarthric Speech
Figure 4 for Interpretable Deep Learning Model for the Detection and Reconstruction of Dysarthric Speech
Viaarxiv icon