Alert button
Picture for Rodrigo Mira

Rodrigo Mira

Alert button

BRAVEn: Improving Self-Supervised Pre-training for Visual and Auditory Speech Recognition

Add code
Bookmark button
Alert button
Apr 02, 2024
Alexandros Haliassos, Andreas Zinonos, Rodrigo Mira, Stavros Petridis, Maja Pantic

Viaarxiv icon

Laughing Matters: Introducing Laughing-Face Generation using Diffusion Models

Add code
Bookmark button
Alert button
May 15, 2023
Antoni Bigata Casademunt, Rodrigo Mira, Nikita Drobyshev, Konstantinos Vougioukas, Stavros Petridis, Maja Pantic

Figure 1 for Laughing Matters: Introducing Laughing-Face Generation using Diffusion Models
Figure 2 for Laughing Matters: Introducing Laughing-Face Generation using Diffusion Models
Figure 3 for Laughing Matters: Introducing Laughing-Face Generation using Diffusion Models
Figure 4 for Laughing Matters: Introducing Laughing-Face Generation using Diffusion Models
Viaarxiv icon

Jointly Learning Visual and Auditory Speech Representations from Raw Data

Add code
Bookmark button
Alert button
Dec 12, 2022
Alexandros Haliassos, Pingchuan Ma, Rodrigo Mira, Stavros Petridis, Maja Pantic

Figure 1 for Jointly Learning Visual and Auditory Speech Representations from Raw Data
Figure 2 for Jointly Learning Visual and Auditory Speech Representations from Raw Data
Figure 3 for Jointly Learning Visual and Auditory Speech Representations from Raw Data
Figure 4 for Jointly Learning Visual and Auditory Speech Representations from Raw Data
Viaarxiv icon

LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders

Add code
Bookmark button
Alert button
Nov 20, 2022
Rodrigo Mira, Buye Xu, Jacob Donley, Anurag Kumar, Stavros Petridis, Vamsi Krishna Ithapu, Maja Pantic

Figure 1 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Figure 2 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Figure 3 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Figure 4 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Viaarxiv icon

SVTS: Scalable Video-to-Speech Synthesis

Add code
Bookmark button
Alert button
May 04, 2022
Rodrigo Mira, Alexandros Haliassos, Stavros Petridis, Björn W. Schuller, Maja Pantic

Figure 1 for SVTS: Scalable Video-to-Speech Synthesis
Figure 2 for SVTS: Scalable Video-to-Speech Synthesis
Figure 3 for SVTS: Scalable Video-to-Speech Synthesis
Figure 4 for SVTS: Scalable Video-to-Speech Synthesis
Viaarxiv icon

Leveraging Real Talking Faces via Self-Supervision for Robust Forgery Detection

Add code
Bookmark button
Alert button
Jan 18, 2022
Alexandros Haliassos, Rodrigo Mira, Stavros Petridis, Maja Pantic

Figure 1 for Leveraging Real Talking Faces via Self-Supervision for Robust Forgery Detection
Figure 2 for Leveraging Real Talking Faces via Self-Supervision for Robust Forgery Detection
Figure 3 for Leveraging Real Talking Faces via Self-Supervision for Robust Forgery Detection
Figure 4 for Leveraging Real Talking Faces via Self-Supervision for Robust Forgery Detection
Viaarxiv icon

LiRA: Learning Visual Speech Representations from Audio through Self-supervision

Add code
Bookmark button
Alert button
Jun 16, 2021
Pingchuan Ma, Rodrigo Mira, Stavros Petridis, Björn W. Schuller, Maja Pantic

Figure 1 for LiRA: Learning Visual Speech Representations from Audio through Self-supervision
Figure 2 for LiRA: Learning Visual Speech Representations from Audio through Self-supervision
Figure 3 for LiRA: Learning Visual Speech Representations from Audio through Self-supervision
Figure 4 for LiRA: Learning Visual Speech Representations from Audio through Self-supervision
Viaarxiv icon

End-to-End Video-To-Speech Synthesis using Generative Adversarial Networks

Add code
Bookmark button
Alert button
Apr 30, 2021
Rodrigo Mira, Konstantinos Vougioukas, Pingchuan Ma, Stavros Petridis, Björn W. Schuller, Maja Pantic

Figure 1 for End-to-End Video-To-Speech Synthesis using Generative Adversarial Networks
Figure 2 for End-to-End Video-To-Speech Synthesis using Generative Adversarial Networks
Figure 3 for End-to-End Video-To-Speech Synthesis using Generative Adversarial Networks
Figure 4 for End-to-End Video-To-Speech Synthesis using Generative Adversarial Networks
Viaarxiv icon