Alert button
Picture for Vamsi Krishna Ithapu

Vamsi Krishna Ithapu

Alert button

Hearing Loss Detection from Facial Expressions in One-on-one Conversations

Add code
Bookmark button
Alert button
Jan 17, 2024
Yufeng Yin, Ishwarya Ananthabhotla, Vamsi Krishna Ithapu, Stavros Petridis, Yu-Hsiang Wu, Christi Miller

Viaarxiv icon

The Audio-Visual Conversational Graph: From an Egocentric-Exocentric Perspective

Add code
Bookmark button
Alert button
Dec 20, 2023
Wenqi Jia, Miao Liu, Hao Jiang, Ishwarya Ananthabhotla, James M. Rehg, Vamsi Krishna Ithapu, Ruohan Gao

Viaarxiv icon

Egocentric Auditory Attention Localization in Conversations

Add code
Bookmark button
Alert button
Mar 28, 2023
Fiona Ryan, Hao Jiang, Abhinav Shukla, James M. Rehg, Vamsi Krishna Ithapu

Figure 1 for Egocentric Auditory Attention Localization in Conversations
Figure 2 for Egocentric Auditory Attention Localization in Conversations
Figure 3 for Egocentric Auditory Attention Localization in Conversations
Figure 4 for Egocentric Auditory Attention Localization in Conversations
Viaarxiv icon

Novel-View Acoustic Synthesis

Add code
Bookmark button
Alert button
Jan 23, 2023
Changan Chen, Alexander Richard, Roman Shapovalov, Vamsi Krishna Ithapu, Natalia Neverova, Kristen Grauman, Andrea Vedaldi

Figure 1 for Novel-View Acoustic Synthesis
Figure 2 for Novel-View Acoustic Synthesis
Figure 3 for Novel-View Acoustic Synthesis
Figure 4 for Novel-View Acoustic Synthesis
Viaarxiv icon

Chat2Map: Efficient Scene Mapping from Multi-Ego Conversations

Add code
Bookmark button
Alert button
Jan 04, 2023
Sagnik Majumder, Hao Jiang, Pierre Moulon, Ethan Henderson, Paul Calamia, Kristen Grauman, Vamsi Krishna Ithapu

Figure 1 for Chat2Map: Efficient Scene Mapping from Multi-Ego Conversations
Figure 2 for Chat2Map: Efficient Scene Mapping from Multi-Ego Conversations
Figure 3 for Chat2Map: Efficient Scene Mapping from Multi-Ego Conversations
Figure 4 for Chat2Map: Efficient Scene Mapping from Multi-Ego Conversations
Viaarxiv icon

LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders

Add code
Bookmark button
Alert button
Nov 20, 2022
Rodrigo Mira, Buye Xu, Jacob Donley, Anurag Kumar, Stavros Petridis, Vamsi Krishna Ithapu, Maja Pantic

Figure 1 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Figure 2 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Figure 3 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Figure 4 for LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Viaarxiv icon

Leveraging Heteroscedastic Uncertainty in Learning Complex Spectral Mapping for Single-channel Speech Enhancement

Add code
Bookmark button
Alert button
Nov 16, 2022
Kuan-Lin Chen, Daniel D. E. Wong, Ke Tan, Buye Xu, Anurag Kumar, Vamsi Krishna Ithapu

Figure 1 for Leveraging Heteroscedastic Uncertainty in Learning Complex Spectral Mapping for Single-channel Speech Enhancement
Figure 2 for Leveraging Heteroscedastic Uncertainty in Learning Complex Spectral Mapping for Single-channel Speech Enhancement
Figure 3 for Leveraging Heteroscedastic Uncertainty in Learning Complex Spectral Mapping for Single-channel Speech Enhancement
Viaarxiv icon

Towards Improved Room Impulse Response Estimation for Speech Recognition

Add code
Bookmark button
Alert button
Nov 08, 2022
Anton Ratnarajah, Ishwarya Ananthabhotla, Vamsi Krishna Ithapu, Pablo Hoffmann, Dinesh Manocha, Paul Calamia

Figure 1 for Towards Improved Room Impulse Response Estimation for Speech Recognition
Figure 2 for Towards Improved Room Impulse Response Estimation for Speech Recognition
Figure 3 for Towards Improved Room Impulse Response Estimation for Speech Recognition
Figure 4 for Towards Improved Room Impulse Response Estimation for Speech Recognition
Viaarxiv icon

RemixIT: Continual self-training of speech enhancement models via bootstrapped remixing

Add code
Bookmark button
Alert button
Feb 22, 2022
Efthymios Tzinis, Yossi Adi, Vamsi Krishna Ithapu, Buye Xu, Paris Smaragdis, Anurag Kumar

Figure 1 for RemixIT: Continual self-training of speech enhancement models via bootstrapped remixing
Figure 2 for RemixIT: Continual self-training of speech enhancement models via bootstrapped remixing
Figure 3 for RemixIT: Continual self-training of speech enhancement models via bootstrapped remixing
Figure 4 for RemixIT: Continual self-training of speech enhancement models via bootstrapped remixing
Viaarxiv icon

Deep Impulse Responses: Estimating and Parameterizing Filters with Deep Networks

Add code
Bookmark button
Alert button
Feb 07, 2022
Alexander Richard, Peter Dodds, Vamsi Krishna Ithapu

Figure 1 for Deep Impulse Responses: Estimating and Parameterizing Filters with Deep Networks
Figure 2 for Deep Impulse Responses: Estimating and Parameterizing Filters with Deep Networks
Figure 3 for Deep Impulse Responses: Estimating and Parameterizing Filters with Deep Networks
Figure 4 for Deep Impulse Responses: Estimating and Parameterizing Filters with Deep Networks
Viaarxiv icon