Alert button
Picture for Jaesung Huh

Jaesung Huh

Alert button

TIM: A Time Interval Machine for Audio-Visual Action Recognition

Add code
Bookmark button
Alert button
Apr 09, 2024
Jacob Chalk, Jaesung Huh, Evangelos Kazakos, Andrew Zisserman, Dima Damen

Viaarxiv icon

Look, Listen and Recognise: Character-Aware Audio-Visual Subtitling

Add code
Bookmark button
Alert button
Jan 22, 2024
Bruno Korbar, Jaesung Huh, Andrew Zisserman

Viaarxiv icon

OxfordVGG Submission to the EGO4D AV Transcription Challenge

Add code
Bookmark button
Alert button
Jul 18, 2023
Jaesung Huh, Max Bain, Andrew Zisserman

Figure 1 for OxfordVGG Submission to the EGO4D AV Transcription Challenge
Figure 2 for OxfordVGG Submission to the EGO4D AV Transcription Challenge
Viaarxiv icon

VoxSRC 2022: The Fourth VoxCeleb Speaker Recognition Challenge

Add code
Bookmark button
Alert button
Mar 06, 2023
Jaesung Huh, Andrew Brown, Jee-weon Jung, Joon Son Chung, Arsha Nagrani, Daniel Garcia-Romero, Andrew Zisserman

Figure 1 for VoxSRC 2022: The Fourth VoxCeleb Speaker Recognition Challenge
Figure 2 for VoxSRC 2022: The Fourth VoxCeleb Speaker Recognition Challenge
Figure 3 for VoxSRC 2022: The Fourth VoxCeleb Speaker Recognition Challenge
Figure 4 for VoxSRC 2022: The Fourth VoxCeleb Speaker Recognition Challenge
Viaarxiv icon

WhisperX: Time-Accurate Speech Transcription of Long-Form Audio

Add code
Bookmark button
Alert button
Mar 01, 2023
Max Bain, Jaesung Huh, Tengda Han, Andrew Zisserman

Figure 1 for WhisperX: Time-Accurate Speech Transcription of Long-Form Audio
Figure 2 for WhisperX: Time-Accurate Speech Transcription of Long-Form Audio
Figure 3 for WhisperX: Time-Accurate Speech Transcription of Long-Form Audio
Figure 4 for WhisperX: Time-Accurate Speech Transcription of Long-Form Audio
Viaarxiv icon

Epic-Sounds: A Large-scale Dataset of Actions That Sound

Add code
Bookmark button
Alert button
Feb 01, 2023
Jaesung Huh, Jacob Chalk, Evangelos Kazakos, Dima Damen, Andrew Zisserman

Figure 1 for Epic-Sounds: A Large-scale Dataset of Actions That Sound
Figure 2 for Epic-Sounds: A Large-scale Dataset of Actions That Sound
Figure 3 for Epic-Sounds: A Large-scale Dataset of Actions That Sound
Figure 4 for Epic-Sounds: A Large-scale Dataset of Actions That Sound
Viaarxiv icon

In search of strong embedding extractors for speaker diarisation

Add code
Bookmark button
Alert button
Oct 26, 2022
Jee-weon Jung, Hee-Soo Heo, Bong-Jin Lee, Jaesung Huh, Andrew Brown, Youngki Kwon, Shinji Watanabe, Joon Son Chung

Figure 1 for In search of strong embedding extractors for speaker diarisation
Figure 2 for In search of strong embedding extractors for speaker diarisation
Figure 3 for In search of strong embedding extractors for speaker diarisation
Viaarxiv icon

VoxSRC 2021: The Third VoxCeleb Speaker Recognition Challenge

Add code
Bookmark button
Alert button
Jan 12, 2022
Andrew Brown, Jaesung Huh, Joon Son Chung, Arsha Nagrani, Andrew Zisserman

Figure 1 for VoxSRC 2021: The Third VoxCeleb Speaker Recognition Challenge
Figure 2 for VoxSRC 2021: The Third VoxCeleb Speaker Recognition Challenge
Figure 3 for VoxSRC 2021: The Third VoxCeleb Speaker Recognition Challenge
Figure 4 for VoxSRC 2021: The Third VoxCeleb Speaker Recognition Challenge
Viaarxiv icon

With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition

Add code
Bookmark button
Alert button
Nov 01, 2021
Evangelos Kazakos, Jaesung Huh, Arsha Nagrani, Andrew Zisserman, Dima Damen

Figure 1 for With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition
Figure 2 for With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition
Figure 3 for With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition
Figure 4 for With a Little Help from my Temporal Context: Multimodal Egocentric Action Recognition
Viaarxiv icon