Alert button
Picture for Marc Delcroix

Marc Delcroix

Alert button

Mask-based Neural Beamforming for Moving Speakers with Self-Attention-based Tracking

Add code
Bookmark button
Alert button
May 07, 2022
Tsubasa Ochiai, Marc Delcroix, Tomohiro Nakatani, Shoko Araki

Figure 1 for Mask-based Neural Beamforming for Moving Speakers with Self-Attention-based Tracking
Figure 2 for Mask-based Neural Beamforming for Moving Speakers with Self-Attention-based Tracking
Figure 3 for Mask-based Neural Beamforming for Moving Speakers with Self-Attention-based Tracking
Figure 4 for Mask-based Neural Beamforming for Moving Speakers with Self-Attention-based Tracking
Viaarxiv icon

Listen only to me! How well can target speech extraction handle false alarms?

Add code
Bookmark button
Alert button
Apr 11, 2022
Marc Delcroix, Keisuke Kinoshita, Tsubasa Ochiai, Katerina Zmolikova, Hiroshi Sato, Tomohiro Nakatani

Figure 1 for Listen only to me! How well can target speech extraction handle false alarms?
Figure 2 for Listen only to me! How well can target speech extraction handle false alarms?
Figure 3 for Listen only to me! How well can target speech extraction handle false alarms?
Figure 4 for Listen only to me! How well can target speech extraction handle false alarms?
Viaarxiv icon

SoundBeam: Target sound extraction conditioned on sound-class labels and enrollment clues for increased performance and continuous learning

Add code
Bookmark button
Alert button
Apr 08, 2022
Marc Delcroix, Jorge Bennasar Vázquez, Tsubasa Ochiai, Keisuke Kinoshita, Yasunori Ohishi, Shoko Araki

Figure 1 for SoundBeam: Target sound extraction conditioned on sound-class labels and enrollment clues for increased performance and continuous learning
Figure 2 for SoundBeam: Target sound extraction conditioned on sound-class labels and enrollment clues for increased performance and continuous learning
Figure 3 for SoundBeam: Target sound extraction conditioned on sound-class labels and enrollment clues for increased performance and continuous learning
Figure 4 for SoundBeam: Target sound extraction conditioned on sound-class labels and enrollment clues for increased performance and continuous learning
Viaarxiv icon

Tight integration of neural- and clustering-based diarization through deep unfolding of infinite Gaussian mixture model

Add code
Bookmark button
Alert button
Feb 14, 2022
Keisuke Kinoshita, Marc Delcroix, Tomoharu Iwata

Figure 1 for Tight integration of neural- and clustering-based diarization through deep unfolding of infinite Gaussian mixture model
Figure 2 for Tight integration of neural- and clustering-based diarization through deep unfolding of infinite Gaussian mixture model
Figure 3 for Tight integration of neural- and clustering-based diarization through deep unfolding of infinite Gaussian mixture model
Viaarxiv icon

How Bad Are Artifacts?: Analyzing the Impact of Speech Enhancement Errors on ASR

Add code
Bookmark button
Alert button
Jan 18, 2022
Kazuma Iwamoto, Tsubasa Ochiai, Marc Delcroix, Rintaro Ikeshita, Hiroshi Sato, Shoko Araki, Shigeru Katagiri

Figure 1 for How Bad Are Artifacts?: Analyzing the Impact of Speech Enhancement Errors on ASR
Figure 2 for How Bad Are Artifacts?: Analyzing the Impact of Speech Enhancement Errors on ASR
Figure 3 for How Bad Are Artifacts?: Analyzing the Impact of Speech Enhancement Errors on ASR
Figure 4 for How Bad Are Artifacts?: Analyzing the Impact of Speech Enhancement Errors on ASR
Viaarxiv icon

Learning to Enhance or Not: Neural Network-Based Switching of Enhanced and Observed Signals for Overlapping Speech Recognition

Add code
Bookmark button
Alert button
Jan 11, 2022
Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Naoyuki Kamo, Takafumi Moriya

Figure 1 for Learning to Enhance or Not: Neural Network-Based Switching of Enhanced and Observed Signals for Overlapping Speech Recognition
Figure 2 for Learning to Enhance or Not: Neural Network-Based Switching of Enhanced and Observed Signals for Overlapping Speech Recognition
Figure 3 for Learning to Enhance or Not: Neural Network-Based Switching of Enhanced and Observed Signals for Overlapping Speech Recognition
Viaarxiv icon

Attention-based Multi-hypothesis Fusion for Speech Summarization

Add code
Bookmark button
Alert button
Nov 16, 2021
Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe

Figure 1 for Attention-based Multi-hypothesis Fusion for Speech Summarization
Figure 2 for Attention-based Multi-hypothesis Fusion for Speech Summarization
Figure 3 for Attention-based Multi-hypothesis Fusion for Speech Summarization
Figure 4 for Attention-based Multi-hypothesis Fusion for Speech Summarization
Viaarxiv icon

Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model

Add code
Bookmark button
Alert button
Oct 31, 2021
Martin Kocour, Kateřina Žmolíková, Lucas Ondel, Ján Švec, Marc Delcroix, Tsubasa Ochiai, Lukáš Burget, Jan Černocký

Figure 1 for Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model
Figure 2 for Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model
Figure 3 for Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model
Figure 4 for Revisiting joint decoding based multi-talker speech recognition with DNN acoustic model
Viaarxiv icon

SA-SDR: A novel loss function for separation of meeting style data

Add code
Bookmark button
Alert button
Oct 29, 2021
Thilo von Neumann, Keisuke Kinoshita, Christoph Boeddeker, Marc Delcroix, Reinhold Haeb-Umbach

Figure 1 for SA-SDR: A novel loss function for separation of meeting style data
Figure 2 for SA-SDR: A novel loss function for separation of meeting style data
Viaarxiv icon

Graph-PIT: Generalized permutation invariant training for continuous separation of arbitrary numbers of speakers

Add code
Bookmark button
Alert button
Jul 30, 2021
Thilo von Neumann, Keisuke Kinoshita, Christoph Boeddeker, Marc Delcroix, Reinhold Haeb-Umbach

Figure 1 for Graph-PIT: Generalized permutation invariant training for continuous separation of arbitrary numbers of speakers
Figure 2 for Graph-PIT: Generalized permutation invariant training for continuous separation of arbitrary numbers of speakers
Figure 3 for Graph-PIT: Generalized permutation invariant training for continuous separation of arbitrary numbers of speakers
Viaarxiv icon