Alert button
Picture for John R. Hershey

John R. Hershey

Alert button

DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement

Add code
Bookmark button
Alert button
Jun 30, 2021
Yuma Koizumi, Shigeki Karita, Scott Wisdom, Hakan Erdogan, John R. Hershey, Llion Jones, Michiel Bacchiani

Figure 1 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Figure 2 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Figure 3 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Figure 4 for DF-Conformer: Integrated architecture of Conv-TasNet and Conformer using linear complexity self-attention for speech enhancement
Viaarxiv icon

Improving On-Screen Sound Separation for Open Domain Videos with Audio-Visual Self-attention

Add code
Bookmark button
Alert button
Jun 17, 2021
Efthymios Tzinis, Scott Wisdom, Tal Remez, John R. Hershey

Figure 1 for Improving On-Screen Sound Separation for Open Domain Videos with Audio-Visual Self-attention
Figure 2 for Improving On-Screen Sound Separation for Open Domain Videos with Audio-Visual Self-attention
Figure 3 for Improving On-Screen Sound Separation for Open Domain Videos with Audio-Visual Self-attention
Figure 4 for Improving On-Screen Sound Separation for Open Domain Videos with Audio-Visual Self-attention
Viaarxiv icon

Sparse, Efficient, and Semantic Mixture Invariant Training: Taming In-the-Wild Unsupervised Sound Separation

Add code
Bookmark button
Alert button
Jun 01, 2021
Scott Wisdom, Aren Jansen, Ron J. Weiss, Hakan Erdogan, John R. Hershey

Figure 1 for Sparse, Efficient, and Semantic Mixture Invariant Training: Taming In-the-Wild Unsupervised Sound Separation
Figure 2 for Sparse, Efficient, and Semantic Mixture Invariant Training: Taming In-the-Wild Unsupervised Sound Separation
Viaarxiv icon

Self-Supervised Learning from Automatically Separated Sound Scenes

Add code
Bookmark button
Alert button
May 05, 2021
Eduardo Fonseca, Aren Jansen, Daniel P. W. Ellis, Scott Wisdom, Marco Tagliasacchi, John R. Hershey, Manoj Plakal, Shawn Hershey, R. Channing Moore, Xavier Serra

Figure 1 for Self-Supervised Learning from Automatically Separated Sound Scenes
Figure 2 for Self-Supervised Learning from Automatically Separated Sound Scenes
Figure 3 for Self-Supervised Learning from Automatically Separated Sound Scenes
Figure 4 for Self-Supervised Learning from Automatically Separated Sound Scenes
Viaarxiv icon

End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings

Add code
Bookmark button
Alert button
May 05, 2021
Soumi Maiti, Hakan Erdogan, Kevin Wilson, Scott Wisdom, Shinji Watanabe, John R. Hershey

Figure 1 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Figure 2 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Figure 3 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Figure 4 for End-to-End Diarization for Variable Number of Speakers with Local-Global Networks and Discriminative Speaker Embeddings
Viaarxiv icon

Into the Wild with AudioScope: Unsupervised Audio-Visual Separation of On-Screen Sounds

Add code
Bookmark button
Alert button
Nov 02, 2020
Efthymios Tzinis, Scott Wisdom, Aren Jansen, Shawn Hershey, Tal Remez, Daniel P. W. Ellis, John R. Hershey

Figure 1 for Into the Wild with AudioScope: Unsupervised Audio-Visual Separation of On-Screen Sounds
Figure 2 for Into the Wild with AudioScope: Unsupervised Audio-Visual Separation of On-Screen Sounds
Figure 3 for Into the Wild with AudioScope: Unsupervised Audio-Visual Separation of On-Screen Sounds
Figure 4 for Into the Wild with AudioScope: Unsupervised Audio-Visual Separation of On-Screen Sounds
Viaarxiv icon

Unsupervised Sound Separation Using Mixtures of Mixtures

Add code
Bookmark button
Alert button
Jun 23, 2020
Scott Wisdom, Efthymios Tzinis, Hakan Erdogan, Ron J. Weiss, Kevin Wilson, John R. Hershey

Figure 1 for Unsupervised Sound Separation Using Mixtures of Mixtures
Figure 2 for Unsupervised Sound Separation Using Mixtures of Mixtures
Figure 3 for Unsupervised Sound Separation Using Mixtures of Mixtures
Figure 4 for Unsupervised Sound Separation Using Mixtures of Mixtures
Viaarxiv icon

Alternating Between Spectral and Spatial Estimation for Speech Separation and Enhancement

Add code
Bookmark button
Alert button
Nov 18, 2019
Zhong-Qiu Wang, Scott Wisdom, Kevin Wilson, John R. Hershey

Figure 1 for Alternating Between Spectral and Spatial Estimation for Speech Separation and Enhancement
Figure 2 for Alternating Between Spectral and Spatial Estimation for Speech Separation and Enhancement
Viaarxiv icon

Improving Universal Sound Separation Using Sound Classification

Add code
Bookmark button
Alert button
Nov 18, 2019
Efthymios Tzinis, Scott Wisdom, John R. Hershey, Aren Jansen, Daniel P. W. Ellis

Figure 1 for Improving Universal Sound Separation Using Sound Classification
Figure 2 for Improving Universal Sound Separation Using Sound Classification
Figure 3 for Improving Universal Sound Separation Using Sound Classification
Figure 4 for Improving Universal Sound Separation Using Sound Classification
Viaarxiv icon

Universal Sound Separation

Add code
Bookmark button
Alert button
May 08, 2019
Ilya Kavalerov, Scott Wisdom, Hakan Erdogan, Brian Patton, Kevin Wilson, Jonathan Le Roux, John R. Hershey

Figure 1 for Universal Sound Separation
Figure 2 for Universal Sound Separation
Figure 3 for Universal Sound Separation
Figure 4 for Universal Sound Separation
Viaarxiv icon