Alert button

"speech": models, code, and papers
Alert button

ChatGPT v.s. Media Bias: A Comparative Study of GPT-3.5 and Fine-tuned Language Models

Mar 29, 2024
Zehao Wen, Rabih Younes

Figure 1 for ChatGPT v.s. Media Bias: A Comparative Study of GPT-3.5 and Fine-tuned Language Models
Figure 2 for ChatGPT v.s. Media Bias: A Comparative Study of GPT-3.5 and Fine-tuned Language Models
Figure 3 for ChatGPT v.s. Media Bias: A Comparative Study of GPT-3.5 and Fine-tuned Language Models
Viaarxiv icon

FFSTC: Fongbe to French Speech Translation Corpus

Mar 08, 2024
D. Fortune Kponou, Frejus A. A. Laleye, Eugene C. Ezin

Figure 1 for FFSTC: Fongbe to French Speech Translation Corpus
Figure 2 for FFSTC: Fongbe to French Speech Translation Corpus
Figure 3 for FFSTC: Fongbe to French Speech Translation Corpus
Viaarxiv icon

HAM-TTS: Hierarchical Acoustic Modeling for Token-Based Zero-Shot Text-to-Speech with Model and Data Scaling

Mar 09, 2024
Chunhui Wang, Chang Zeng, Bowen Zhang, Ziyang Ma, Yefan Zhu, Zifeng Cai, Jian Zhao, Zhonglin Jiang, Yong Chen

Figure 1 for HAM-TTS: Hierarchical Acoustic Modeling for Token-Based Zero-Shot Text-to-Speech with Model and Data Scaling
Figure 2 for HAM-TTS: Hierarchical Acoustic Modeling for Token-Based Zero-Shot Text-to-Speech with Model and Data Scaling
Figure 3 for HAM-TTS: Hierarchical Acoustic Modeling for Token-Based Zero-Shot Text-to-Speech with Model and Data Scaling
Figure 4 for HAM-TTS: Hierarchical Acoustic Modeling for Token-Based Zero-Shot Text-to-Speech with Model and Data Scaling
Viaarxiv icon

Project MOSLA: Recording Every Moment of Second Language Acquisition

Mar 26, 2024
Masato Hagiwara, Joshua Tanner

Viaarxiv icon

A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement

Mar 03, 2024
Ravi Shankar, Ke Tan, Buye Xu, Anurag Kumar

Figure 1 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Figure 2 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Figure 3 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Figure 4 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Viaarxiv icon

NLP for Counterspeech against Hate: A Survey and How-To Guide

Mar 29, 2024
Helena Bonaldi, Yi-Ling Chung, Gavin Abercrombie, Marco Guerini

Figure 1 for NLP for Counterspeech against Hate: A Survey and How-To Guide
Figure 2 for NLP for Counterspeech against Hate: A Survey and How-To Guide
Figure 3 for NLP for Counterspeech against Hate: A Survey and How-To Guide
Figure 4 for NLP for Counterspeech against Hate: A Survey and How-To Guide
Viaarxiv icon

Skipformer: A Skip-and-Recover Strategy for Efficient Speech Recognition

Mar 13, 2024
Wenjing Zhu, Sining Sun, Changhao Shan, Peng Fan, Qing Yang

Figure 1 for Skipformer: A Skip-and-Recover Strategy for Efficient Speech Recognition
Figure 2 for Skipformer: A Skip-and-Recover Strategy for Efficient Speech Recognition
Figure 3 for Skipformer: A Skip-and-Recover Strategy for Efficient Speech Recognition
Figure 4 for Skipformer: A Skip-and-Recover Strategy for Efficient Speech Recognition
Viaarxiv icon

Extending Multilingual Speech Synthesis to 100+ Languages without Transcribed Data

Feb 29, 2024
Takaaki Saeki, Gary Wang, Nobuyuki Morioka, Isaac Elias, Kyle Kastner, Andrew Rosenberg, Bhuvana Ramabhadran, Heiga Zen, Françoise Beaufays, Hadar Shemtov

Viaarxiv icon

Decode Neural signal as Speech

Add code
Bookmark button
Alert button
Mar 04, 2024
Yiqian Yang, Yiqun Duan, Qiang Zhang, Renjing Xu, Hui Xiong

Figure 1 for Decode Neural signal as Speech
Figure 2 for Decode Neural signal as Speech
Figure 3 for Decode Neural signal as Speech
Figure 4 for Decode Neural signal as Speech
Viaarxiv icon

A Cross-Modal Approach to Silent Speech with LLM-Enhanced Recognition

Add code
Bookmark button
Alert button
Mar 02, 2024
Tyler Benster, Guy Wilson, Reshef Elisha, Francis R Willett, Shaul Druckmann

Figure 1 for A Cross-Modal Approach to Silent Speech with LLM-Enhanced Recognition
Figure 2 for A Cross-Modal Approach to Silent Speech with LLM-Enhanced Recognition
Figure 3 for A Cross-Modal Approach to Silent Speech with LLM-Enhanced Recognition
Figure 4 for A Cross-Modal Approach to Silent Speech with LLM-Enhanced Recognition
Viaarxiv icon