Alert button

"Text": models, code, and papers
Alert button

PainPoints: A Framework for Language-based Detection of Chronic Pain and Expert-Collaborative Text-Summarization

Sep 14, 2022
Shreyas Fadnavis, Amit Dhurandhar, Raquel Norel, Jenna M Reinen, Carla Agurto, Erica Secchettin, Vittorio Schweiger, Giovanni Perini, Guillermo Cecchi

Figure 1 for PainPoints: A Framework for Language-based Detection of Chronic Pain and Expert-Collaborative Text-Summarization
Figure 2 for PainPoints: A Framework for Language-based Detection of Chronic Pain and Expert-Collaborative Text-Summarization
Figure 3 for PainPoints: A Framework for Language-based Detection of Chronic Pain and Expert-Collaborative Text-Summarization
Figure 4 for PainPoints: A Framework for Language-based Detection of Chronic Pain and Expert-Collaborative Text-Summarization
Viaarxiv icon

Discrete Cross-Modal Alignment Enables Zero-Shot Speech Translation

Oct 18, 2022
Chen Wang, Yuchen Liu, Boxing Chen, Jiajun Zhang, Wei Luo, Zhongqiang Huang, Chengqing Zong

Figure 1 for Discrete Cross-Modal Alignment Enables Zero-Shot Speech Translation
Figure 2 for Discrete Cross-Modal Alignment Enables Zero-Shot Speech Translation
Figure 3 for Discrete Cross-Modal Alignment Enables Zero-Shot Speech Translation
Figure 4 for Discrete Cross-Modal Alignment Enables Zero-Shot Speech Translation
Viaarxiv icon

A Unified Understanding of Deep NLP Models for Text Classification

Jun 19, 2022
Zhen Li, Xiting Wang, Weikai Yang, Jing Wu, Zhengyan Zhang, Zhiyuan Liu, Maosong Sun, Hui Zhang, Shixia Liu

Figure 1 for A Unified Understanding of Deep NLP Models for Text Classification
Figure 2 for A Unified Understanding of Deep NLP Models for Text Classification
Figure 3 for A Unified Understanding of Deep NLP Models for Text Classification
Figure 4 for A Unified Understanding of Deep NLP Models for Text Classification
Viaarxiv icon

Extremely Low-light Image Enhancement with Scene Text Restoration

Apr 01, 2022
Pohao Hsu, Che-Tsung Lin, Chun Chet Ng, Jie-Long Kew, Mei Yih Tan, Shang-Hong Lai, Chee Seng Chan, Christopher Zach

Figure 1 for Extremely Low-light Image Enhancement with Scene Text Restoration
Figure 2 for Extremely Low-light Image Enhancement with Scene Text Restoration
Figure 3 for Extremely Low-light Image Enhancement with Scene Text Restoration
Figure 4 for Extremely Low-light Image Enhancement with Scene Text Restoration
Viaarxiv icon

TGAVC: Improving Autoencoder Voice Conversion with Text-Guided and Adversarial Training

Aug 08, 2022
Huaizhen Tang, Xulong Zhang, Jianzong Wang, Ning Cheng, Zhen Zeng, Edward Xiao, Jing Xiao

Figure 1 for TGAVC: Improving Autoencoder Voice Conversion with Text-Guided and Adversarial Training
Figure 2 for TGAVC: Improving Autoencoder Voice Conversion with Text-Guided and Adversarial Training
Figure 3 for TGAVC: Improving Autoencoder Voice Conversion with Text-Guided and Adversarial Training
Figure 4 for TGAVC: Improving Autoencoder Voice Conversion with Text-Guided and Adversarial Training
Viaarxiv icon

Book Cover Synthesis from the Summary

Nov 03, 2022
Emdadul Haque, Md. Faraz Kabir Khan, Mohammad Imrul Jubair, Jarin Anjum, Abrar Zahir Niloy

Figure 1 for Book Cover Synthesis from the Summary
Figure 2 for Book Cover Synthesis from the Summary
Figure 3 for Book Cover Synthesis from the Summary
Figure 4 for Book Cover Synthesis from the Summary
Viaarxiv icon

Language models are better than humans at next-token prediction

Dec 21, 2022
Buck Shlegeris, Fabien Roger, Lawrence Chan, Euan McLean

Figure 1 for Language models are better than humans at next-token prediction
Figure 2 for Language models are better than humans at next-token prediction
Figure 3 for Language models are better than humans at next-token prediction
Viaarxiv icon

Zero-Shot Style Transfer for Gesture Animation driven by Text and Speech using Adversarial Disentanglement of Multimodal Style Encoding

Aug 03, 2022
Mireille Fares, Michele Grimaldi, Catherine Pelachaud, Nicolas Obin

Figure 1 for Zero-Shot Style Transfer for Gesture Animation driven by Text and Speech using Adversarial Disentanglement of Multimodal Style Encoding
Figure 2 for Zero-Shot Style Transfer for Gesture Animation driven by Text and Speech using Adversarial Disentanglement of Multimodal Style Encoding
Figure 3 for Zero-Shot Style Transfer for Gesture Animation driven by Text and Speech using Adversarial Disentanglement of Multimodal Style Encoding
Figure 4 for Zero-Shot Style Transfer for Gesture Animation driven by Text and Speech using Adversarial Disentanglement of Multimodal Style Encoding
Viaarxiv icon

A Large-Scale Outdoor Multi-modal Dataset and Benchmark for Novel View Synthesis and Implicit Scene Reconstruction

Jan 17, 2023
Chongshan Lu, Fukun Yin, Xin Chen, Tao Chen, Gang YU, Jiayuan Fan

Figure 1 for A Large-Scale Outdoor Multi-modal Dataset and Benchmark for Novel View Synthesis and Implicit Scene Reconstruction
Figure 2 for A Large-Scale Outdoor Multi-modal Dataset and Benchmark for Novel View Synthesis and Implicit Scene Reconstruction
Figure 3 for A Large-Scale Outdoor Multi-modal Dataset and Benchmark for Novel View Synthesis and Implicit Scene Reconstruction
Figure 4 for A Large-Scale Outdoor Multi-modal Dataset and Benchmark for Novel View Synthesis and Implicit Scene Reconstruction
Viaarxiv icon

Effectiveness of Mining Audio and Text Pairs from Public Data for Improving ASR Systems for Low-Resource Languages

Aug 26, 2022
Kaushal Santosh Bhogale, Abhigyan Raman, Tahir Javed, Sumanth Doddapaneni, Anoop Kunchukuttan, Pratyush Kumar, Mitesh M. Khapra

Figure 1 for Effectiveness of Mining Audio and Text Pairs from Public Data for Improving ASR Systems for Low-Resource Languages
Figure 2 for Effectiveness of Mining Audio and Text Pairs from Public Data for Improving ASR Systems for Low-Resource Languages
Figure 3 for Effectiveness of Mining Audio and Text Pairs from Public Data for Improving ASR Systems for Low-Resource Languages
Figure 4 for Effectiveness of Mining Audio and Text Pairs from Public Data for Improving ASR Systems for Low-Resource Languages
Viaarxiv icon