Get our free extension to see links to code for papers anywhere online!

Chrome logo Add to Chrome

Firefox logo Add to Firefox

"speech": models, code, and papers

Twitter Dataset on the Russo-Ukrainian War

Apr 07, 2022
Alexander Shevtsov, Christos Tzagkarakis, Despoina Antonakaki, Polyvios Pratikakis, Sotiris Ioannidis

On 24 February 2022, Russia invaded Ukraine, also known now as Russo-Ukrainian War. We have initiated an ongoing dataset acquisition from Twitter API. Until the day this paper was written the dataset has reached the amount of 57.3 million tweets, originating from 7.7 million users. We apply an initial volume and sentiment analysis, while the dataset can be used to further exploratory investigation towards topic analysis, hate speech, propaganda recognition, or even show potential malicious entities like botnets.

  Access Paper or Ask Questions

Seq-2-Seq based Refinement of ASR Output for Spoken Name Capture

Mar 29, 2022
Karan Singla, Shahab Jalalvand, Yeon-Jun Kim, Ryan Price, Daniel Pressel, Srinivas Bangalore

Person name capture from human speech is a difficult task in human-machine conversations. In this paper, we propose a novel approach to capture the person names from the caller utterances in response to the prompt "say and spell your first/last name". Inspired from work on spell correction, disfluency removal and text normalization, we propose a lightweight Seq-2-Seq system which generates a name spell from a varying user input. Our proposed method outperforms the strong baseline which is based on LM-driven rule-based approach.

* Under review at InterSpeech 2022 

  Access Paper or Ask Questions

UzBERT: pretraining a BERT model for Uzbek

Aug 22, 2021
B. Mansurov, A. Mansurov

Pretrained language models based on the Transformer architecture have achieved state-of-the-art results in various natural language processing tasks such as part-of-speech tagging, named entity recognition, and question answering. However, no such monolingual model for the Uzbek language is publicly available. In this paper, we introduce UzBERT, a pretrained Uzbek language model based on the BERT architecture. Our model greatly outperforms multilingual BERT on masked language model accuracy. We make the model publicly available under the MIT open-source license.

* 9 pages, 1 table 

  Access Paper or Ask Questions

NNVLP: A Neural Network-Based Vietnamese Language Processing Toolkit

Oct 19, 2017
Thai-Hoang Pham, Xuan-Khoai Pham, Tuan-Anh Nguyen, Phuong Le-Hong

This paper demonstrates neural network-based toolkit namely NNVLP for essential Vietnamese language processing tasks including part-of-speech (POS) tagging, chunking, named entity recognition (NER). Our toolkit is a combination of bidirectional Long Short-Term Memory (Bi-LSTM), Convolutional Neural Network (CNN), Conditional Random Field (CRF), using pre-trained word embeddings as input, which achieves state-of-the-art results on these three tasks. We provide both API and web demo for this toolkit.

* 4 pages, 5 figures, 6 tables, accepted to IJCNLP 2017 

  Access Paper or Ask Questions

Incorporating Language Level Information into Acoustic Models

Dec 14, 2016
Peidong Wang, Deliang Wang

This paper proposed a class of novel Deep Recurrent Neural Networks which can incorporate language-level information into acoustic models. For simplicity, we named these networks Recurrent Deep Language Networks (RDLNs). Multiple variants of RDLNs were considered, including two kinds of context information, two methods to process the context, and two methods to incorporate the language-level information. RDLNs provided possible methods to fine-tune the whole Automatic Speech Recognition (ASR) system in the acoustic modeling process.

  Access Paper or Ask Questions

The First Parallel Multilingual Corpus of Persian: Toward a Persian BLARK

Apr 17, 2014
Behrang Qasemizadeh, Saeed Rahimi, Behrooz Mahmoodi Bakhtiari

In this article, we have introduced the first parallel corpus of Persian with more than 10 other European languages. This article describes primary steps toward preparing a Basic Language Resources Kit (BLARK) for Persian. Up to now, we have proposed morphosyntactic specification of Persian based on EAGLE/MULTEXT guidelines and specific resources of MULTEXT-East. The article introduces Persian Language, with emphasis on its orthography and morphosyntactic features, then a new Part-of-Speech categorization and orthography for Persian in digital environments is proposed. Finally, the corpus and related statistic will be analyzed.

  Access Paper or Ask Questions

Evaluation of a Grammar of French Determiners

Nov 21, 2007
Eric Laporte

Existing syntactic grammars of natural languages, even with a far from complete coverage, are complex objects. Assessments of the quality of parts of such grammars are useful for the validation of their construction. We evaluated the quality of a grammar of French determiners that takes the form of a recursive transition network. The result of the application of this local grammar gives deeper syntactic information than chunking or information available in treebanks. We performed the evaluation by comparison with a corpus independently annotated with information on determiners. We obtained 86% precision and 92% recall on text not tagged for parts of speech.

* Dans Annals of the 27th Congress of the Brazilian Society of Computation - Evaluation of a Grammar of French Determiners, Rio de Janeiro : Br\'esil (2007) 
* 10 pages 

  Access Paper or Ask Questions

Semantic robust parsing for noun extraction from natural language queries

Sep 02, 1999
Afzal Ballim, Vincenzo Pallotta

This paper describes how robust parsing techniques can be fruitful applied for building a query generation module which is part of a pipelined NLP architecture aimed at process natural language queries in a restricted domain. We want to show that semantic robustness represents a key issue in those NLP systems where it is more likely to have partial and ill-formed utterances due to various factors (e.g. noisy environments, low quality of speech recognition modules, etc...) and where it is necessary to succeed, even if partially, in extracting some meaningful information.

* Proceedings of WPDI'99 (Workshop on Procedures in Discourse Interpretation),1999, Iasi - Romania 

  Access Paper or Ask Questions

A Maximum Entropy Approach to Identifying Sentence Boundaries

Apr 09, 1997
Jeffrey C. Reynar, Adwait Ratnaparkhi

We present a trainable model for identifying sentence boundaries in raw text. Given a corpus annotated with sentence boundaries, our model learns to classify each occurrence of ., ?, and ! as either a valid or invalid sentence boundary. The training procedure requires no hand-crafted rules, lexica, part-of-speech tags, or domain-specific information. The model can therefore be trained easily on any genre of English, and should be trainable on any other Roman-alphabet language. Performance is comparable to or better than the performance of similar systems, but we emphasize the simplicity of retraining for new domains.

* Proceedings of the 5th ANLP Conference, 1997 
* 4 pages, uses aclap.sty and covingtn.sty 

  Access Paper or Ask Questions

Interleaving Syntax and Semantics in an Efficient Bottom-Up Parser

Jul 05, 1994
John Dowding, Robert Moore, Francois Andry, Douglas Moran

We describe an efficient bottom-up parser that interleaves syntactic and semantic structure building. Two techniques are presented for reducing search by reducing local ambiguity: Limited left-context constraints are used to reduce local syntactic ambiguity, and deferred sortal-constraint application is used to reduce local semantic ambiguity. We experimentally evaluate these techniques, and show dramatic reductions in both number of chart-edges and total parsing time. The robust processing capabilities of the parser are demonstrated in its use in improving the accuracy of a speech recognizer.

* 32nd ACL, Las Cruces, New Mexico, June 1994, pp. 110-116 
* 8 pages, postscript 

  Access Paper or Ask Questions