Alert button
Picture for Nikolay Babakov

Nikolay Babakov

Alert button

Don't lose the message while paraphrasing: A study on content preserving style transfer

Aug 17, 2023
Nikolay Babakov, David Dale, Ilya Gusev, Irina Krotova, Alexander Panchenko

Text style transfer techniques are gaining popularity in natural language processing allowing paraphrasing text in the required form: from toxic to neural, from formal to informal, from old to the modern English language, etc. Solving the task is not sufficient to generate some neural/informal/modern text, but it is important to preserve the original content unchanged. This requirement becomes even more critical in some applications such as style transfer of goal-oriented dialogues where the factual information shall be kept to preserve the original message, e.g. ordering a certain type of pizza to a certain address at a certain time. The aspect of content preservation is critical for real-world applications of style transfer studies, but it has received little attention. To bridge this gap we perform a comparison of various style transfer models on the example of the formality transfer domain. To perform a study of the content preservation abilities of various style transfer methods we create a parallel dataset of formal vs. informal task-oriented dialogues. The key difference between our dataset and the existing ones like GYAFC [17] is the presence of goal-oriented dialogues with predefined semantic slots essential to be kept during paraphrasing, e.g. named entities. This additional annotation allowed us to conduct a precise comparative study of several state-of-the-art techniques for style transfer. Another result of our study is a modification of the unsupervised method LEWIS [19] which yields a substantial improvement over the original method and all evaluated baselines on the proposed task.

* Published at the NLDB 2023 conference 
Viaarxiv icon

Error syntax aware augmentation of feedback comment generation dataset

Dec 29, 2022
Nikolay Babakov, Maria Lysyuk, Alexander Shvets, Lilya Kazakova, Alexander Panchenko

Figure 1 for Error syntax aware augmentation of feedback comment generation dataset
Figure 2 for Error syntax aware augmentation of feedback comment generation dataset
Figure 3 for Error syntax aware augmentation of feedback comment generation dataset
Figure 4 for Error syntax aware augmentation of feedback comment generation dataset

This paper presents a solution to the GenChal 2022 shared task dedicated to feedback comment generation for writing learning. In terms of this task given a text with an error and a span of the error, a system generates an explanatory note that helps the writer (language learner) to improve their writing skills. Our solution is based on fine-tuning the T5 model on the initial dataset augmented according to syntactical dependencies of the words located within indicated error span. The solution of our team "nigula" obtained second place according to manual evaluation by the organizers.

* Accepted to publication on INLG 2023 
Viaarxiv icon

Studying the role of named entities for content preservation in text style transfer

Jun 20, 2022
Nikolay Babakov, David Dale, Varvara Logacheva, Irina Krotova, Alexander Panchenko

Text style transfer techniques are gaining popularity in Natural Language Processing, finding various applications such as text detoxification, sentiment, or formality transfer. However, the majority of the existing approaches were tested on such domains as online communications on public platforms, music, or entertainment yet none of them were applied to the domains which are typical for task-oriented production systems, such as personal plans arrangements (e.g. booking of flights or reserving a table in a restaurant). We fill this gap by studying formality transfer in this domain. We noted that the texts in this domain are full of named entities, which are very important for keeping the original sense of the text. Indeed, if for example, someone communicates the destination city of a flight it must not be altered. Thus, we concentrate on the role of named entities in content preservation for formality text style transfer. We collect a new dataset for the evaluation of content similarity measures in text style transfer. It is taken from a corpus of task-oriented dialogues and contains many important entities related to realistic requests that make this dataset particularly useful for testing style transfer models before using them in production. Besides, we perform an error analysis of a pre-trained formality transfer model and introduce a simple technique to use information about named entities to enhance the performance of baseline content similarity measures used in text style transfer.

* Natural Language Processing and Information Systems. NLDB 2022. Lecture Notes in Computer Science, vol 13286. Springer, Cham, p.437--448  
Viaarxiv icon

Beyond Plain Toxic: Detection of Inappropriate Statements on Flammable Topics for the Russian Language

Mar 04, 2022
Nikolay Babakov, Varvara Logacheva, Alexander Panchenko

Figure 1 for Beyond Plain Toxic: Detection of Inappropriate Statements on Flammable Topics for the Russian Language
Figure 2 for Beyond Plain Toxic: Detection of Inappropriate Statements on Flammable Topics for the Russian Language
Figure 3 for Beyond Plain Toxic: Detection of Inappropriate Statements on Flammable Topics for the Russian Language
Figure 4 for Beyond Plain Toxic: Detection of Inappropriate Statements on Flammable Topics for the Russian Language

Toxicity on the Internet, such as hate speech, offenses towards particular users or groups of people, or the use of obscene words, is an acknowledged problem. However, there also exist other types of inappropriate messages which are usually not viewed as toxic, e.g. as they do not contain explicit offences. Such messages can contain covered toxicity or generalizations, incite harmful actions (crime, suicide, drug use), provoke "heated" discussions. Such messages are often related to particular sensitive topics, e.g. on politics, sexual minorities, social injustice which more often than other topics, e.g. cars or computing, yield toxic emotional reactions. At the same time, clearly not all messages within such flammable topics are inappropriate. Towards this end, in this work, we present two text collections labelled according to binary notion of inapropriateness and a multinomial notion of sensitive topic. Assuming that the notion of inappropriateness is common among people of the same culture, we base our approach on human intuitive understanding of what is not acceptable and harmful. To objectivise the notion of inappropriateness, we define it in a data-driven way though crowdsourcing. Namely we run a large-scale annotation study asking workers if a given chatbot textual statement could harm reputation of a company created it. Acceptably high values of inter-annotator agreement suggest that the notion of inappropriateness exists and can be uniformly understood by different people. To define the notion of sensitive topics in an objective way we use on guidelines suggested commonly by specialists of legal and PR department of a large public company as potentially harmful.

* arXiv admin note: text overlap with arXiv:2103.05345 
Viaarxiv icon

Detecting Inappropriate Messages on Sensitive Topics that Could Harm a Company's Reputation

Mar 09, 2021
Nikolay Babakov, Varvara Logacheva, Olga Kozlova, Nikita Semenov, Alexander Panchenko

Figure 1 for Detecting Inappropriate Messages on Sensitive Topics that Could Harm a Company's Reputation
Figure 2 for Detecting Inappropriate Messages on Sensitive Topics that Could Harm a Company's Reputation
Figure 3 for Detecting Inappropriate Messages on Sensitive Topics that Could Harm a Company's Reputation
Figure 4 for Detecting Inappropriate Messages on Sensitive Topics that Could Harm a Company's Reputation

Not all topics are equally "flammable" in terms of toxicity: a calm discussion of turtles or fishing less often fuels inappropriate toxic dialogues than a discussion of politics or sexual minorities. We define a set of sensitive topics that can yield inappropriate and toxic messages and describe the methodology of collecting and labeling a dataset for appropriateness. While toxicity in user-generated data is well-studied, we aim at defining a more fine-grained notion of inappropriateness. The core of inappropriateness is that it can harm the reputation of a speaker. This is different from toxicity in two respects: (i) inappropriateness is topic-related, and (ii) inappropriate message is not toxic but still unacceptable. We collect and release two datasets for Russian: a topic-labeled dataset and an appropriateness-labeled dataset. We also release pre-trained classification models trained on this data.

* Accepted to the Balto-Slavic NLP workshop 2021 co-located with EACL-2021 
Viaarxiv icon