Alert button
Picture for Zae Myung Kim

Zae Myung Kim

Alert button

An Analysis of Reader Engagement in Literary Fiction through Eye Tracking and Linguistic Features

Jun 06, 2023
Rose Neis, Karin de Langis, Zae Myung Kim, Dongyeop Kang

Figure 1 for An Analysis of Reader Engagement in Literary Fiction through Eye Tracking and Linguistic Features
Figure 2 for An Analysis of Reader Engagement in Literary Fiction through Eye Tracking and Linguistic Features
Figure 3 for An Analysis of Reader Engagement in Literary Fiction through Eye Tracking and Linguistic Features
Figure 4 for An Analysis of Reader Engagement in Literary Fiction through Eye Tracking and Linguistic Features

Capturing readers' engagement in fiction is a challenging but important aspect of narrative understanding. In this study, we collected 23 readers' reactions to 2 short stories through eye tracking, sentence-level annotations, and an overall engagement scale survey. We analyzed the significance of various qualities of the text in predicting how engaging a reader is likely to find it. As enjoyment of fiction is highly contextual, we also investigated individual differences in our data. Furthering our understanding of what captivates readers in fiction will help better inform models used in creative narrative generation and collaborative writing tools.

* 9 pages, 4 figures 
Viaarxiv icon

Diffusion Models in NLP: A Survey

May 24, 2023
Hao Zou, Zae Myung Kim, Dongyeop Kang

Figure 1 for Diffusion Models in NLP: A Survey
Figure 2 for Diffusion Models in NLP: A Survey
Figure 3 for Diffusion Models in NLP: A Survey
Figure 4 for Diffusion Models in NLP: A Survey

This survey paper provides a comprehensive review of the use of diffusion models in natural language processing (NLP). Diffusion models are a class of mathematical models that aim to capture the diffusion of information or signals across a network or manifold. In NLP, diffusion models have been used in a variety of applications, such as natural language generation, sentiment analysis, topic modeling, and machine translation. This paper discusses the different formulations of diffusion models used in NLP, their strengths and limitations, and their applications. We also perform a thorough comparison between diffusion models and alternative generative models, specifically highlighting the autoregressive (AR) models, while also examining how diverse architectures incorporate the Transformer in conjunction with diffusion models. Compared to AR models, diffusion models have significant advantages for parallel generation, text interpolation, token-level controls such as syntactic structures and semantic contents, and robustness. Exploring further permutations of integrating Transformers into diffusion models would be a valuable pursuit. Also, the development of multimodal diffusion models and large-scale diffusion language models with notable capabilities for few-shot learning would be important directions for the future advance of diffusion models in NLP.

Viaarxiv icon

"Is the Pope Catholic?" Applying Chain-of-Thought Reasoning to Understanding Conversational Implicatures

May 23, 2023
Zae Myung Kim, David E. Taylor, Dongyeop Kang

Figure 1 for "Is the Pope Catholic?" Applying Chain-of-Thought Reasoning to Understanding Conversational Implicatures
Figure 2 for "Is the Pope Catholic?" Applying Chain-of-Thought Reasoning to Understanding Conversational Implicatures

Conversational implicatures are pragmatic inferences that require listeners to deduce the intended meaning conveyed by a speaker from their explicit utterances. Although such inferential reasoning is fundamental to human communication, recent research indicates that large language models struggle to comprehend these implicatures as effectively as the average human. This paper demonstrates that by incorporating Grice's Four Maxims into the model through chain-of-thought prompting, we can significantly enhance its performance, surpassing even the average human performance on this task.

Viaarxiv icon

Improving Iterative Text Revision by Learning Where to Edit from Other Revision Tasks

Dec 02, 2022
Zae Myung Kim, Wanyu Du, Vipul Raheja, Dhruv Kumar, Dongyeop Kang

Figure 1 for Improving Iterative Text Revision by Learning Where to Edit from Other Revision Tasks
Figure 2 for Improving Iterative Text Revision by Learning Where to Edit from Other Revision Tasks
Figure 3 for Improving Iterative Text Revision by Learning Where to Edit from Other Revision Tasks
Figure 4 for Improving Iterative Text Revision by Learning Where to Edit from Other Revision Tasks

Iterative text revision improves text quality by fixing grammatical errors, rephrasing for better readability or contextual appropriateness, or reorganizing sentence structures throughout a document. Most recent research has focused on understanding and classifying different types of edits in the iterative revision process from human-written text instead of building accurate and robust systems for iterative text revision. In this work, we aim to build an end-to-end text revision system that can iteratively generate helpful edits by explicitly detecting editable spans (where-to-edit) with their corresponding edit intents and then instructing a revision model to revise the detected edit spans. Leveraging datasets from other related text editing NLP tasks, combined with the specification of editable spans, leads our system to more accurately model the process of iterative text refinement, as evidenced by empirical results and human evaluations. Our system significantly outperforms previous baselines on our text revision tasks and other standard text revision tasks, including grammatical error correction, text simplification, sentence fusion, and style transfer. Through extensive qualitative and quantitative analysis, we make vital connections between edit intentions and writing quality, and better computational modeling of iterative text revisions.

* 14 pages, accepted at EMNLP 2022 conference as a full paper 
Viaarxiv icon

Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision

Apr 07, 2022
Wanyu Du, Zae Myung Kim, Vipul Raheja, Dhruv Kumar, Dongyeop Kang

Figure 1 for Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision
Figure 2 for Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision
Figure 3 for Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision
Figure 4 for Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision

Revision is an essential part of the human writing process. It tends to be strategic, adaptive, and, more importantly, iterative in nature. Despite the success of large language models on text revision tasks, they are limited to non-iterative, one-shot revisions. Examining and evaluating the capability of large language models for making continuous revisions and collaborating with human writers is a critical step towards building effective writing assistants. In this work, we present a human-in-the-loop iterative text revision system, Read, Revise, Repeat (R3), which aims at achieving high quality text revisions with minimal human efforts by reading model-generated revisions and user feedbacks, revising documents, and repeating human-machine interactions. In R3, a text revision model provides text editing suggestions for human writers, who can accept or reject the suggested edits. The accepted edits are then incorporated into the model for the next iteration of document revision. Writers can therefore revise documents iteratively by interacting with the system and simply accepting/rejecting its suggested edits until the text revision model stops making further revisions or reaches a predefined maximum number of revisions. Empirical experiments show that R3 can generate revisions with comparable acceptance rate to human writers at early revision depths, and the human-machine interaction can get higher quality revisions with fewer iterations and edits. The collected human-model interaction dataset and system code are available at \url{https://github.com/vipulraheja/IteraTeR}. Our system demonstration is available at \url{https://youtu.be/lK08tIpEoaE}.

* Accepted by The First Workshop on Intelligent and Interactive Writing Assistants at ACL2022 
Viaarxiv icon

Understanding Iterative Revision from Human-Written Text

Mar 16, 2022
Wanyu Du, Vipul Raheja, Dhruv Kumar, Zae Myung Kim, Melissa Lopez, Dongyeop Kang

Figure 1 for Understanding Iterative Revision from Human-Written Text
Figure 2 for Understanding Iterative Revision from Human-Written Text
Figure 3 for Understanding Iterative Revision from Human-Written Text
Figure 4 for Understanding Iterative Revision from Human-Written Text

Writing is, by nature, a strategic, adaptive, and more importantly, an iterative process. A crucial part of writing is editing and revising the text. Previous works on text revision have focused on defining edit intention taxonomies within a single domain or developing computational models with a single level of edit granularity, such as sentence-level edits, which differ from human's revision cycles. This work describes IteraTeR: the first large-scale, multi-domain, edit-intention annotated corpus of iteratively revised text. In particular, IteraTeR is collected based on a new framework to comprehensively model the iterative text revisions that generalize to various domains of formal writing, edit intentions, revision depths, and granularities. When we incorporate our annotated edit intentions, both generative and edit-based text revision models significantly improve automatic evaluations. Through our work, we better understand the text revision process, making vital connections between edit intentions and writing quality, enabling the creation of diverse corpora to support computational modeling of iterative text revisions.

* To appear in ACL2022 
Viaarxiv icon

Do Multilingual Neural Machine Translation Models Contain Language Pair Specific Attention Heads?

May 31, 2021
Zae Myung Kim, Laurent Besacier, Vassilina Nikoulina, Didier Schwab

Figure 1 for Do Multilingual Neural Machine Translation Models Contain Language Pair Specific Attention Heads?
Figure 2 for Do Multilingual Neural Machine Translation Models Contain Language Pair Specific Attention Heads?
Figure 3 for Do Multilingual Neural Machine Translation Models Contain Language Pair Specific Attention Heads?

Recent studies on the analysis of the multilingual representations focus on identifying whether there is an emergence of language-independent representations, or whether a multilingual model partitions its weights among different languages. While most of such work has been conducted in a "black-box" manner, this paper aims to analyze individual components of a multilingual neural translation (NMT) model. In particular, we look at the encoder self-attention and encoder-decoder attention heads (in a many-to-one NMT model) that are more specific to the translation of a certain language pair than others by (1) employing metrics that quantify some aspects of the attention weights such as "variance" or "confidence", and (2) systematically ranking the importance of attention heads with respect to translation quality. Experimental results show that surprisingly, the set of most important attention heads are very similar across the language pairs and that it is possible to remove nearly one-third of the less important heads without hurting the translation quality greatly.

* 10 pages, accepted at Findings of ACL 2021 (short) 
Viaarxiv icon

A Multilingual Neural Machine Translation Model for Biomedical Data

Aug 06, 2020
Alexandre Bérard, Zae Myung Kim, Vassilina Nikoulina, Eunjeong L. Park, Matthias Gallé

Figure 1 for A Multilingual Neural Machine Translation Model for Biomedical Data
Figure 2 for A Multilingual Neural Machine Translation Model for Biomedical Data
Figure 3 for A Multilingual Neural Machine Translation Model for Biomedical Data

We release a multilingual neural machine translation model, which can be used to translate text in the biomedical domain. The model can translate from 5 languages (French, German, Italian, Korean and Spanish) into English. It is trained with large amounts of generic and biomedical data, using domain tags. Our benchmarks show that it performs near state-of-the-art both on news (generic domain) and biomedical test sets, and that it outperforms the existing publicly released models. We believe that this release will help the large-scale multilingual analysis of the digital content of the COVID-19 crisis and of its effects on society, economy, and healthcare policies. We also release a test set of biomedical text for Korean-English. It consists of 758 sentences from official guidelines and recent papers, all about COVID-19.

* https://github.com/naver/covid19-nmt 
Viaarxiv icon