Alert button
Picture for Chenghua Lin

Chenghua Lin

Alert button

Affective Decoding for Empathetic Response Generation

Aug 18, 2021
Chengkun Zheng, Guanyi Chen, Chenghua Lin, Ruizhe Li, Zhigang Chen

Figure 1 for Affective Decoding for Empathetic Response Generation
Figure 2 for Affective Decoding for Empathetic Response Generation
Figure 3 for Affective Decoding for Empathetic Response Generation
Figure 4 for Affective Decoding for Empathetic Response Generation

Understanding speaker's feelings and producing appropriate responses with emotion connection is a key communicative skill for empathetic dialogue systems. In this paper, we propose a simple technique called Affective Decoding for empathetic response generation. Our method can effectively incorporate emotion signals during each decoding step, and can additionally be augmented with an auxiliary dual emotion encoder, which learns separate embeddings for the speaker and listener given the emotion base of the dialogue. Extensive empirical studies show that our models are perceived to be more empathetic by human evaluations, in comparison to several strong mainstream methods for empathetic responding.

* Long paper accepted to INLG 2021 
Viaarxiv icon

Guiding the Growth: Difficulty-Controllable Question Generation through Step-by-Step Rewriting

May 25, 2021
Yi Cheng, Siyao Li, Bang Liu, Ruihui Zhao, Sujian Li, Chenghua Lin, Yefeng Zheng

Figure 1 for Guiding the Growth: Difficulty-Controllable Question Generation through Step-by-Step Rewriting
Figure 2 for Guiding the Growth: Difficulty-Controllable Question Generation through Step-by-Step Rewriting
Figure 3 for Guiding the Growth: Difficulty-Controllable Question Generation through Step-by-Step Rewriting
Figure 4 for Guiding the Growth: Difficulty-Controllable Question Generation through Step-by-Step Rewriting

This paper explores the task of Difficulty-Controllable Question Generation (DCQG), which aims at generating questions with required difficulty levels. Previous research on this task mainly defines the difficulty of a question as whether it can be correctly answered by a Question Answering (QA) system, lacking interpretability and controllability. In our work, we redefine question difficulty as the number of inference steps required to answer it and argue that Question Generation (QG) systems should have stronger control over the logic of generated questions. To this end, we propose a novel framework that progressively increases question difficulty through step-by-step rewriting under the guidance of an extracted reasoning chain. A dataset is automatically constructed to facilitate the research, on which extensive experiments are conducted to test the performance of our method.

* Accepted by ACL 2021 (long paper) 
Viaarxiv icon

Highly Efficient Knowledge Graph Embedding Learning with Orthogonal Procrustes Analysis

Apr 17, 2021
Xutan Peng, Guanyi Chen, Chenghua Lin, Mark Stevenson

Figure 1 for Highly Efficient Knowledge Graph Embedding Learning with Orthogonal Procrustes Analysis
Figure 2 for Highly Efficient Knowledge Graph Embedding Learning with Orthogonal Procrustes Analysis
Figure 3 for Highly Efficient Knowledge Graph Embedding Learning with Orthogonal Procrustes Analysis
Figure 4 for Highly Efficient Knowledge Graph Embedding Learning with Orthogonal Procrustes Analysis

Knowledge Graph Embeddings (KGEs) have been intensively explored in recent years due to their promise for a wide range of applications. However, existing studies focus on improving the final model performance without acknowledging the computational cost of the proposed approaches, in terms of execution time and environmental impact. This paper proposes a simple yet effective KGE framework which can reduce the training time and carbon footprint by orders of magnitudes compared with state-of-the-art approaches, while producing competitive performance. We highlight three technical innovations: full batch learning via relational matrices, closed-form Orthogonal Procrustes Analysis for KGEs, and non-negative-sampling training. In addition, as the first KGE method whose entity embeddings also store full relation information, our trained models encode rich semantics and are highly interpretable. Comprehensive experiments and ablation studies involving 13 strong baselines and two standard datasets verify the effectiveness and efficiency of our algorithm.

* To appear at NAACL 2021 
Viaarxiv icon

Cross-Lingual Word Embedding Refinement by $\ell_{1}$ Norm Optimisation

Apr 11, 2021
Xutan Peng, Chenghua Lin, Mark Stevenson

Figure 1 for Cross-Lingual Word Embedding Refinement by $\ell_{1}$ Norm Optimisation
Figure 2 for Cross-Lingual Word Embedding Refinement by $\ell_{1}$ Norm Optimisation
Figure 3 for Cross-Lingual Word Embedding Refinement by $\ell_{1}$ Norm Optimisation
Figure 4 for Cross-Lingual Word Embedding Refinement by $\ell_{1}$ Norm Optimisation

Cross-Lingual Word Embeddings (CLWEs) encode words from two or more languages in a shared high-dimensional space in which vectors representing words with similar meaning (regardless of language) are closely located. Existing methods for building high-quality CLWEs learn mappings that minimise the $\ell_{2}$ norm loss function. However, this optimisation objective has been demonstrated to be sensitive to outliers. Based on the more robust Manhattan norm (aka. $\ell_{1}$ norm) goodness-of-fit criterion, this paper proposes a simple post-processing step to improve CLWEs. An advantage of this approach is that it is fully agnostic to the training process of the original CLWEs and can therefore be applied widely. Extensive experiments are performed involving ten diverse languages and embeddings trained on different corpora. Evaluation results based on bilingual lexicon induction and cross-lingual transfer for natural language inference tasks show that the $\ell_{1}$ refinement substantially outperforms four state-of-the-art baselines in both supervised and unsupervised settings. It is therefore recommended that this strategy be adopted as a standard for CLWE methods.

* To appear at NAACL 2021 
Viaarxiv icon

Interpreting Verbal Metaphors by Paraphrasing

Apr 07, 2021
Rui Mao, Chenghua Lin, Frank Guerin

Figure 1 for Interpreting Verbal Metaphors by Paraphrasing
Figure 2 for Interpreting Verbal Metaphors by Paraphrasing
Figure 3 for Interpreting Verbal Metaphors by Paraphrasing

Metaphorical expressions are difficult linguistic phenomena, challenging diverse Natural Language Processing tasks. Previous works showed that paraphrasing a metaphor as its literal counterpart can help machines better process metaphors on downstream tasks. In this paper, we interpret metaphors with BERT and WordNet hypernyms and synonyms in an unsupervised manner, showing that our method significantly outperforms the state-of-the-art baseline. We also demonstrate that our method can help a machine translation system improve its accuracy in translating English metaphors to 8 target languages.

Viaarxiv icon

Combining Pre-trained Word Embeddings and Linguistic Features for Sequential Metaphor Identification

Apr 07, 2021
Rui Mao, Chenghua Lin, Frank Guerin

Figure 1 for Combining Pre-trained Word Embeddings and Linguistic Features for Sequential Metaphor Identification
Figure 2 for Combining Pre-trained Word Embeddings and Linguistic Features for Sequential Metaphor Identification
Figure 3 for Combining Pre-trained Word Embeddings and Linguistic Features for Sequential Metaphor Identification
Figure 4 for Combining Pre-trained Word Embeddings and Linguistic Features for Sequential Metaphor Identification

We tackle the problem of identifying metaphors in text, treated as a sequence tagging task. The pre-trained word embeddings GloVe, ELMo and BERT have individually shown good performance on sequential metaphor identification. These embeddings are generated by different models, training targets and corpora, thus encoding different semantic and syntactic information. We show that leveraging GloVe, ELMo and feature-based BERT based on a multi-channel CNN and a Bidirectional LSTM model can significantly outperform any single word embedding method and the combination of the two embeddings. Incorporating linguistic features into our model can further improve model performance, yielding state-of-the-art performance on three public metaphor datasets. We also provide in-depth analysis on the effectiveness of leveraging multiple word embeddings, including analysing the spatial distribution of different embedding methods for metaphors and literals, and showing how well the embeddings complement each other in different genres and parts of speech.

Viaarxiv icon

Summarising Historical Text in Modern Languages

Jan 27, 2021
Xutan Peng, Yi Zheng, Chenghua Lin, Advaith Siddharthan

Figure 1 for Summarising Historical Text in Modern Languages
Figure 2 for Summarising Historical Text in Modern Languages
Figure 3 for Summarising Historical Text in Modern Languages
Figure 4 for Summarising Historical Text in Modern Languages

We introduce the task of historical text summarisation, where documents in historical forms of a language are summarised in the corresponding modern language. This is a fundamentally important routine to historians and digital humanities researchers but has never been automated. We compile a high-quality gold-standard text summarisation dataset, which consists of historical German and Chinese news from hundreds of years ago summarised in modern German or Chinese. Based on cross-lingual transfer learning techniques, we propose a summarisation model that can be trained even with no cross-lingual (historical to modern) parallel data, and further benchmark it against state-of-the-art algorithms. We report automatic and human evaluations that distinguish the historic to modern language summarisation task from standard cross-lingual summarisation (i.e., modern to modern language), highlight the distinctness and value of our dataset, and demonstrate that our transfer learning approach outperforms standard cross-lingual benchmarks on this task.

* To appear at EACL 2021 
Viaarxiv icon

Generating Descriptions for Sequential Images with Local-Object Attention and Global Semantic Context Modelling

Dec 02, 2020
Jing Su, Chenghua Lin, Mian Zhou, Qingyun Dai, Haoyu Lv

Figure 1 for Generating Descriptions for Sequential Images with Local-Object Attention and Global Semantic Context Modelling
Figure 2 for Generating Descriptions for Sequential Images with Local-Object Attention and Global Semantic Context Modelling
Figure 3 for Generating Descriptions for Sequential Images with Local-Object Attention and Global Semantic Context Modelling
Figure 4 for Generating Descriptions for Sequential Images with Local-Object Attention and Global Semantic Context Modelling

In this paper, we propose an end-to-end CNN-LSTM model for generating descriptions for sequential images with a local-object attention mechanism. To generate coherent descriptions, we capture global semantic context using a multi-layer perceptron, which learns the dependencies between sequential images. A paralleled LSTM network is exploited for decoding the sequence descriptions. Experimental results show that our model outperforms the baseline across three different evaluation metrics on the datasets published by Microsoft.

* Accepted by INLG 2018 
Viaarxiv icon