Alert button
Picture for Mengting Hu

Mengting Hu

Alert button

Uncertainty in Natural Language Processing: Sources, Quantification, and Applications

Jun 05, 2023
Mengting Hu, Zhen Zhang, Shiwan Zhao, Minlie Huang, Bingzhe Wu

Figure 1 for Uncertainty in Natural Language Processing: Sources, Quantification, and Applications
Figure 2 for Uncertainty in Natural Language Processing: Sources, Quantification, and Applications
Figure 3 for Uncertainty in Natural Language Processing: Sources, Quantification, and Applications
Figure 4 for Uncertainty in Natural Language Processing: Sources, Quantification, and Applications

As a main field of artificial intelligence, natural language processing (NLP) has achieved remarkable success via deep neural networks. Plenty of NLP tasks have been addressed in a unified manner, with various tasks being associated with each other through sharing the same paradigm. However, neural networks are black boxes and rely on probability computation. Making mistakes is inevitable. Therefore, estimating the reliability and trustworthiness (in other words, uncertainty) of neural networks becomes a key research direction, which plays a crucial role in reducing models' risks and making better decisions. Therefore, in this survey, we provide a comprehensive review of uncertainty-relevant works in the NLP field. Considering the data and paradigms characteristics, we first categorize the sources of uncertainty in natural language into three types, including input, system, and output. Then, we systemically review uncertainty quantification approaches and the main applications. Finally, we discuss the challenges of uncertainty estimation in NLP and discuss potential future directions, taking into account recent trends in the field. Though there have been a few surveys about uncertainty estimation, our work is the first to review uncertainty from the NLP perspective.

* This work has been submitted to the IEEE for possible publication 
Viaarxiv icon

Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction

Jun 03, 2023
Mengting Hu, Yinhao Bai, Yike Wu, Zhen Zhang, Liqi Zhang, Hang Gao, Shiwan Zhao, Minlie Huang

Figure 1 for Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction
Figure 2 for Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction
Figure 3 for Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction
Figure 4 for Uncertainty-Aware Unlikelihood Learning Improves Generative Aspect Sentiment Quad Prediction

Recently, aspect sentiment quad prediction has received widespread attention in the field of aspect-based sentiment analysis. Existing studies extract quadruplets via pre-trained generative language models to paraphrase the original sentence into a templated target sequence. However, previous works only focus on what to generate but ignore what not to generate. We argue that considering the negative samples also leads to potential benefits. In this work, we propose a template-agnostic method to control the token-level generation, which boosts original learning and reduces mistakes simultaneously. Specifically, we introduce Monte Carlo dropout to understand the built-in uncertainty of pre-trained language models, acquiring the noises and errors. We further propose marginalized unlikelihood learning to suppress the uncertainty-aware mistake tokens. Finally, we introduce minimization entropy to balance the effects of marginalized unlikelihood learning. Extensive experiments on four public datasets demonstrate the effectiveness of our approach on various generation templates.

* Accepted by ACL Findings (2023) 
Viaarxiv icon

E-NER: Evidential Deep Learning for Trustworthy Named Entity Recognition

May 29, 2023
Zhen Zhang, Mengting Hu, Shiwan Zhao, Minlie Huang, Haotian Wang, Lemao Liu, Zhirui Zhang, Zhe Liu, Bingzhe Wu

Figure 1 for E-NER: Evidential Deep Learning for Trustworthy Named Entity Recognition
Figure 2 for E-NER: Evidential Deep Learning for Trustworthy Named Entity Recognition
Figure 3 for E-NER: Evidential Deep Learning for Trustworthy Named Entity Recognition
Figure 4 for E-NER: Evidential Deep Learning for Trustworthy Named Entity Recognition

Most named entity recognition (NER) systems focus on improving model performance, ignoring the need to quantify model uncertainty, which is critical to the reliability of NER systems in open environments. Evidential deep learning (EDL) has recently been proposed as a promising solution to explicitly model predictive uncertainty for classification tasks. However, directly applying EDL to NER applications faces two challenges, i.e., the problems of sparse entities and OOV/OOD entities in NER tasks. To address these challenges, we propose a trustworthy NER framework named E-NER by introducing two uncertainty-guided loss terms to the conventional EDL, along with a series of uncertainty-guided training strategies. Experiments show that E-NER can be applied to multiple NER paradigms to obtain accurate uncertainty estimation. Furthermore, compared to state-of-the-art baselines, the proposed method achieves a better OOV/OOD detection performance and better generalization ability on OOV entities.

* accepted by ACL Findings (2023) 
Viaarxiv icon

Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation

Oct 19, 2022
Mengting Hu, Yike Wu, Hang Gao, Yinhao Bai, Shiwan Zhao

Figure 1 for Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation
Figure 2 for Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation
Figure 3 for Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation
Figure 4 for Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation

Recently, aspect sentiment quad prediction (ASQP) has become a popular task in the field of aspect-level sentiment analysis. Previous work utilizes a predefined template to paraphrase the original sentence into a structure target sequence, which can be easily decoded as quadruplets of the form (aspect category, aspect term, opinion term, sentiment polarity). The template involves the four elements in a fixed order. However, we observe that this solution contradicts with the order-free property of the ASQP task, since there is no need to fix the template order as long as the quadruplet is extracted correctly. Inspired by the observation, we study the effects of template orders and find that some orders help the generative model achieve better performance. It is hypothesized that different orders provide various views of the quadruplet. Therefore, we propose a simple but effective method to identify the most proper orders, and further combine multiple proper templates as data augmentation to improve the ASQP task. Specifically, we use the pre-trained language model to select the orders with minimal entropy. By fine-tuning the pre-trained language model with these template orders, our approach improves the performance of quad prediction, and outperforms state-of-the-art methods significantly in low-resource settings.

* EMNLP 2022 
Viaarxiv icon

Classical Sequence Match is a Competitive Few-Shot One-Class Learner

Sep 14, 2022
Mengting Hu, Hang Gao, Yinhao Bai, Mingming Liu

Figure 1 for Classical Sequence Match is a Competitive Few-Shot One-Class Learner
Figure 2 for Classical Sequence Match is a Competitive Few-Shot One-Class Learner
Figure 3 for Classical Sequence Match is a Competitive Few-Shot One-Class Learner
Figure 4 for Classical Sequence Match is a Competitive Few-Shot One-Class Learner

Nowadays, transformer-based models gradually become the default choice for artificial intelligence pioneers. The models also show superiority even in the few-shot scenarios. In this paper, we revisit the classical methods and propose a new few-shot alternative. Specifically, we investigate the few-shot one-class problem, which actually takes a known sample as a reference to detect whether an unknown instance belongs to the same class. This problem can be studied from the perspective of sequence match. It is shown that with meta-learning, the classical sequence match method, i.e. Compare-Aggregate, significantly outperforms transformer ones. The classical approach requires much less training cost. Furthermore, we perform an empirical comparison between two kinds of sequence match approaches under simple fine-tuning and meta-learning. Meta-learning causes the transformer models' features to have high-correlation dimensions. The reason is closely related to the number of layers and heads of transformer models. Experimental codes and data are available at https://github.com/hmt2014/FewOne

* COLING 2022 
Viaarxiv icon

Efficient Mind-Map Generation via Sequence-to-Graph and Reinforced Graph Refinement

Sep 06, 2021
Mengting Hu, Honglei Guo, Shiwan Zhao, Hang Gao, Zhong Su

Figure 1 for Efficient Mind-Map Generation via Sequence-to-Graph and Reinforced Graph Refinement
Figure 2 for Efficient Mind-Map Generation via Sequence-to-Graph and Reinforced Graph Refinement
Figure 3 for Efficient Mind-Map Generation via Sequence-to-Graph and Reinforced Graph Refinement
Figure 4 for Efficient Mind-Map Generation via Sequence-to-Graph and Reinforced Graph Refinement

A mind-map is a diagram that represents the central concept and key ideas in a hierarchical way. Converting plain text into a mind-map will reveal its key semantic structure and be easier to understand. Given a document, the existing automatic mind-map generation method extracts the relationships of every sentence pair to generate the directed semantic graph for this document. The computation complexity increases exponentially with the length of the document. Moreover, it is difficult to capture the overall semantics. To deal with the above challenges, we propose an efficient mind-map generation network that converts a document into a graph via sequence-to-graph. To guarantee a meaningful mind-map, we design a graph refinement module to adjust the relation graph in a reinforcement learning manner. Extensive experimental results demonstrate that the proposed approach is more effective and efficient than the existing methods. The inference time is reduced by thousands of times compared with the existing methods. The case studies verify that the generated mind-maps better reveal the underlying semantic structures of the document.

* Accepted by EMNLP 2021 main conference 
Viaarxiv icon

Multi-Label Few-Shot Learning for Aspect Category Detection

May 29, 2021
Mengting Hu, Shiwan Zhao, Honglei Guo, Chao Xue, Hang Gao, Tiegang Gao, Renhong Cheng, Zhong Su

Figure 1 for Multi-Label Few-Shot Learning for Aspect Category Detection
Figure 2 for Multi-Label Few-Shot Learning for Aspect Category Detection
Figure 3 for Multi-Label Few-Shot Learning for Aspect Category Detection
Figure 4 for Multi-Label Few-Shot Learning for Aspect Category Detection

Aspect category detection (ACD) in sentiment analysis aims to identify the aspect categories mentioned in a sentence. In this paper, we formulate ACD in the few-shot learning scenario. However, existing few-shot learning approaches mainly focus on single-label predictions. These methods can not work well for the ACD task since a sentence may contain multiple aspect categories. Therefore, we propose a multi-label few-shot learning method based on the prototypical network. To alleviate the noise, we design two effective attention mechanisms. The support-set attention aims to extract better prototypes by removing irrelevant aspects. The query-set attention computes multiple prototype-specific representations for each query instance, which are then used to compute accurate distances with the corresponding prototypes. To achieve multi-label inference, we further learn a dynamic threshold per instance by a policy network. Extensive experimental results on three datasets demonstrate that the proposed method significantly outperforms strong baselines.

* Accepted by ACL 2021 main conference 
Viaarxiv icon

Hierarchical Ranking for Answer Selection

Feb 01, 2021
Hang Gao, Mengting Hu, Renhong Cheng, Tiegang Gao

Figure 1 for Hierarchical Ranking for Answer Selection
Figure 2 for Hierarchical Ranking for Answer Selection
Figure 3 for Hierarchical Ranking for Answer Selection
Figure 4 for Hierarchical Ranking for Answer Selection

Answer selection is a task to choose the positive answers from a pool of candidate answers for a given question. In this paper, we propose a novel strategy for answer selection, called hierarchical ranking. We introduce three levels of ranking: point-level ranking, pair-level ranking, and list-level ranking. They formulate their optimization objectives by employing supervisory information from different perspectives to achieve the same goal of ranking candidate answers. Therefore, the three levels of ranking are related and they can promote each other. We take the well-performed compare-aggregate model as the backbone and explore three schemes to implement the idea of applying the hierarchical rankings jointly: the scheme under the Multi-Task Learning (MTL) strategy, the Ranking Integration (RI) scheme, and the Progressive Ranking Integration (PRI) scheme. Experimental results on two public datasets, WikiQA and TREC-QA, demonstrate that the proposed hierarchical ranking is effective. Our method achieves state-of-the-art (non-BERT) performance on both TREC-QA and WikiQA.

Viaarxiv icon

Learning to Detect Opinion Snippet for Aspect-Based Sentiment Analysis

Sep 25, 2019
Mengting Hu, Shiwan Zhao, Honglei Guo, Renhong Cheng, Zhong Su

Figure 1 for Learning to Detect Opinion Snippet for Aspect-Based Sentiment Analysis
Figure 2 for Learning to Detect Opinion Snippet for Aspect-Based Sentiment Analysis
Figure 3 for Learning to Detect Opinion Snippet for Aspect-Based Sentiment Analysis
Figure 4 for Learning to Detect Opinion Snippet for Aspect-Based Sentiment Analysis

Aspect-based sentiment analysis (ABSA) is to predict the sentiment polarity towards a particular aspect in a sentence. Recently, this task has been widely addressed by the neural attention mechanism, which computes attention weights to softly select words for generating aspect-specific sentence representations. The attention is expected to concentrate on opinion words for accurate sentiment prediction. However, attention is prone to be distracted by noisy or misleading words, or opinion words from other aspects. In this paper, we propose an alternative hard-selection approach, which determines the start and end positions of the opinion snippet, and selects the words between these two positions for sentiment prediction. Specifically, we learn deep associations between the sentence and aspect, and the long-term dependencies within the sentence by leveraging the pre-trained BERT model. We further detect the opinion snippet by self-critical reinforcement learning. Especially, experimental results demonstrate the effectiveness of our method and prove that our hard-selection approach outperforms soft-selection approaches when handling multi-aspect sentences.

* Accepted by CoNLL 2019 
Viaarxiv icon

Domain-Invariant Feature Distillation for Cross-Domain Sentiment Classification

Aug 24, 2019
Mengting Hu, Yike Wu, Shiwan Zhao, Honglei Guo, Renhong Cheng, Zhong Su

Figure 1 for Domain-Invariant Feature Distillation for Cross-Domain Sentiment Classification
Figure 2 for Domain-Invariant Feature Distillation for Cross-Domain Sentiment Classification
Figure 3 for Domain-Invariant Feature Distillation for Cross-Domain Sentiment Classification
Figure 4 for Domain-Invariant Feature Distillation for Cross-Domain Sentiment Classification

Cross-domain sentiment classification has drawn much attention in recent years. Most existing approaches focus on learning domain-invariant representations in both the source and target domains, while few of them pay attention to the domain-specific information. Despite the non-transferability of the domain-specific information, simultaneously learning domain-dependent representations can facilitate the learning of domain-invariant representations. In this paper, we focus on aspect-level cross-domain sentiment classification, and propose to distill the domain-invariant sentiment features with the help of an orthogonal domain-dependent task, i.e. aspect detection, which is built on the aspects varying widely in different domains. We conduct extensive experiments on three public datasets and the experimental results demonstrate the effectiveness of our method.

* Accepted by EMNLP 2019 
Viaarxiv icon