Alert button
Picture for Kalina Bontcheva

Kalina Bontcheva

Alert button

Gold Standard Online Debates Summaries and First Experiments Towards Automatic Summarization of Online Debate Data

Aug 15, 2017
Nattapong Sanchan, Ahmet Aker, Kalina Bontcheva

Figure 1 for Gold Standard Online Debates Summaries and First Experiments Towards Automatic Summarization of Online Debate Data
Figure 2 for Gold Standard Online Debates Summaries and First Experiments Towards Automatic Summarization of Online Debate Data
Figure 3 for Gold Standard Online Debates Summaries and First Experiments Towards Automatic Summarization of Online Debate Data
Figure 4 for Gold Standard Online Debates Summaries and First Experiments Towards Automatic Summarization of Online Debate Data

Usage of online textual media is steadily increasing. Daily, more and more news stories, blog posts and scientific articles are added to the online volumes. These are all freely accessible and have been employed extensively in multiple research areas, e.g. automatic text summarization, information retrieval, information extraction, etc. Meanwhile, online debate forums have recently become popular, but have remained largely unexplored. For this reason, there are no sufficient resources of annotated debate data available for conducting research in this genre. In this paper, we collected and annotated debate data for an automatic summarization task. Similar to extractive gold standard summary generation our data contains sentences worthy to include into a summary. Five human annotators performed this task. Inter-annotator agreement, based on semantic similarity, is 36% for Cohen's kappa and 48% for Krippendorff's alpha. Moreover, we also implement an extractive summarization system for online debates and discuss prominent features for the task of summarizing online debate data automatically.

* accepted and presented at the CICLING 2017 - 18th International Conference on Intelligent Text Processing and Computational Linguistics 
Viaarxiv icon

Automatic Summarization of Online Debates

Aug 15, 2017
Nattapong Sanchan, Ahmet Aker, Kalina Bontcheva

Figure 1 for Automatic Summarization of Online Debates
Figure 2 for Automatic Summarization of Online Debates
Figure 3 for Automatic Summarization of Online Debates
Figure 4 for Automatic Summarization of Online Debates

Debate summarization is one of the novel and challenging research areas in automatic text summarization which has been largely unexplored. In this paper, we develop a debate summarization pipeline to summarize key topics which are discussed or argued in the two opposing sides of online debates. We view that the generation of debate summaries can be achieved by clustering, cluster labeling, and visualization. In our work, we investigate two different clustering approaches for the generation of the summaries. In the first approach, we generate the summaries by applying purely term-based clustering and cluster labeling. The second approach makes use of X-means for clustering and Mutual Information for labeling the clusters. Both approaches are driven by ontologies. We visualize the results using bar charts. We think that our results are a smooth entry for users aiming to receive the first impression about what is discussed within a debate topic containing waste number of argumentations.

* Accepted and to be published in Natural Language Processing and Information Retrieval workshop, Recent Advances in Natural Language Processing 2017 (RANLP 2017) 
Viaarxiv icon

SemEval-2017 Task 8: RumourEval: Determining rumour veracity and support for rumours

Apr 20, 2017
Leon Derczynski, Kalina Bontcheva, Maria Liakata, Rob Procter, Geraldine Wong Sak Hoi, Arkaitz Zubiaga

Figure 1 for SemEval-2017 Task 8: RumourEval: Determining rumour veracity and support for rumours
Figure 2 for SemEval-2017 Task 8: RumourEval: Determining rumour veracity and support for rumours
Figure 3 for SemEval-2017 Task 8: RumourEval: Determining rumour veracity and support for rumours
Figure 4 for SemEval-2017 Task 8: RumourEval: Determining rumour veracity and support for rumours

Media is full of false claims. Even Oxford Dictionaries named "post-truth" as the word of 2016. This makes it more important than ever to build systems that can identify the veracity of a story, and the kind of discourse there is around it. RumourEval is a SemEval shared task that aims to identify and handle rumours and reactions to them, in text. We present an annotation scheme, a large dataset covering multiple topics - each having their own families of claims and replies - and use these to pose two concrete challenges as well as the results achieved by participants on these challenges.

Viaarxiv icon

Generalisation in Named Entity Recognition: A Quantitative Analysis

Mar 07, 2017
Isabelle Augenstein, Leon Derczynski, Kalina Bontcheva

Figure 1 for Generalisation in Named Entity Recognition: A Quantitative Analysis
Figure 2 for Generalisation in Named Entity Recognition: A Quantitative Analysis
Figure 3 for Generalisation in Named Entity Recognition: A Quantitative Analysis
Figure 4 for Generalisation in Named Entity Recognition: A Quantitative Analysis

Named Entity Recognition (NER) is a key NLP task, which is all the more challenging on Web and user-generated content with their diverse and continuously changing language. This paper aims to quantify how this diversity impacts state-of-the-art NER methods, by measuring named entity (NE) and context variability, feature sparsity, and their effects on precision and recall. In particular, our findings indicate that NER approaches struggle to generalise in diverse genres with limited training data. Unseen NEs, in particular, play an important role, which have a higher incidence in diverse genres such as social media than in more regular genres such as newswire. Coupled with a higher incidence of unseen features more generally and the lack of large training corpora, this leads to significantly lower F1 scores for diverse genres as compared to more regular ones. We also find that leading systems rely heavily on surface forms found in training data, having problems generalising beyond these, and offer explanations for this observation.

* Preprint, accepted to Computer Speech and Language 
Viaarxiv icon

Stance Detection with Bidirectional Conditional Encoding

Sep 26, 2016
Isabelle Augenstein, Tim Rocktäschel, Andreas Vlachos, Kalina Bontcheva

Figure 1 for Stance Detection with Bidirectional Conditional Encoding
Figure 2 for Stance Detection with Bidirectional Conditional Encoding
Figure 3 for Stance Detection with Bidirectional Conditional Encoding
Figure 4 for Stance Detection with Bidirectional Conditional Encoding

Stance detection is the task of classifying the attitude expressed in a text towards a target such as Hillary Clinton to be "positive", negative" or "neutral". Previous work has assumed that either the target is mentioned in the text or that training data for every target is given. This paper considers the more challenging version of this task, where targets are not always mentioned and no training data is available for the test targets. We experiment with conditional LSTM encoding, which builds a representation of the tweet that is dependent on the target, and demonstrate that it outperforms encoding the tweet and the target independently. Performance is improved further when the conditional model is augmented with bidirectional encoding. We evaluate our approach on the SemEval 2016 Task 6 Twitter Stance Detection corpus achieving performance second best only to a system trained on semi-automatically labelled tweets for the test target. When such weak supervision is added, our approach achieves state-of-the-art results.

* EMNLP 2016  
* 10 pages 
Viaarxiv icon

Using Gaussian Processes for Rumour Stance Classification in Social Media

Sep 07, 2016
Michal Lukasik, Kalina Bontcheva, Trevor Cohn, Arkaitz Zubiaga, Maria Liakata, Rob Procter

Figure 1 for Using Gaussian Processes for Rumour Stance Classification in Social Media
Figure 2 for Using Gaussian Processes for Rumour Stance Classification in Social Media
Figure 3 for Using Gaussian Processes for Rumour Stance Classification in Social Media
Figure 4 for Using Gaussian Processes for Rumour Stance Classification in Social Media

Social media tend to be rife with rumours while new reports are released piecemeal during breaking news. Interestingly, one can mine multiple reactions expressed by social media users in those situations, exploring their stance towards rumours, ultimately enabling the flagging of highly disputed rumours as being potentially false. In this work, we set out to develop an automated, supervised classifier that uses multi-task learning to classify the stance expressed in each individual tweet in a rumourous conversation as either supporting, denying or questioning the rumour. Using a classifier based on Gaussian Processes, and exploring its effectiveness on two datasets with very different characteristics and varying distributions of stances, we show that our approach consistently outperforms competitive baseline classifiers. Our classifier is especially effective in estimating the distribution of different types of stance associated with a given rumour, which we set forth as a desired characteristic for a rumour-tracking system that will warn both ordinary users of Twitter and professional news practitioners when a rumour is being rebutted.

Viaarxiv icon

USFD: Twitter NER with Drift Compensation and Linked Data

Nov 10, 2015
Leon Derczynski, Isabelle Augenstein, Kalina Bontcheva

Figure 1 for USFD: Twitter NER with Drift Compensation and Linked Data
Figure 2 for USFD: Twitter NER with Drift Compensation and Linked Data
Figure 3 for USFD: Twitter NER with Drift Compensation and Linked Data
Figure 4 for USFD: Twitter NER with Drift Compensation and Linked Data

This paper describes a pilot NER system for Twitter, comprising the USFD system entry to the W-NUT 2015 NER shared task. The goal is to correctly label entities in a tweet dataset, using an inventory of ten types. We employ structured learning, drawing on gazetteers taken from Linked Data, and on unsupervised clustering features, and attempting to compensate for stylistic and topic drift - a key challenge in social media text. Our result is competitive; we provide an analysis of the components of our methodology, and an examination of the target dataset in the context of this task.

* Proceedings of the ACL Workshop on Noisy User-generated Text (2015), pp. 48--53  
* Paper in ACL anthology: https://aclweb.org/anthology/W/W15/W15-4306.bib 
Viaarxiv icon

Classifying Tweet Level Judgements of Rumours in Social Media

Sep 10, 2015
Michal Lukasik, Trevor Cohn, Kalina Bontcheva

Figure 1 for Classifying Tweet Level Judgements of Rumours in Social Media
Figure 2 for Classifying Tweet Level Judgements of Rumours in Social Media
Figure 3 for Classifying Tweet Level Judgements of Rumours in Social Media
Figure 4 for Classifying Tweet Level Judgements of Rumours in Social Media

Social media is a rich source of rumours and corresponding community reactions. Rumours reflect different characteristics, some shared and some individual. We formulate the problem of classifying tweet level judgements of rumours as a supervised learning task. Both supervised and unsupervised domain adaptation are considered, in which tweets from a rumour are classified on the basis of other annotated rumours. We demonstrate how multi-task learning helps achieve good results on rumours from the 2011 England riots.

Viaarxiv icon

Analysis of Named Entity Recognition and Linking for Tweets

Oct 27, 2014
Leon Derczynski, Diana Maynard, Giuseppe Rizzo, Marieke van Erp, Genevieve Gorrell, Raphaël Troncy, Johann Petrak, Kalina Bontcheva

Figure 1 for Analysis of Named Entity Recognition and Linking for Tweets
Figure 2 for Analysis of Named Entity Recognition and Linking for Tweets
Figure 3 for Analysis of Named Entity Recognition and Linking for Tweets
Figure 4 for Analysis of Named Entity Recognition and Linking for Tweets

Applying natural language processing for mining and intelligent information access to tweets (a form of microblog) is a challenging, emerging research area. Unlike carefully authored news text and other longer content, tweets pose a number of new challenges, due to their short, noisy, context-dependent, and dynamic nature. Information extraction from tweets is typically performed in a pipeline, comprising consecutive stages of language identification, tokenisation, part-of-speech tagging, named entity recognition and entity disambiguation (e.g. with respect to DBpedia). In this work, we describe a new Twitter entity disambiguation dataset, and conduct an empirical analysis of named entity recognition and disambiguation, investigating how robust a number of state-of-the-art systems are on such noisy texts, what the main sources of error are, and which problems should be further investigated to improve the state of the art.

* Information Processing & Management 51 (2), 32-49, 2014  
* 35 pages, accepted to journal Information Processing and Management 
Viaarxiv icon