Alert button
Picture for Vasudeva Varma

Vasudeva Varma

Alert button

Neural models for Factual Inconsistency Classification with Explanations

Jun 15, 2023
Tathagata Raha, Mukund Choudhary, Abhinav Menon, Harshit Gupta, KV Aditya Srivatsa, Manish Gupta, Vasudeva Varma

Figure 1 for Neural models for Factual Inconsistency Classification with Explanations
Figure 2 for Neural models for Factual Inconsistency Classification with Explanations
Figure 3 for Neural models for Factual Inconsistency Classification with Explanations
Figure 4 for Neural models for Factual Inconsistency Classification with Explanations

Factual consistency is one of the most important requirements when editing high quality documents. It is extremely important for automatic text generation systems like summarization, question answering, dialog modeling, and language modeling. Still, automated factual inconsistency detection is rather under-studied. Existing work has focused on (a) finding fake news keeping a knowledge base in context, or (b) detecting broad contradiction (as part of natural language inference literature). However, there has been no work on detecting and explaining types of factual inconsistencies in text, without any knowledge base in context. In this paper, we leverage existing work in linguistics to formally define five types of factual inconsistencies. Based on this categorization, we contribute a novel dataset, FICLE (Factual Inconsistency CLassification with Explanation), with ~8K samples where each sample consists of two sentences (claim and context) annotated with type and span of inconsistency. When the inconsistency relates to an entity type, it is labeled as well at two levels (coarse and fine-grained). Further, we leverage this dataset to train a pipeline of four neural models to predict inconsistency type with explanations, given a (claim, context) sentence pair. Explanations include inconsistent claim fact triple, inconsistent context span, inconsistent claim component, coarse and fine-grained inconsistent entity types. The proposed system first predicts inconsistent spans from claim and context; and then uses them to predict inconsistency types and inconsistent entity types (when inconsistency is due to entities). We experiment with multiple Transformer-based natural language classification as well as generative models, and find that DeBERTa performs the best. Our proposed methods provide a weighted F1 of ~87% for inconsistency type classification across the five classes.

* ECML-PKDD 2023 
Viaarxiv icon

LLM-RM at SemEval-2023 Task 2: Multilingual Complex NER using XLM-RoBERTa

May 05, 2023
Rahul Mehta, Vasudeva Varma

Figure 1 for LLM-RM at SemEval-2023 Task 2: Multilingual Complex NER using XLM-RoBERTa
Figure 2 for LLM-RM at SemEval-2023 Task 2: Multilingual Complex NER using XLM-RoBERTa
Figure 3 for LLM-RM at SemEval-2023 Task 2: Multilingual Complex NER using XLM-RoBERTa
Figure 4 for LLM-RM at SemEval-2023 Task 2: Multilingual Complex NER using XLM-RoBERTa

Named Entity Recognition(NER) is a task of recognizing entities at a token level in a sentence. This paper focuses on solving NER tasks in a multilingual setting for complex named entities. Our team, LLM-RM participated in the recently organized SemEval 2023 task, Task 2: MultiCoNER II,Multilingual Complex Named Entity Recognition. We approach the problem by leveraging cross-lingual representation provided by fine-tuning XLM-Roberta base model on datasets of all of the 12 languages provided -- Bangla, Chinese, English, Farsi, French, German, Hindi, Italian, Portuguese, Spanish, Swedish and Ukrainian

* Submitted to SemEval-2023, The 17th International Workshop on Semantic Evaluation 
Viaarxiv icon

Summarizing Indian Languages using Multilingual Transformers based Models

Mar 29, 2023
Dhaval Taunk, Vasudeva Varma

Figure 1 for Summarizing Indian Languages using Multilingual Transformers based Models
Figure 2 for Summarizing Indian Languages using Multilingual Transformers based Models
Figure 3 for Summarizing Indian Languages using Multilingual Transformers based Models
Figure 4 for Summarizing Indian Languages using Multilingual Transformers based Models

With the advent of multilingual models like mBART, mT5, IndicBART etc., summarization in low resource Indian languages is getting a lot of attention now a days. But still the number of datasets is low in number. In this work, we (Team HakunaMatata) study how these multilingual models perform on the datasets which have Indian languages as source and target text while performing summarization. We experimented with IndicBART and mT5 models to perform the experiments and report the ROUGE-1, ROUGE-2, ROUGE-3 and ROUGE-4 scores as a performance metric.

* Forum for Information Retrieval Evaluation, December 9-13, 2022, India  
Viaarxiv icon

GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering

Mar 22, 2023
Dhaval Taunk, Lakshya Khanna, Pavan Kandru, Vasudeva Varma, Charu Sharma, Makarand Tapaswi

Figure 1 for GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering
Figure 2 for GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering
Figure 3 for GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering
Figure 4 for GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering

Commonsense question-answering (QA) methods combine the power of pre-trained Language Models (LM) with the reasoning provided by Knowledge Graphs (KG). A typical approach collects nodes relevant to the QA pair from a KG to form a Working Graph (WG) followed by reasoning using Graph Neural Networks(GNNs). This faces two major challenges: (i) it is difficult to capture all the information from the QA in the WG, and (ii) the WG contains some irrelevant nodes from the KG. To address these, we propose GrapeQA with two simple improvements on the WG: (i) Prominent Entities for Graph Augmentation identifies relevant text chunks from the QA pair and augments the WG with corresponding latent representations from the LM, and (ii) Context-Aware Node Pruning removes nodes that are less relevant to the QA pair. We evaluate our results on OpenBookQA, CommonsenseQA and MedQA-USMLE and see that GrapeQA shows consistent improvements over its LM + KG predecessor (QA-GNN in particular) and large improvements on OpenBookQA.

Viaarxiv icon

XWikiGen: Cross-lingual Summarization for Encyclopedic Text Generation in Low Resource Languages

Mar 22, 2023
Dhaval Taunk, Shivprasad Sagare, Anupam Patil, Shivansh Subramanian, Manish Gupta, Vasudeva Varma

Figure 1 for XWikiGen: Cross-lingual Summarization for Encyclopedic Text Generation in Low Resource Languages
Figure 2 for XWikiGen: Cross-lingual Summarization for Encyclopedic Text Generation in Low Resource Languages
Figure 3 for XWikiGen: Cross-lingual Summarization for Encyclopedic Text Generation in Low Resource Languages
Figure 4 for XWikiGen: Cross-lingual Summarization for Encyclopedic Text Generation in Low Resource Languages

Lack of encyclopedic text contributors, especially on Wikipedia, makes automated text generation for \emph{low resource (LR) languages} a critical problem. Existing work on Wikipedia text generation has focused on \emph{English only} where English reference articles are summarized to generate English Wikipedia pages. But, for low-resource languages, the scarcity of reference articles makes monolingual summarization ineffective in solving this problem. Hence, in this work, we propose \task{}, which is the task of cross-lingual multi-document summarization of text from multiple reference articles, written in various languages, to generate Wikipedia-style text. Accordingly, we contribute a benchmark dataset, \data{}, spanning $\sim$69K Wikipedia articles covering five domains and eight languages. We harness this dataset to train a two-stage system where the input is a set of citations and a section title and the output is a section-specific LR summary. The proposed system is based on a novel idea of neural unsupervised extractive summarization to coarsely identify salient information followed by a neural abstractive model to generate the section-specific text. Extensive experiments show that multi-domain training is better than the multi-lingual setup on average.

Viaarxiv icon

Massively Multilingual Language Models for Cross Lingual Fact Extraction from Low Resource Indian Languages

Feb 09, 2023
Bhavyajeet Singh, Pavan Kandru, Anubhav Sharma, Vasudeva Varma

Figure 1 for Massively Multilingual Language Models for Cross Lingual Fact Extraction from Low Resource Indian Languages
Figure 2 for Massively Multilingual Language Models for Cross Lingual Fact Extraction from Low Resource Indian Languages
Figure 3 for Massively Multilingual Language Models for Cross Lingual Fact Extraction from Low Resource Indian Languages
Figure 4 for Massively Multilingual Language Models for Cross Lingual Fact Extraction from Low Resource Indian Languages

Massive knowledge graphs like Wikidata attempt to capture world knowledge about multiple entities. Recent approaches concentrate on automatically enriching these KGs from text. However a lot of information present in the form of natural text in low resource languages is often missed out. Cross Lingual Information Extraction aims at extracting factual information in the form of English triples from low resource Indian Language text. Despite its massive potential, progress made on this task is lagging when compared to Monolingual Information Extraction. In this paper, we propose the task of Cross Lingual Fact Extraction(CLFE) from text and devise an end-to-end generative approach for the same which achieves an overall F1 score of 77.46.

* 5 pages, 2 page Apendix, 3 figures, accepted at 19th International Conference on Natural Language Processing 
Viaarxiv icon

Investigating Strategies for Clause Recommendation

Jan 21, 2023
Sagar Joshi, Sumanth Balaji, Jerrin Thomas, Aparna Garimella, Vasudeva Varma

Figure 1 for Investigating Strategies for Clause Recommendation
Figure 2 for Investigating Strategies for Clause Recommendation
Figure 3 for Investigating Strategies for Clause Recommendation
Figure 4 for Investigating Strategies for Clause Recommendation

Clause recommendation is the problem of recommending a clause to a legal contract, given the context of the contract in question and the clause type to which the clause should belong. With not much prior work being done toward the generation of legal contracts, this problem was proposed as a first step toward the bigger problem of contract generation. As an open-ended text generation problem, the distinguishing characteristics of this problem lie in the nature of legal language as a sublanguage and the considerable similarity of textual content within the clauses of a specific type. This similarity aspect in legal clauses drives us to investigate the importance of similar contracts' representation for recommending clauses. In our work, we experiment with generating clauses for 15 commonly occurring clause types in contracts expanding upon the previous work on this problem and analyzing clause recommendations in varying settings using information derived from similar contracts.

* Volume 362: Legal Knowledge and Information Systems (2022), Frontiers in Artificial Intelligence and Applications  
* Published in Legal Knowledge and Information Systems (JURIX) 2022. (10 pages, 4 figures) 
Viaarxiv icon

Graph-based Keyword Planning for Legal Clause Generation from Topics

Jan 07, 2023
Sagar Joshi, Sumanth Balaji, Aparna Garimella, Vasudeva Varma

Figure 1 for Graph-based Keyword Planning for Legal Clause Generation from Topics
Figure 2 for Graph-based Keyword Planning for Legal Clause Generation from Topics
Figure 3 for Graph-based Keyword Planning for Legal Clause Generation from Topics
Figure 4 for Graph-based Keyword Planning for Legal Clause Generation from Topics

Generating domain-specific content such as legal clauses based on minimal user-provided information can be of significant benefit in automating legal contract generation. In this paper, we propose a controllable graph-based mechanism that can generate legal clauses using only the topic or type of the legal clauses. Our pipeline consists of two stages involving a graph-based planner followed by a clause generator. The planner outlines the content of a legal clause as a sequence of keywords in the order of generic to more specific clause information based on the input topic using a controllable graph-based mechanism. The generation stage takes in a given plan and generates a clause. The pipeline consists of a graph-based planner followed by text generation. We illustrate the effectiveness of our proposed two-stage approach on a broad set of clause topics in contracts.

* To be published in the Natural Legal Language Processing Workshop, EMNLP 2022 (11 pages, 7 figures) 
Viaarxiv icon

Towards Proactively Forecasting Sentence-Specific Information Popularity within Online News Documents

Dec 31, 2022
Sayar Ghosh Roy, Anshul Padhi, Risubh Jain, Manish Gupta, Vasudeva Varma

Figure 1 for Towards Proactively Forecasting Sentence-Specific Information Popularity within Online News Documents
Figure 2 for Towards Proactively Forecasting Sentence-Specific Information Popularity within Online News Documents
Figure 3 for Towards Proactively Forecasting Sentence-Specific Information Popularity within Online News Documents
Figure 4 for Towards Proactively Forecasting Sentence-Specific Information Popularity within Online News Documents

Multiple studies have focused on predicting the prospective popularity of an online document as a whole, without paying attention to the contributions of its individual parts. We introduce the task of proactively forecasting popularities of sentences within online news documents solely utilizing their natural language content. We model sentence-specific popularity forecasting as a sequence regression task. For training our models, we curate InfoPop, the first dataset containing popularity labels for over 1.7 million sentences from over 50,000 online news documents. To the best of our knowledge, this is the first dataset automatically created using streams of incoming search engine queries to generate sentence-level popularity annotations. We propose a novel transfer learning approach involving sentence salience prediction as an auxiliary task. Our proposed technique coupled with a BERT-based neural model exceeds nDCG values of 0.8 for proactive sentence-specific popularity forecasting. Notably, our study presents a non-trivial takeaway: though popularity and salience are different concepts, transfer learning from salience prediction enhances popularity forecasting. We release InfoPop and make our code publicly available: https://github.com/sayarghoshroy/InfoPopularity

* In HT '22. Association for Computing Machinery, New York, NY, USA, 11-20 (2022)  
* In 33rd ACM Conference on Hypertext and Social Media [HT '22] (Main Track), Link: https://dl.acm.org/doi/10.1145/3511095.3531268 
Viaarxiv icon

XF2T: Cross-lingual Fact-to-Text Generation for Low-Resource Languages

Sep 22, 2022
Shivprasad Sagare, Tushar Abhishek, Bhavyajeet Singh, Anubhav Sharma, Manish Gupta, Vasudeva Varma

Figure 1 for XF2T: Cross-lingual Fact-to-Text Generation for Low-Resource Languages
Figure 2 for XF2T: Cross-lingual Fact-to-Text Generation for Low-Resource Languages
Figure 3 for XF2T: Cross-lingual Fact-to-Text Generation for Low-Resource Languages
Figure 4 for XF2T: Cross-lingual Fact-to-Text Generation for Low-Resource Languages

Multiple business scenarios require an automated generation of descriptive human-readable text from structured input data. Hence, fact-to-text generation systems have been developed for various downstream tasks like generating soccer reports, weather and financial reports, medical reports, person biographies, etc. Unfortunately, previous work on fact-to-text (F2T) generation has focused primarily on English mainly due to the high availability of relevant datasets. Only recently, the problem of cross-lingual fact-to-text (XF2T) was proposed for generation across multiple languages alongwith a dataset, XALIGN for eight languages. However, there has been no rigorous work on the actual XF2T generation problem. We extend XALIGN dataset with annotated data for four more languages: Punjabi, Malayalam, Assamese and Oriya. We conduct an extensive study using popular Transformer-based text generation models on our extended multi-lingual dataset, which we call XALIGNV2. Further, we investigate the performance of different text generation strategies: multiple variations of pretraining, fact-aware embeddings and structure-aware input encoding. Our extensive experiments show that a multi-lingual mT5 model which uses fact-aware embeddings with structure-aware input encoding leads to best results on average across the twelve languages. We make our code, dataset and model publicly available, and hope that this will help advance further research in this critical area.

Viaarxiv icon