Alert button
Picture for Jialong Tang

Jialong Tang

Alert button

Harvesting Event Schemas from Large Language Models

May 12, 2023
Jialong Tang, Hongyu Lin, Zhuoqun Li, Yaojie Lu, Xianpei Han, Le Sun

Figure 1 for Harvesting Event Schemas from Large Language Models
Figure 2 for Harvesting Event Schemas from Large Language Models
Figure 3 for Harvesting Event Schemas from Large Language Models
Figure 4 for Harvesting Event Schemas from Large Language Models

Event schema provides a conceptual, structural and formal language to represent events and model the world event knowledge. Unfortunately, it is challenging to automatically induce high-quality and high-coverage event schemas due to the open nature of real-world events, the diversity of event expressions, and the sparsity of event knowledge. In this paper, we propose a new paradigm for event schema induction -- knowledge harvesting from large-scale pre-trained language models, which can effectively resolve the above challenges by discovering, conceptualizing and structuralizing event schemas from PLMs. And an Event Schema Harvester (ESHer) is designed to automatically induce high-quality event schemas via in-context generation-based conceptualization, confidence-aware schema structuralization and graph-based schema aggregation. Empirical results show that ESHer can induce high-quality and high-coverage event schemas on varying domains.

* 14 pages 
Viaarxiv icon

Procedural Text Understanding via Scene-Wise Evolution

Mar 15, 2022
Jialong Tang, Hongyu Lin, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun, Weijian Xie, Jin Xu

Figure 1 for Procedural Text Understanding via Scene-Wise Evolution
Figure 2 for Procedural Text Understanding via Scene-Wise Evolution
Figure 3 for Procedural Text Understanding via Scene-Wise Evolution
Figure 4 for Procedural Text Understanding via Scene-Wise Evolution

Procedural text understanding requires machines to reason about entity states within the dynamical narratives. Current procedural text understanding approaches are commonly \textbf{entity-wise}, which separately track each entity and independently predict different states of each entity. Such an entity-wise paradigm does not consider the interaction between entities and their states. In this paper, we propose a new \textbf{scene-wise} paradigm for procedural text understanding, which jointly tracks states of all entities in a scene-by-scene manner. Based on this paradigm, we propose \textbf{S}cene \textbf{G}raph \textbf{R}easoner (\textbf{SGR}), which introduces a series of dynamically evolving scene graphs to jointly formulate the evolution of entities, states and their associations throughout the narrative. In this way, the deep interactions between all entities and states can be jointly captured and simultaneously derived from scene graphs. Experiments show that SGR not only achieves the new state-of-the-art performance but also significantly accelerates the speed of reasoning.

* AAAI 2022  
* 9 pages, 2 figures 
Viaarxiv icon

Text2Event: Controllable Sequence-to-Structure Generation for End-to-end Event Extraction

Jun 17, 2021
Yaojie Lu, Hongyu Lin, Jin Xu, Xianpei Han, Jialong Tang, Annan Li, Le Sun, Meng Liao, Shaoyi Chen

Figure 1 for Text2Event: Controllable Sequence-to-Structure Generation for End-to-end Event Extraction
Figure 2 for Text2Event: Controllable Sequence-to-Structure Generation for End-to-end Event Extraction
Figure 3 for Text2Event: Controllable Sequence-to-Structure Generation for End-to-end Event Extraction
Figure 4 for Text2Event: Controllable Sequence-to-Structure Generation for End-to-end Event Extraction

Event extraction is challenging due to the complex structure of event records and the semantic gap between text and event. Traditional methods usually extract event records by decomposing the complex structure prediction task into multiple subtasks. In this paper, we propose Text2Event, a sequence-to-structure generation paradigm that can directly extract events from the text in an end-to-end manner. Specifically, we design a sequence-to-structure network for unified event extraction, a constrained decoding algorithm for event knowledge injection during inference, and a curriculum learning algorithm for efficient model learning. Experimental results show that, by uniformly modeling all tasks in a single model and universally predicting different labels, our method can achieve competitive performance using only record-level annotations in both supervised learning and transfer learning settings.

* Accepted to ACL2021 (main conference) 
Viaarxiv icon

From Discourse to Narrative: Knowledge Projection for Event Relation Extraction

Jun 16, 2021
Jialong Tang, Hongyu Lin, Meng Liao, Yaojie Lu, Xianpei Han, Le Sun, Weijian Xie, Jin Xu

Figure 1 for From Discourse to Narrative: Knowledge Projection for Event Relation Extraction
Figure 2 for From Discourse to Narrative: Knowledge Projection for Event Relation Extraction
Figure 3 for From Discourse to Narrative: Knowledge Projection for Event Relation Extraction
Figure 4 for From Discourse to Narrative: Knowledge Projection for Event Relation Extraction

Current event-centric knowledge graphs highly rely on explicit connectives to mine relations between events. Unfortunately, due to the sparsity of connectives, these methods severely undermine the coverage of EventKGs. The lack of high-quality labelled corpora further exacerbates that problem. In this paper, we propose a knowledge projection paradigm for event relation extraction: projecting discourse knowledge to narratives by exploiting the commonalities between them. Specifically, we propose Multi-tier Knowledge Projection Network (MKPNet), which can leverage multi-tier discourse knowledge effectively for event relation extraction. In this way, the labelled data requirement is significantly reduced, and implicit event relations can be effectively extracted. Intrinsic experimental results show that MKPNet achieves the new state-of-the-art performance, and extrinsic experimental results verify the value of the extracted event relations.

* ACL 2021  
* 11 pages 
Viaarxiv icon

Syntactic and Semantic-driven Learning for Open Information Extraction

Mar 05, 2021
Jialong Tang, Yaojie Lu, Hongyu Lin, Xianpei Han, Le Sun, Xinyan Xiao, Hua Wu

Figure 1 for Syntactic and Semantic-driven Learning for Open Information Extraction
Figure 2 for Syntactic and Semantic-driven Learning for Open Information Extraction
Figure 3 for Syntactic and Semantic-driven Learning for Open Information Extraction
Figure 4 for Syntactic and Semantic-driven Learning for Open Information Extraction

One of the biggest bottlenecks in building accurate, high coverage neural open IE systems is the need for large labelled corpora. The diversity of open domain corpora and the variety of natural language expressions further exacerbate this problem. In this paper, we propose a syntactic and semantic-driven learning approach, which can learn neural open IE models without any human-labelled data by leveraging syntactic and semantic knowledge as noisier, higher-level supervisions. Specifically, we first employ syntactic patterns as data labelling functions and pretrain a base model using the generated labels. Then we propose a syntactic and semantic-driven reinforcement learning algorithm, which can effectively generalize the base model to open situations with high accuracy. Experimental results show that our approach significantly outperforms the supervised counterparts, and can even achieve competitive performance to supervised state-of-the-art (SoA) model

* Findings of ACL: EMNLP 2020  
* 11 pages 
Viaarxiv icon

Enhanced Aspect-Based Sentiment Analysis Models with Progressive Self-supervised Attention Learning

Mar 05, 2021
Jinsong Su, Jialong Tang, Hui Jiang, Ziyao Lu, Yubin Ge, Linfeng Song, Deyi Xiong, Le Sun, Jiebo Luo

Figure 1 for Enhanced Aspect-Based Sentiment Analysis Models with Progressive Self-supervised Attention Learning
Figure 2 for Enhanced Aspect-Based Sentiment Analysis Models with Progressive Self-supervised Attention Learning
Figure 3 for Enhanced Aspect-Based Sentiment Analysis Models with Progressive Self-supervised Attention Learning
Figure 4 for Enhanced Aspect-Based Sentiment Analysis Models with Progressive Self-supervised Attention Learning

In aspect-based sentiment analysis (ABSA), many neural models are equipped with an attention mechanism to quantify the contribution of each context word to sentiment prediction. However, such a mechanism suffers from one drawback: only a few frequent words with sentiment polarities are tended to be taken into consideration for final sentiment decision while abundant infrequent sentiment words are ignored by models. To deal with this issue, we propose a progressive self-supervised attention learning approach for attentional ABSA models. In this approach, we iteratively perform sentiment prediction on all training instances, and continually learn useful attention supervision information in the meantime. During training, at each iteration, context words with the highest impact on sentiment prediction, identified based on their attention weights or gradients, are extracted as words with active/misleading influence on the correct/incorrect prediction for each instance. Words extracted in this way are masked for subsequent iterations. To exploit these extracted words for refining ABSA models, we augment the conventional training objective with a regularization term that encourages ABSA models to not only take full advantage of the extracted active context words but also decrease the weights of those misleading words. We integrate the proposed approach into three state-of-the-art neural ABSA models. Experiment results and in-depth analyses show that our approach yields better attention results and significantly enhances the performance of all three models. We release the source code and trained models at https://github.com/DeepLearnXMU/PSSAttention.

* Artificial Intelligence 2021  
* 31 pages. arXiv admin note: text overlap with arXiv:1906.01213 
Viaarxiv icon

End-to-End Neural Event Coreference Resolution

Sep 17, 2020
Yaojie Lu, Hongyu Lin, Jialong Tang, Xianpei Han, Le Sun

Figure 1 for End-to-End Neural Event Coreference Resolution
Figure 2 for End-to-End Neural Event Coreference Resolution
Figure 3 for End-to-End Neural Event Coreference Resolution
Figure 4 for End-to-End Neural Event Coreference Resolution

Traditional event coreference systems usually rely on pipeline framework and hand-crafted features, which often face error propagation problem and have poor generalization ability. In this paper, we propose an End-to-End Event Coreference approach -- E3C neural network, which can jointly model event detection and event coreference resolution tasks, and learn to extract features from raw text automatically. Furthermore, because event mentions are highly diversified and event coreference is intricately governed by long-distance, semantic-dependent decisions, a type-guided event coreference mechanism is further proposed in our E3C neural network. Experiments show that our method achieves new state-of-the-art performance on two standard datasets.

Viaarxiv icon

Progressive Self-Supervised Attention Learning for Aspect-Level Sentiment Analysis

Jun 06, 2019
Jialong Tang, Ziyao Lu, Jinsong Su, Yubin Ge, Linfeng Song, Le Sun, Jiebo Luo

Figure 1 for Progressive Self-Supervised Attention Learning for Aspect-Level Sentiment Analysis
Figure 2 for Progressive Self-Supervised Attention Learning for Aspect-Level Sentiment Analysis
Figure 3 for Progressive Self-Supervised Attention Learning for Aspect-Level Sentiment Analysis
Figure 4 for Progressive Self-Supervised Attention Learning for Aspect-Level Sentiment Analysis

In aspect-level sentiment classification (ASC), it is prevalent to equip dominant neural models with attention mechanisms, for the sake of acquiring the importance of each context word on the given aspect. However, such a mechanism tends to excessively focus on a few frequent words with sentiment polarities, while ignoring infrequent ones. In this paper, we propose a progressive self-supervised attention learning approach for neural ASC models, which automatically mines useful attention supervision information from a training corpus to refine attention mechanisms. Specifically, we iteratively conduct sentiment predictions on all training instances. Particularly, at each iteration, the context word with the maximum attention weight is extracted as the one with active/misleading influence on the correct/incorrect prediction of every instance, and then the word itself is masked for subsequent iterations. Finally, we augment the conventional training objective with a regularization term, which enables ASC models to continue equally focusing on the extracted active context words while decreasing weights of those misleading ones. Experimental results on multiple datasets show that our proposed approach yields better attention mechanisms, leading to substantial improvements over the two state-of-the-art neural ASC models. Source code and trained models are available at https://github.com/DeepLearnXMU/PSSAttention.

* ACL 2019  
* 10 pages 
Viaarxiv icon