Alert button
Picture for Patrick Xia

Patrick Xia

Alert button

Few-Shot Adaptation for Parsing Contextual Utterances with LLMs

Sep 18, 2023
Kevin Lin, Patrick Xia, Hao Fang

We evaluate the ability of semantic parsers based on large language models (LLMs) to handle contextual utterances. In real-world settings, there typically exists only a limited number of annotated contextual utterances due to annotation cost, resulting in an imbalance compared to non-contextual utterances. Therefore, parsers must adapt to contextual utterances with a few training examples. We examine four major paradigms for doing so in conversational semantic parsing i.e., Parse-with-Utterance-History, Parse-with-Reference-Program, Parse-then-Resolve, and Rewrite-then-Parse. To facilitate such cross-paradigm comparisons, we construct SMCalFlow-EventQueries, a subset of contextual examples from SMCalFlow with additional annotations. Experiments with in-context learning and fine-tuning suggest that Rewrite-then-Parse is the most promising paradigm when holistically considering parsing accuracy, annotation cost, and error types.

* Findings of IJCNLP-AACL 2023 
Viaarxiv icon

Natural Language Decomposition and Interpretation of Complex Utterances

May 15, 2023
Harsh Jhamtani, Hao Fang, Patrick Xia, Eran Levy, Jacob Andreas, Ben Van Durme

Figure 1 for Natural Language Decomposition and Interpretation of Complex Utterances
Figure 2 for Natural Language Decomposition and Interpretation of Complex Utterances
Figure 3 for Natural Language Decomposition and Interpretation of Complex Utterances
Figure 4 for Natural Language Decomposition and Interpretation of Complex Utterances

Natural language interfaces often require supervised data to translate user requests into programs, database queries, or other structured intent representations. During data collection, it can be difficult to anticipate and formalize the full range of user needs -- for example, in a system designed to handle simple requests (like $\textit{find my meetings tomorrow}$ or $\textit{move my meeting with my manager to noon})$, users may also express more elaborate requests (like $\textit{swap all my calls on Monday and Tuesday}$). We introduce an approach for equipping a simple language-to-code model to handle complex utterances via a process of hierarchical natural language decomposition. Our approach uses a pre-trained language model to decompose a complex utterance into a sequence of smaller natural language steps, then interprets each step using the language-to-code model. To test our approach, we collect and release DeCU -- a new NL-to-program benchmark to evaluate Decomposition of Complex Utterances. Experiments show that the proposed approach enables the interpretation of complex utterances with almost no complex training data, while outperforming standard few-shot prompting approaches.

Viaarxiv icon

Multilingual Coreference Resolution in Multiparty Dialogue

Aug 02, 2022
Boyuan Zheng, Patrick Xia, Mahsa Yarmohammadi, Benjamin Van Durme

Figure 1 for Multilingual Coreference Resolution in Multiparty Dialogue
Figure 2 for Multilingual Coreference Resolution in Multiparty Dialogue
Figure 3 for Multilingual Coreference Resolution in Multiparty Dialogue
Figure 4 for Multilingual Coreference Resolution in Multiparty Dialogue

Existing multiparty dialogue datasets for coreference resolution are nascent, and many challenges are still unaddressed. We create a large-scale dataset, Multilingual Multiparty Coref (MMC), for this task based on TV transcripts. Due to the availability of gold-quality subtitles in multiple languages, we propose reusing the annotations to create silver coreference data in other languages (Chinese and Farsi) via annotation projection. On the gold (English) data, off-the-shelf models perform relatively poorly on MMC, suggesting that MMC has broader coverage of multiparty coreference than prior datasets. On the silver data, we find success both using it for data augmentation and training from scratch, which effectively simulates the zero-shot cross-lingual setting.

Viaarxiv icon

Pruning Pretrained Encoders with a Multitask Objective

Dec 10, 2021
Patrick Xia, Richard Shin

Figure 1 for Pruning Pretrained Encoders with a Multitask Objective
Figure 2 for Pruning Pretrained Encoders with a Multitask Objective
Figure 3 for Pruning Pretrained Encoders with a Multitask Objective
Figure 4 for Pruning Pretrained Encoders with a Multitask Objective

The sizes of pretrained language models make them challenging and expensive to use when there are multiple desired downstream tasks. In this work, we adopt recent strategies for model pruning during finetuning to explore the question of whether it is possible to prune a single encoder so that it can be used for multiple tasks. We allocate a fixed parameter budget and compare pruning a single model with a multitask objective against the best ensemble of single-task models. We find that under two pruning strategies (element-wise and rank pruning), the approach with the multitask objective outperforms training models separately when averaged across all tasks, and it is competitive on each individual one. Additional analysis finds that using a multitask objective during pruning can also be an effective method for reducing model sizes for low-resource tasks.

* ENLSP NeurIPS 2021 
Viaarxiv icon

On Generalization in Coreference Resolution

Sep 20, 2021
Shubham Toshniwal, Patrick Xia, Sam Wiseman, Karen Livescu, Kevin Gimpel

Figure 1 for On Generalization in Coreference Resolution
Figure 2 for On Generalization in Coreference Resolution
Figure 3 for On Generalization in Coreference Resolution
Figure 4 for On Generalization in Coreference Resolution

While coreference resolution is defined independently of dataset domain, most models for performing coreference resolution do not transfer well to unseen domains. We consolidate a set of 8 coreference resolution datasets targeting different domains to evaluate the off-the-shelf performance of models. We then mix three datasets for training; even though their domain, annotation guidelines, and metadata differ, we propose a method for jointly training a single model on this heterogeneous data mixture by using data augmentation to account for annotation differences and sampling to balance the data quantities. We find that in a zero-shot setting, models trained on a single dataset transfer poorly while joint training yields improved overall performance, leading to better generalization in coreference resolution models. This work contributes a new benchmark for robust coreference resolution and multiple new state-of-the-art results.

* CRAC 2021 
Viaarxiv icon

Moving on from OntoNotes: Coreference Resolution Model Transfer

Apr 17, 2021
Patrick Xia, Benjamin Van Durme

Figure 1 for Moving on from OntoNotes: Coreference Resolution Model Transfer
Figure 2 for Moving on from OntoNotes: Coreference Resolution Model Transfer
Figure 3 for Moving on from OntoNotes: Coreference Resolution Model Transfer
Figure 4 for Moving on from OntoNotes: Coreference Resolution Model Transfer

Academic neural models for coreference resolution are typically trained on a single dataset (OntoNotes) and model improvements are then benchmarked on that dataset. However, real-world usages of coreference resolution models depend on the annotation guidelines and the domain of the target dataset, which often differ from those of OntoNotes. We aim to quantify transferability of coreference resolution models based on the number of annotated documents available in the target dataset. We examine five target datasets and find that continued training is consistently effective and especially beneficial when there are few target documents. We establish new benchmarks across several datasets, including state-of-the-art results on LitBank and PreCo.

Viaarxiv icon

Adaptive Active Learning for Coreference Resolution

Apr 15, 2021
Michelle Yuan, Patrick Xia, Benjamin Van Durme, Jordan Boyd-Graber

Figure 1 for Adaptive Active Learning for Coreference Resolution
Figure 2 for Adaptive Active Learning for Coreference Resolution
Figure 3 for Adaptive Active Learning for Coreference Resolution

Training coreference resolution models require comprehensively labeled data. A model trained on one dataset may not successfully transfer to new domains. This paper investigates an approach to active learning for coreference resolution that feeds discrete annotations to an incremental clustering model. The recent developments in incremental coreference resolution allow for a novel approach to active learning in this setting. Through this new framework, we analyze important factors in data acquisition, like sources of model uncertainty and balancing reading and labeling costs. We explore different settings through simulated labeling with gold data. By lowering the data barrier for coreference, coreference resolvers can rapidly adapt to a series of previously unconsidered domains.

Viaarxiv icon

LOME: Large Ontology Multilingual Extraction

Jan 28, 2021
Patrick Xia, Guanghui Qin, Siddharth Vashishtha, Yunmo Chen, Tongfei Chen, Chandler May, Craig Harman, Kyle Rawlins, Aaron Steven White, Benjamin Van Durme

Figure 1 for LOME: Large Ontology Multilingual Extraction
Figure 2 for LOME: Large Ontology Multilingual Extraction
Figure 3 for LOME: Large Ontology Multilingual Extraction
Figure 4 for LOME: Large Ontology Multilingual Extraction

We present LOME, a system for performing multilingual information extraction. Given a text document as input, our core system identifies spans of textual entity and event mentions with a FrameNet (Baker et al., 1998) parser. It subsequently performs coreference resolution, fine-grained entity typing, and temporal relation prediction between events. By doing so, the system constructs an event and entity focused knowledge graph. We can further apply third-party modules for other types of annotation, like relation extraction. Our (multilingual) first-party modules either outperform or are competitive with the (monolingual) state-of-the-art. We achieve this through the use of multilingual encoders like XLM-R (Conneau et al., 2020) and leveraging multilingual training data. LOME is available as a Docker container on Docker Hub. In addition, a lightweight version of the system is accessible as a web demo.

Viaarxiv icon

CopyNext: Explicit Span Copying and Alignment in Sequence to Sequence Models

Oct 28, 2020
Abhinav Singh, Patrick Xia, Guanghui Qin, Mahsa Yarmohammadi, Benjamin Van Durme

Figure 1 for CopyNext: Explicit Span Copying and Alignment in Sequence to Sequence Models
Figure 2 for CopyNext: Explicit Span Copying and Alignment in Sequence to Sequence Models
Figure 3 for CopyNext: Explicit Span Copying and Alignment in Sequence to Sequence Models
Figure 4 for CopyNext: Explicit Span Copying and Alignment in Sequence to Sequence Models

Copy mechanisms are employed in sequence to sequence models (seq2seq) to generate reproductions of words from the input to the output. These frameworks, operating at the lexical type level, fail to provide an explicit alignment that records where each token was copied from. Further, they require contiguous token sequences from the input (spans) to be copied individually. We present a model with an explicit token-level copy operation and extend it to copying entire spans. Our model provides hard alignments between spans in the input and output, allowing for nontraditional applications of seq2seq, like information extraction. We demonstrate the approach on Nested Named Entity Recognition, achieving near state-of-the-art accuracy with an order of magnitude increase in decoding speed.

* 4th Workshop on Structured Prediction for NLP (EMNLP 2020) 
Viaarxiv icon

Which *BERT? A Survey Organizing Contextualized Encoders

Oct 02, 2020
Patrick Xia, Shijie Wu, Benjamin Van Durme

Pretrained contextualized text encoders are now a staple of the NLP community. We present a survey on language representation learning with the aim of consolidating a series of shared lessons learned across a variety of recent efforts. While significant advancements continue at a rapid pace, we find that enough has now been discovered, in different directions, that we can begin to organize advances according to common themes. Through this organization, we highlight important considerations when interpreting recent contributions and choosing which model to use.

* EMNLP 2020 
Viaarxiv icon