Alert button
Picture for Anne Lauscher

Anne Lauscher

Alert button

Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender

Feb 24, 2022
Anne Lauscher, Archie Crowley, Dirk Hovy

Figure 1 for Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender
Figure 2 for Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender
Figure 3 for Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender
Figure 4 for Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender

The world of pronouns is changing. From a closed class of words with few members to a much more open set of terms to reflect identities. However, Natural Language Processing (NLP) is barely reflecting this linguistic shift, even though recent work outlined the harms of gender-exclusive language technology. Particularly problematic is the current modeling 3rd person pronouns, as it largely ignores various phenomena like neopronouns, i.e., pronoun sets that are novel and not (yet) widely established. This omission contributes to the discrimination of marginalized and underrepresented groups, e.g., non-binary individuals. However, other identity-expression phenomena beyond gender are also ignored by current NLP technology. In this paper, we provide an overview of 3rd person pronoun issues for NLP. Based on our observations and ethical considerations, we define a series of desiderata for modeling pronouns in language technology. We evaluate existing and novel modeling approaches w.r.t. these desiderata qualitatively, and quantify the impact of a more discrimination-free approach on established benchmark data.

Viaarxiv icon

DS-TOD: Efficient Domain Specialization for Task Oriented Dialog

Oct 15, 2021
Chia-Chien Hung, Anne Lauscher, Simone Paolo Ponzetto, Goran Glavaš

Figure 1 for DS-TOD: Efficient Domain Specialization for Task Oriented Dialog
Figure 2 for DS-TOD: Efficient Domain Specialization for Task Oriented Dialog
Figure 3 for DS-TOD: Efficient Domain Specialization for Task Oriented Dialog
Figure 4 for DS-TOD: Efficient Domain Specialization for Task Oriented Dialog

Recent work has shown that self-supervised dialog-specific pretraining on large conversational datasets yields substantial gains over traditional language modeling (LM) pretraining in downstream task-oriented dialog (TOD). These approaches, however, exploit general dialogic corpora (e.g., Reddit) and thus presumably fail to reliably embed domain-specific knowledge useful for concrete downstream TOD domains. In this work, we investigate the effects of domain specialization of pretrained language models (PLMs) for task-oriented dialog. Within our DS-TOD framework, we first automatically extract salient domain-specific terms, and then use them to construct DomainCC and DomainReddit -- resources that we leverage for domain-specific pretraining, based on (i) masked language modeling (MLM) and (ii) response selection (RS) objectives, respectively. We further propose a resource-efficient and modular domain specialization by means of domain adapters -- additional parameter-light layers in which we encode the domain knowledge. Our experiments with two prominent TOD tasks -- dialog state tracking (DST) and response retrieval (RR) -- encompassing five domains from the MultiWOZ TOD benchmark demonstrate the effectiveness of our domain specialization approach. Moreover, we show that the light-weight adapter-based specialization (1) performs comparably to full fine-tuning in single-domain setups and (2) is particularly suitable for multi-domain specialization, in which, besides advantageous computational footprint, it can offer better downstream performance.

Viaarxiv icon

Sustainable Modular Debiasing of Language Models

Sep 08, 2021
Anne Lauscher, Tobias Lüken, Goran Glavaš

Figure 1 for Sustainable Modular Debiasing of Language Models
Figure 2 for Sustainable Modular Debiasing of Language Models
Figure 3 for Sustainable Modular Debiasing of Language Models
Figure 4 for Sustainable Modular Debiasing of Language Models

Unfair stereotypical biases (e.g., gender, racial, or religious biases) encoded in modern pretrained language models (PLMs) have negative ethical implications for widespread adoption of state-of-the-art language technology. To remedy for this, a wide range of debiasing techniques have recently been introduced to remove such stereotypical biases from PLMs. Existing debiasing methods, however, directly modify all of the PLMs parameters, which -- besides being computationally expensive -- comes with the inherent risk of (catastrophic) forgetting of useful language knowledge acquired in pretraining. In this work, we propose a more sustainable modular debiasing approach based on dedicated debiasing adapters, dubbed ADELE. Concretely, we (1) inject adapter modules into the original PLM layers and (2) update only the adapters (i.e., we keep the original PLM parameters frozen) via language modeling training on a counterfactually augmented corpus. We showcase ADELE, in gender debiasing of BERT: our extensive evaluation, encompassing three intrinsic and two extrinsic bias measures, renders ADELE, very effective in bias mitigation. We further show that -- due to its modular nature -- ADELE, coupled with task adapters, retains fairness even after large-scale downstream training. Finally, by means of multilingual BERT, we successfully transfer ADELE, to six target languages.

* Accepted for EMNLP-Findings 2021 
Viaarxiv icon

Diachronic Analysis of German Parliamentary Proceedings: Ideological Shifts through the Lens of Political Biases

Aug 13, 2021
Tobias Walter, Celina Kirschner, Steffen Eger, Goran Glavaš, Anne Lauscher, Simone Paolo Ponzetto

Figure 1 for Diachronic Analysis of German Parliamentary Proceedings: Ideological Shifts through the Lens of Political Biases
Figure 2 for Diachronic Analysis of German Parliamentary Proceedings: Ideological Shifts through the Lens of Political Biases
Figure 3 for Diachronic Analysis of German Parliamentary Proceedings: Ideological Shifts through the Lens of Political Biases
Figure 4 for Diachronic Analysis of German Parliamentary Proceedings: Ideological Shifts through the Lens of Political Biases

We analyze bias in historical corpora as encoded in diachronic distributional semantic models by focusing on two specific forms of bias, namely a political (i.e., anti-communism) and racist (i.e., antisemitism) one. For this, we use a new corpus of German parliamentary proceedings, DeuPARL, spanning the period 1867--2020. We complement this analysis of historical biases in diachronic word embeddings with a novel measure of bias on the basis of term co-occurrences and graph-based label propagation. The results of our bias measurements align with commonly perceived historical trends of antisemitic and anti-communist biases in German politics in different time periods, thus indicating the viability of analyzing historical bias trends using semantic spaces induced from historical corpora.

* Accepted for JCDL2021 
Viaarxiv icon

MultiCite: Modeling realistic citations requires moving beyond the single-sentence single-label setting

Aug 01, 2021
Anne Lauscher, Brandon Ko, Bailey Kuehl, Sophie Johnson, David Jurgens, Arman Cohan, Kyle Lo

Figure 1 for MultiCite: Modeling realistic citations requires moving beyond the single-sentence single-label setting
Figure 2 for MultiCite: Modeling realistic citations requires moving beyond the single-sentence single-label setting
Figure 3 for MultiCite: Modeling realistic citations requires moving beyond the single-sentence single-label setting
Figure 4 for MultiCite: Modeling realistic citations requires moving beyond the single-sentence single-label setting

Citation context analysis (CCA) is an important task in natural language processing that studies how and why scholars discuss each others' work. Despite decades of study, traditional frameworks for CCA have largely relied on overly-simplistic assumptions of how authors cite, which ignore several important phenomena. For instance, scholarly papers often contain rich discussions of cited work that span multiple sentences and express multiple intents concurrently. Yet, CCA is typically approached as a single-sentence, single-label classification task, and thus existing datasets fail to capture this interesting discourse. In our work, we address this research gap by proposing a novel framework for CCA as a document-level context extraction and labeling task. We release MultiCite, a new dataset of 12,653 citation contexts from over 1,200 computational linguistics papers. Not only is it the largest collection of expert-annotated citation contexts to-date, MultiCite contains multi-sentence, multi-label citation contexts within full paper texts. Finally, we demonstrate how our dataset, while still usable for training classic CCA models, also supports the development of new types of models for CCA beyond fixed-width text classification. We release our code and dataset at https://github.com/allenai/multicite.

Viaarxiv icon

Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation

Jul 31, 2021
Anne Lauscher, Henning Wachsmuth, Iryna Gurevych, Goran Glavaš

Figure 1 for Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation
Figure 2 for Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation
Figure 3 for Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation
Figure 4 for Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation

Despite extensive research efforts in the recent years, computational modeling of argumentation remains one of the most challenging areas of natural language processing (NLP). This is primarily due to inherent complexity of the cognitive processes behind human argumentation, which commonly combine and integrate plethora of different types of knowledge, requiring from computational models capabilities that are far beyond what is needed for most other (i.e., simpler) natural language understanding tasks. The existing large body of work on mining, assessing, generating, and reasoning over arguments largely acknowledges that much more common sense and world knowledge needs to be integrated into computational models that would accurately model argumentation. A systematic overview and organization of the types of knowledge introduced in existing models of computational argumentation (CA) is, however, missing and this hinders targeted progress in the field. In this survey paper, we fill this gap by (1) proposing a pyramid of types of knowledge required in CA tasks, (2) analysing the state of the art with respect to the reliance and exploitation of these types of knowledge, for each of the for main research areas in CA, and (3) outlining and discussing directions for future research efforts in CA.

Viaarxiv icon

RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models

Jun 07, 2021
Soumya Barikeri, Anne Lauscher, Ivan Vulić, Goran Glavaš

Figure 1 for RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models
Figure 2 for RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models
Figure 3 for RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models
Figure 4 for RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models

Text representation models are prone to exhibit a range of societal biases, reflecting the non-controlled and biased nature of the underlying pretraining data, which consequently leads to severe ethical issues and even bias amplification. Recent work has predominantly focused on measuring and mitigating bias in pretrained language models. Surprisingly, the landscape of bias measurements and mitigation resources and methods for conversational language models is still very scarce: it is limited to only a few types of bias, artificially constructed resources, and completely ignores the impact that debiasing methods may have on the final performance in dialog tasks, e.g., conversational response generation. In this work, we present RedditBias, the first conversational data set grounded in the actual human conversations from Reddit, allowing for bias measurement and mitigation across four important bias dimensions: gender, race, religion, and queerness. Further, we develop an evaluation framework which simultaneously 1) measures bias on the developed RedditBias resource, and 2) evaluates model capability in dialog tasks after model debiasing. We use the evaluation framework to benchmark the widely used conversational DialoGPT model along with the adaptations of four debiasing methods. Our results indicate that DialoGPT is biased with respect to religious groups and that some debiasing techniques can remove this bias while preserving downstream task performance.

* Accepted for ACL21 
Viaarxiv icon

DebIE: A Platform for Implicit and Explicit Debiasing of Word Embedding Spaces

Mar 11, 2021
Niklas Friedrich, Anne Lauscher, Simone Paolo Ponzetto, Goran Glavaš

Figure 1 for DebIE: A Platform for Implicit and Explicit Debiasing of Word Embedding Spaces
Figure 2 for DebIE: A Platform for Implicit and Explicit Debiasing of Word Embedding Spaces
Figure 3 for DebIE: A Platform for Implicit and Explicit Debiasing of Word Embedding Spaces

Recent research efforts in NLP have demonstrated that distributional word vector spaces often encode stereotypical human biases, such as racism and sexism. With word representations ubiquitously used in NLP models and pipelines, this raises ethical issues and jeopardizes the fairness of language technologies. While there exists a large body of work on bias measures and debiasing methods, to date, there is no platform that would unify these research efforts and make bias measuring and debiasing of representation spaces widely accessible. In this work, we present DebIE, the first integrated platform for (1) measuring and (2) mitigating bias in word embeddings. Given an (i) embedding space (users can choose between the predefined spaces or upload their own) and (ii) a bias specification (users can choose between existing bias specifications or create their own), DebIE can (1) compute several measures of implicit and explicit bias and modify the embedding space by executing two (mutually composable) debiasing models. DebIE's functionality can be accessed through four different interfaces: (a) a web application, (b) a desktop application, (c) a REST-ful API, and (d) as a command-line application. DebIE is available at: debie.informatik.uni-mannheim.de.

* Accepted as EACL21 Demo 
Viaarxiv icon