Alert button
Picture for Jason Naradowsky

Jason Naradowsky

Alert button

Break it Down for Me: A Study in Automated Lyric Annotation

Aug 11, 2017
Lucas Sterckx, Jason Naradowsky, Bill Byrne, Thomas Demeester, Chris Develder

Figure 1 for Break it Down for Me: A Study in Automated Lyric Annotation
Figure 2 for Break it Down for Me: A Study in Automated Lyric Annotation
Figure 3 for Break it Down for Me: A Study in Automated Lyric Annotation
Figure 4 for Break it Down for Me: A Study in Automated Lyric Annotation

Comprehending lyrics, as found in songs and poems, can pose a challenge to human and machine readers alike. This motivates the need for systems that can understand the ambiguity and jargon found in such creative texts, and provide commentary to aid readers in reaching the correct interpretation. We introduce the task of automated lyric annotation (ALA). Like text simplification, a goal of ALA is to rephrase the original text in a more easily understandable manner. However, in ALA the system must often include additional information to clarify niche terminology and abstract concepts. To stimulate research on this task, we release a large collection of crowdsourced annotations for song lyrics. We analyze the performance of translation and retrieval models on this task, measuring performance with both automated and human evaluation. We find that each model captures a unique type of information important to the task.

* To appear in Proceedings of EMNLP 2017 
Viaarxiv icon

Programming with a Differentiable Forth Interpreter

Jul 23, 2017
Matko Bošnjak, Tim Rocktäschel, Jason Naradowsky, Sebastian Riedel

Figure 1 for Programming with a Differentiable Forth Interpreter
Figure 2 for Programming with a Differentiable Forth Interpreter
Figure 3 for Programming with a Differentiable Forth Interpreter
Figure 4 for Programming with a Differentiable Forth Interpreter

Given that in practice training data is scarce for all but a small set of problems, a core question is how to incorporate prior knowledge into a model. In this paper, we consider the case of prior procedural knowledge for neural networks, such as knowing how a program should traverse a sequence, but not what local actions should be performed at each step. To this end, we present an end-to-end differentiable interpreter for the programming language Forth which enables programmers to write program sketches with slots that can be filled with behaviour trained from program input-output data. We can optimise this behaviour directly through gradient descent techniques on user-specified objectives, and also integrate the program into any larger neural computation graph. We show empirically that our interpreter is able to effectively leverage different levels of prior program structure and learn complex behaviours such as sequence sorting and addition. When connected to outputs of an LSTM and trained jointly, our interpreter achieves state-of-the-art accuracy for end-to-end reasoning about quantities expressed in natural language stories.

* 34th International Conference on Machine Learning (ICML 2017) 
Viaarxiv icon

Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources

Oct 30, 2016
Jason Naradowsky, Sebastian Riedel

Figure 1 for Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources
Figure 2 for Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources
Figure 3 for Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources
Figure 4 for Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources

In order to extract event information from text, a machine reading model must learn to accurately read and interpret the ways in which that information is expressed. But it must also, as the human reader must, aggregate numerous individual value hypotheses into a single coherent global analysis, applying global constraints which reflect prior knowledge of the domain. In this work we focus on the task of extracting plane crash event information from clusters of related news articles whose labels are derived via distant supervision. Unlike previous machine reading work, we assume that while most target values will occur frequently in most clusters, they may also be missing or incorrect. We introduce a novel neural architecture to explicitly model the noisy nature of the data and to deal with these aforementioned learning issues. Our models are trained end-to-end and achieve an improvement of more than 12.1 F$_1$ over previous work, despite using far less linguistic annotation. We apply factor graph constraints to promote more coherent event analyses, with belief propagation inference formulated within the transitions of a recurrent neural network. We show this technique additionally improves maximum F$_1$ by up to 2.8 points, resulting in a relative improvement of $50\%$ over the previous state-of-the-art.

Viaarxiv icon