Alert button
Picture for Lea Krause

Lea Krause

Alert button

Leveraging Few-Shot Data Augmentation and Waterfall Prompting for Response Generation

Aug 02, 2023
Lea Krause, Selene Báez Santamaría, Michiel van der Meer, Urja Khurana

Figure 1 for Leveraging Few-Shot Data Augmentation and Waterfall Prompting for Response Generation
Figure 2 for Leveraging Few-Shot Data Augmentation and Waterfall Prompting for Response Generation
Figure 3 for Leveraging Few-Shot Data Augmentation and Waterfall Prompting for Response Generation
Figure 4 for Leveraging Few-Shot Data Augmentation and Waterfall Prompting for Response Generation

This paper discusses our approaches for task-oriented conversational modelling using subjective knowledge, with a particular emphasis on response generation. Our methodology was shaped by an extensive data analysis that evaluated key factors such as response length, sentiment, and dialogue acts present in the provided dataset. We used few-shot learning to augment the data with newly generated subjective knowledge items and present three approaches for DSTC11: (1) task-specific model exploration, (2) incorporation of the most frequent question into all generated responses, and (3) a waterfall prompting technique using a combination of both GPT-3 and ChatGPT.

* DSTC11 
Viaarxiv icon

Will It Blend? Mixing Training Paradigms & Prompting for Argument Quality Prediction

Oct 05, 2022
Michiel van der Meer, Myrthe Reuver, Urja Khurana, Lea Krause, Selene Báez Santamaría

Figure 1 for Will It Blend? Mixing Training Paradigms & Prompting for Argument Quality Prediction
Figure 2 for Will It Blend? Mixing Training Paradigms & Prompting for Argument Quality Prediction
Figure 3 for Will It Blend? Mixing Training Paradigms & Prompting for Argument Quality Prediction
Figure 4 for Will It Blend? Mixing Training Paradigms & Prompting for Argument Quality Prediction

This paper describes our contributions to the Shared Task of the 9th Workshop on Argument Mining (2022). Our approach uses Large Language Models for the task of Argument Quality Prediction. We perform prompt engineering using GPT-3, and also investigate the training paradigms multi-task learning, contrastive learning, and intermediate-task training. We find that a mixed prediction setup outperforms single models. Prompting GPT-3 works best for predicting argument validity, and argument novelty is best estimated by a model trained using all three training paradigms.

* Accepted at the 9th Workshop on Argument Mining (2022) 
Viaarxiv icon