Alert button
Picture for Mathieu Ravaut

Mathieu Ravaut

Alert button

PromptSum: Parameter-Efficient Controllable Abstractive Summarization

Aug 06, 2023
Mathieu Ravaut, Hailin Chen, Ruochen Zhao, Chengwei Qin, Shafiq Joty, Nancy Chen

Figure 1 for PromptSum: Parameter-Efficient Controllable Abstractive Summarization
Figure 2 for PromptSum: Parameter-Efficient Controllable Abstractive Summarization
Figure 3 for PromptSum: Parameter-Efficient Controllable Abstractive Summarization
Figure 4 for PromptSum: Parameter-Efficient Controllable Abstractive Summarization

Prompt tuning (PT), a parameter-efficient technique that only tunes the additional prompt embeddings while keeping the backbone pre-trained language model (PLM) frozen, has shown promising results in language understanding tasks, especially in low-resource scenarios. However, effective prompt design methods suitable for generation tasks such as summarization are still lacking. At the same time, summarization guided through instructions (discrete prompts) can achieve a desirable double objective of high quality and controllability in summary generation. Towards a goal of strong summarization performance under the triple conditions of parameter-efficiency, data-efficiency, and controllability, we introduce PromptSum, a method combining PT with a multi-task objective and discrete entity prompts for abstractive summarization. Our model achieves competitive ROUGE results on popular abstractive summarization benchmarks coupled with a strong level of controllability through entities, all while only tuning several orders of magnitude less parameters.

Viaarxiv icon

A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets

Apr 02, 2023
Iva Bojic, Josef Halim, Verena Suharman, Sreeja Tar, Qi Chwen Ong, Duy Phung, Mathieu Ravaut, Shafiq Joty, Josip Car

Figure 1 for A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets
Figure 2 for A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets
Figure 3 for A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets
Figure 4 for A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets

Low-quality data can cause downstream problems in high-stakes applications. Data-centric approach emphasizes on improving dataset quality to enhance model performance. High-quality datasets are needed for general-purpose Large Language Models (LLMs) training, as well as for domain-specific models, which are usually small in size as it is costly to engage a large number of domain experts for their creation. Thus, it is vital to ensure high-quality domain-specific training data. In this paper, we propose a framework for enhancing the data quality of original datasets. We applied the proposed framework to four biomedical datasets and showed relative improvement of up to 33%/40% for fine-tuning of retrieval/reader models on the BioASQ dataset when using back translation to enhance the original dataset quality.

Viaarxiv icon

Unsupervised Summarization Re-ranking

Dec 19, 2022
Mathieu Ravaut, Shafiq Joty, Nancy Chen

Figure 1 for Unsupervised Summarization Re-ranking
Figure 2 for Unsupervised Summarization Re-ranking
Figure 3 for Unsupervised Summarization Re-ranking
Figure 4 for Unsupervised Summarization Re-ranking

With the rise of task-specific pre-training objectives, abstractive summarization models like PEGASUS offer appealing zero-shot performance on downstream summarization tasks. However, the performance of such unsupervised models still lags significantly behind their supervised counterparts. Similarly to the supervised setup, we notice a very high variance in quality among summary candidates from these models whereas only one candidate is kept as the summary output. In this paper, we propose to re-rank summary candidates in an unsupervised manner, aiming to close the performance gap between unsupervised and supervised models. Our approach improves the pre-trained unsupervised PEGASUS by 4.37% to 7.27% relative mean ROUGE across four widely-adopted summarization benchmarks, and achieves relative gains of 7.51% (up to 23.73%) averaged over 30 transfer setups.

Viaarxiv icon

Towards Summary Candidates Fusion

Oct 17, 2022
Mathieu Ravaut, Shafiq Joty, Nancy F. Chen

Figure 1 for Towards Summary Candidates Fusion
Figure 2 for Towards Summary Candidates Fusion
Figure 3 for Towards Summary Candidates Fusion
Figure 4 for Towards Summary Candidates Fusion

Sequence-to-sequence deep neural models fine-tuned for abstractive summarization can achieve great performance on datasets with enough human annotations. Yet, it has been shown that they have not reached their full potential, with a wide gap between the top beam search output and the oracle beam. Recently, re-ranking methods have been proposed, to learn to select a better summary candidate. However, such methods are limited by the summary quality aspects captured by the first-stage candidates. To bypass this limitation, we propose a new paradigm in second-stage abstractive summarization called SummaFusion that fuses several summary candidates to produce a novel abstractive second-stage summary. Our method works well on several summarization datasets, improving both the ROUGE scores and qualitative properties of fused summaries. It is especially good when the candidates to fuse are worse, such as in the few-shot setup where we set a new state-of-the-art. We will make our code and checkpoints available at https://github.com/ntunlp/SummaFusion/.

* 4 Figures, 9 Tables, EMNLP 2022 
Viaarxiv icon

SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization

Mar 13, 2022
Mathieu Ravaut, Shafiq Joty, Nancy F. Chen

Figure 1 for SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization
Figure 2 for SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization
Figure 3 for SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization
Figure 4 for SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization

Sequence-to-sequence neural networks have recently achieved great success in abstractive summarization, especially through fine-tuning large pre-trained language models on the downstream dataset. These models are typically decoded with beam search to generate a unique summary. However, the search space is very large, and with the exposure bias, such decoding is not optimal. In this paper, we show that it is possible to directly train a second-stage model performing re-ranking on a set of summary candidates. Our mixture-of-experts SummaReranker learns to select a better candidate and consistently improves the performance of the base model. With a base PEGASUS, we push ROUGE scores by 5.44% on CNN-DailyMail (47.16 ROUGE-1), 1.31% on XSum (48.12 ROUGE-1) and 9.34% on Reddit TIFU (29.83 ROUGE-1), reaching a new state-of-the-art. Our code and checkpoints will be available at https://github.com/ntunlp/SummaReranker.

* 9 pages, 6 figures, 6 tables, 9 appendix pages, ACL 2022 
Viaarxiv icon

Diabetes Mellitus Forecasting Using Population Health Data in Ontario, Canada

Apr 08, 2019
Mathieu Ravaut, Hamed Sadeghi, Kin Kwan Leung, Maksims Volkovs, Laura C. Rosella

Figure 1 for Diabetes Mellitus Forecasting Using Population Health Data in Ontario, Canada
Figure 2 for Diabetes Mellitus Forecasting Using Population Health Data in Ontario, Canada
Figure 3 for Diabetes Mellitus Forecasting Using Population Health Data in Ontario, Canada
Figure 4 for Diabetes Mellitus Forecasting Using Population Health Data in Ontario, Canada

Leveraging health administrative data (HAD) datasets for predicting the risk of chronic diseases including diabetes has gained a lot of attention in the machine learning community recently. In this paper, we use the largest health records datasets of patients in Ontario,Canada. Provided by the Institute of Clinical Evaluative Sciences (ICES), this database is age, gender and ethnicity-diverse. The datasets include demographics, lab measurements,drug benefits, healthcare system interactions, ambulatory and hospitalizations records. We perform one of the first large-scale machine learning studies with this data to study the task of predicting diabetes in a range of 1-10 years ahead, which requires no additional screening of individuals.In the best setup, we reach a test AUC of 80.3 with a single-model trained on an observation window of 5 years with a one-year buffer using all datasets. A subset of top 15 features alone (out of a total of 963) could provide a test AUC of 79.1. In this paper, we provide extensive machine learning model performance and feature contribution analysis, which enables us to narrow down to the most important features useful for diabetes forecasting. Examples include chronic conditions such as asthma and hypertension, lab results, diagnostic codes in insurance claims, age and geographical information.

* 18 pages, 3 figures, 8 Tables, Submitted to 2019 ML for Healthcare conference 
Viaarxiv icon

ReGAN: RE[LAX|BAR|INFORCE] based Sequence Generation using GANs

May 08, 2018
Aparna Balagopalan, Satya Gorti, Mathieu Ravaut, Raeid Saqur

Figure 1 for ReGAN: RE[LAX|BAR|INFORCE] based Sequence Generation using GANs
Figure 2 for ReGAN: RE[LAX|BAR|INFORCE] based Sequence Generation using GANs
Figure 3 for ReGAN: RE[LAX|BAR|INFORCE] based Sequence Generation using GANs
Figure 4 for ReGAN: RE[LAX|BAR|INFORCE] based Sequence Generation using GANs

Generative Adversarial Networks (GANs) have seen steep ascension to the peak of ML research zeitgeist in recent years. Mostly catalyzed by its success in the domain of image generation, the technique has seen wide range of adoption in a variety of other problem domains. Although GANs have had a lot of success in producing more realistic images than other approaches, they have only seen limited use for text sequences. Generation of longer sequences compounds this problem. Most recently, SeqGAN (Yu et al., 2017) has shown improvements in adversarial evaluation and results with human evaluation compared to a MLE based trained baseline. The main contributions of this paper are three-fold: 1. We show results for sequence generation using a GAN architecture with efficient policy gradient estimators, 2. We attain improved training stability, and 3. We perform a comparative study of recent unbiased low variance gradient estimation techniques such as REBAR (Tucker et al., 2017), RELAX (Grathwohl et al., 2018) and REINFORCE (Williams, 1992). Using a simple grammar on synthetic datasets with varying length, we indicate the quality of sequences generated by the model.

Viaarxiv icon

Gradient descent revisited via an adaptive online learning rate

Apr 08, 2018
Mathieu Ravaut, Satya Gorti

Figure 1 for Gradient descent revisited via an adaptive online learning rate
Figure 2 for Gradient descent revisited via an adaptive online learning rate
Figure 3 for Gradient descent revisited via an adaptive online learning rate
Figure 4 for Gradient descent revisited via an adaptive online learning rate

Any gradient descent optimization requires to choose a learning rate. With deeper and deeper models, tuning that learning rate can easily become tedious and does not necessarily lead to an ideal convergence. We propose a variation of the gradient descent algorithm in the which the learning rate is not fixed. Instead, we learn the learning rate itself, either by another gradient descent (first-order method), or by Newton's method (second-order). This way, gradient descent for any machine learning algorithm can be optimized.

Viaarxiv icon

Truly Multi-modal YouTube-8M Video Classification with Video, Audio, and Text

Jul 10, 2017
Zhe Wang, Kingsley Kuan, Mathieu Ravaut, Gaurav Manek, Sibo Song, Yuan Fang, Seokhwan Kim, Nancy Chen, Luis Fernando D'Haro, Luu Anh Tuan, Hongyuan Zhu, Zeng Zeng, Ngai Man Cheung, Georgios Piliouras, Jie Lin, Vijay Chandrasekhar

Figure 1 for Truly Multi-modal YouTube-8M Video Classification with Video, Audio, and Text
Figure 2 for Truly Multi-modal YouTube-8M Video Classification with Video, Audio, and Text
Figure 3 for Truly Multi-modal YouTube-8M Video Classification with Video, Audio, and Text
Figure 4 for Truly Multi-modal YouTube-8M Video Classification with Video, Audio, and Text

The YouTube-8M video classification challenge requires teams to classify 0.7 million videos into one or more of 4,716 classes. In this Kaggle competition, we placed in the top 3% out of 650 participants using released video and audio features. Beyond that, we extend the original competition by including text information in the classification, making this a truly multi-modal approach with vision, audio and text. The newly introduced text data is termed as YouTube-8M-Text. We present a classification framework for the joint use of text, visual and audio features, and conduct an extensive set of experiments to quantify the benefit that this additional mode brings. The inclusion of text yields state-of-the-art results, e.g. 86.7% GAP on the YouTube-8M-Text validation dataset.

* 8 pages, Accepted to CVPR'17 Workshop on YouTube-8M Large-Scale Video Understanding 
Viaarxiv icon

Deep Learning for Lung Cancer Detection: Tackling the Kaggle Data Science Bowl 2017 Challenge

May 26, 2017
Kingsley Kuan, Mathieu Ravaut, Gaurav Manek, Huiling Chen, Jie Lin, Babar Nazir, Cen Chen, Tse Chiang Howe, Zeng Zeng, Vijay Chandrasekhar

Figure 1 for Deep Learning for Lung Cancer Detection: Tackling the Kaggle Data Science Bowl 2017 Challenge
Figure 2 for Deep Learning for Lung Cancer Detection: Tackling the Kaggle Data Science Bowl 2017 Challenge
Figure 3 for Deep Learning for Lung Cancer Detection: Tackling the Kaggle Data Science Bowl 2017 Challenge
Figure 4 for Deep Learning for Lung Cancer Detection: Tackling the Kaggle Data Science Bowl 2017 Challenge

We present a deep learning framework for computer-aided lung cancer diagnosis. Our multi-stage framework detects nodules in 3D lung CAT scans, determines if each nodule is malignant, and finally assigns a cancer probability based on these results. We discuss the challenges and advantages of our framework. In the Kaggle Data Science Bowl 2017, our framework ranked 41st out of 1972 teams.

Viaarxiv icon