Picture for Simon Mille

Simon Mille

The QCET Taxonomy of Standard Quality Criterion Names and Definitions for the Evaluation of NLP Systems

Add code
Sep 26, 2025
Viaarxiv icon

On the Role of Summary Content Units in Text Summarization Evaluation

Add code
Apr 02, 2024
Figure 1 for On the Role of Summary Content Units in Text Summarization Evaluation
Figure 2 for On the Role of Summary Content Units in Text Summarization Evaluation
Figure 3 for On the Role of Summary Content Units in Text Summarization Evaluation
Figure 4 for On the Role of Summary Content Units in Text Summarization Evaluation
Viaarxiv icon

Needle in a Haystack: An Analysis of Finding Qualified Workers on MTurk for Summarization

Add code
Dec 28, 2022
Figure 1 for Needle in a Haystack: An Analysis of Finding Qualified Workers on MTurk for Summarization
Figure 2 for Needle in a Haystack: An Analysis of Finding Qualified Workers on MTurk for Summarization
Figure 3 for Needle in a Haystack: An Analysis of Finding Qualified Workers on MTurk for Summarization
Figure 4 for Needle in a Haystack: An Analysis of Finding Qualified Workers on MTurk for Summarization
Viaarxiv icon

GEMv2: Multilingual NLG Benchmarking in a Single Line of Code

Add code
Jun 24, 2022
Figure 1 for GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
Figure 2 for GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
Figure 3 for GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
Figure 4 for GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
Viaarxiv icon

Quantified Reproducibility Assessment of NLP Results

Add code
Apr 12, 2022
Figure 1 for Quantified Reproducibility Assessment of NLP Results
Figure 2 for Quantified Reproducibility Assessment of NLP Results
Figure 3 for Quantified Reproducibility Assessment of NLP Results
Figure 4 for Quantified Reproducibility Assessment of NLP Results
Viaarxiv icon

NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation

Add code
Dec 06, 2021
Figure 1 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Figure 2 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Figure 3 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Figure 4 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Viaarxiv icon

Automatic Construction of Evaluation Suites for Natural Language Generation Datasets

Add code
Jun 16, 2021
Figure 1 for Automatic Construction of Evaluation Suites for Natural Language Generation Datasets
Figure 2 for Automatic Construction of Evaluation Suites for Natural Language Generation Datasets
Figure 3 for Automatic Construction of Evaluation Suites for Natural Language Generation Datasets
Figure 4 for Automatic Construction of Evaluation Suites for Natural Language Generation Datasets
Viaarxiv icon

Assessing the Syntactic Capabilities of Transformer-based Multilingual Language Models

Add code
May 10, 2021
Figure 1 for Assessing the Syntactic Capabilities of Transformer-based Multilingual Language Models
Figure 2 for Assessing the Syntactic Capabilities of Transformer-based Multilingual Language Models
Figure 3 for Assessing the Syntactic Capabilities of Transformer-based Multilingual Language Models
Figure 4 for Assessing the Syntactic Capabilities of Transformer-based Multilingual Language Models
Viaarxiv icon

The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics

Add code
Feb 03, 2021
Figure 1 for The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics
Figure 2 for The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics
Figure 3 for The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics
Figure 4 for The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics
Viaarxiv icon