Picture for Nikita Nangia

Nikita Nangia

Shammie

Does Putting a Linguist in the Loop Improve NLU Data Collection?

Add code
Apr 15, 2021
Figure 1 for Does Putting a Linguist in the Loop Improve NLU Data Collection?
Figure 2 for Does Putting a Linguist in the Loop Improve NLU Data Collection?
Figure 3 for Does Putting a Linguist in the Loop Improve NLU Data Collection?
Figure 4 for Does Putting a Linguist in the Loop Improve NLU Data Collection?
Viaarxiv icon

CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models

Add code
Sep 30, 2020
Figure 1 for CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models
Figure 2 for CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models
Figure 3 for CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models
Figure 4 for CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models
Viaarxiv icon

Natural Language Understanding with the Quora Question Pairs Dataset

Add code
Jul 01, 2019
Figure 1 for Natural Language Understanding with the Quora Question Pairs Dataset
Figure 2 for Natural Language Understanding with the Quora Question Pairs Dataset
Figure 3 for Natural Language Understanding with the Quora Question Pairs Dataset
Figure 4 for Natural Language Understanding with the Quora Question Pairs Dataset
Viaarxiv icon

Human vs. Muppet: A Conservative Estimate of Human Performance on the GLUE Benchmark

Add code
Jun 01, 2019
Figure 1 for Human vs. Muppet: A Conservative Estimate of Human Performance on the GLUE Benchmark
Figure 2 for Human vs. Muppet: A Conservative Estimate of Human Performance on the GLUE Benchmark
Figure 3 for Human vs. Muppet: A Conservative Estimate of Human Performance on the GLUE Benchmark
Viaarxiv icon

SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems

Add code
May 02, 2019
Figure 1 for SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems
Figure 2 for SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems
Figure 3 for SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems
Figure 4 for SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems
Viaarxiv icon

ListOps: A Diagnostic Dataset for Latent Tree Learning

Add code
Apr 17, 2018
Figure 1 for ListOps: A Diagnostic Dataset for Latent Tree Learning
Figure 2 for ListOps: A Diagnostic Dataset for Latent Tree Learning
Figure 3 for ListOps: A Diagnostic Dataset for Latent Tree Learning
Figure 4 for ListOps: A Diagnostic Dataset for Latent Tree Learning
Viaarxiv icon

A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference

Add code
Feb 19, 2018
Figure 1 for A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference
Figure 2 for A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference
Figure 3 for A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference
Figure 4 for A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference
Viaarxiv icon

The RepEval 2017 Shared Task: Multi-Genre Natural Language Inference with Sentence Representations

Add code
Jul 25, 2017
Figure 1 for The RepEval 2017 Shared Task: Multi-Genre Natural Language Inference with Sentence Representations
Figure 2 for The RepEval 2017 Shared Task: Multi-Genre Natural Language Inference with Sentence Representations
Figure 3 for The RepEval 2017 Shared Task: Multi-Genre Natural Language Inference with Sentence Representations
Figure 4 for The RepEval 2017 Shared Task: Multi-Genre Natural Language Inference with Sentence Representations
Viaarxiv icon