MiRa
Collection
8 items
•
Updated
input
stringlengths 20
53
| reference
stringlengths 3
9
|
|---|---|
With proper equipment, scuba-diving is very
|
safe
|
With proper equipment, scuba-diving isn't very
|
safe
|
With proper equipment, scuba-diving is very
|
dangerous
|
With proper equipment, scuba-diving isn't very
|
dangerous
|
Traveling in Baghdad is very
|
dangerous
|
Traveling in Baghdad isn't very
|
dangerous
|
Traveling in Baghdad is very
|
safe
|
Traveling in Baghdad isn't very
|
safe
|
In moderation, drinking red wine is
|
good
|
In moderation, drinking red wine isn't
|
good
|
In moderation, drinking red wine is
|
bad
|
In moderation, drinking red wine isn't
|
bad
|
Drinking tap water in developing countries is very
|
dangerous
|
Drinking tap water in developing countries isn't very
|
dangerous
|
Drinking tap water in developing countries is very
|
safe
|
Drinking tap water in developing countries isn't very
|
safe
|
Secondhand furniture is very
|
cheap
|
Secondhand furniture isn't very
|
cheap
|
Secondhand furniture is very
|
expensive
|
Secondhand furniture isn't very
|
expensive
|
Most smokers find that quitting is very
|
difficult
|
Most smokers find that quitting isn't very
|
difficult
|
Most smokers find that quitting is very
|
easy
|
Most smokers find that quitting isn't very
|
easy
|
Old computers may be
|
slow
|
Old computers may not be
|
slow
|
Old computers may be
|
fast
|
Old computers may not be
|
fast
|
A fast food dinner on a first date is very
|
lame
|
A fast food dinner on a first date isn't very
|
lame
|
A fast food dinner on a first date is very
|
romantic
|
A fast food dinner on a first date isn't very
|
romantic
|
Bulletproof vests are very
|
safe
|
Bulletproof vests aren't very
|
safe
|
Bulletproof vests are very
|
dangerous
|
Bulletproof vests aren't very
|
dangerous
|
Terrorist bomb attacks are really
|
dangerous
|
Terrorist bomb attacks aren't really
|
dangerous
|
Terrorist bomb attacks are really
|
safe
|
Terrorist bomb attacks aren't really
|
safe
|
Vitamins and proteins are very
|
good
|
Vitamins and proteins aren't very
|
good
|
Vitamins and proteins are very
|
bad
|
Vitamins and proteins aren't very
|
bad
|
Using strong suntan lotion is
|
safe
|
Using strong suntan lotion isn't
|
safe
|
Using strong suntan lotion is
|
dangerous
|
Using strong suntan lotion isn't
|
dangerous
|
Rockets and missiles are very
|
fast
|
Rockets and missiles aren't very
|
fast
|
Rockets and missiles are very
|
slow
|
Rockets and missiles aren't very
|
slow
|
Keeping the door open for somebody is very
|
polite
|
Keeping the door open for somebody isn't very
|
polite
|
Keeping the door open for somebody is very
|
rude
|
Keeping the door open for somebody isn't very
|
rude
|
A baby bunny's fur is very
|
soft
|
A baby bunny's fur isn't very
|
soft
|
A baby bunny's fur is very
|
hard
|
A baby bunny's fur isn't very
|
hard
|
Businessman Donald Trump is really
|
rich
|
Businessman Donald Trump isn't really
|
rich
|
Businessman Donald Trump is really
|
poor
|
Businessman Donald Trump isn't really
|
poor
|
This repository contains a diagnostic dataset (cprag) for What BERT is not: Lessons from a new suite of psycholinguistic diagnostics for language models, by Allyson Ettinger.
The dataset is released under the MIT License.
@article{10.1162/tacl_a_00298,
author = {Ettinger, Allyson},
title = {What BERT Is Not: Lessons from a New Suite of Psycholinguistic Diagnostics for Language Models},
journal = {Transactions of the Association for Computational Linguistics},
volume = {8},
pages = {34-48},
year = {2020},
month = {01},
abstract = {Pre-training by language modeling has become a popular and successful approach to NLP tasks, but we have yet to understand exactly what linguistic capacities these pre-training processes confer upon models. In this paper we introduce a suite of diagnostics drawn from human language experiments, which allow us to ask targeted questions about information used by language models for generating predictions in context. As a case study, we apply these diagnostics to the popular BERT model, finding that it can generally distinguish good from bad completions involving shared category or role reversal, albeit with less sensitivity than humans, and it robustly retrieves noun hypernyms, but it struggles with challenging inference and role-based event prediction— and, in particular, it shows clear insensitivity to the contextual impacts of negation.},
issn = {2307-387X},
doi = {10.1162/tacl_a_00298},
url = {https://doi.org/10.1162/tacl_a_00298},
eprint = {https://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl_a_00298/1923116/tacl_a_00298.pdf},
}