TY - GEN
T1 - HELP
T2 - 8th Joint Conference on Lexical and Computational Semantics, *SEM@NAACL-HLT 2019
AU - Yanaka, Hitomi
AU - Mineshima, Koji
AU - Bekki, Daisuke
AU - Inui, Kentaro
AU - Sekine, Satoshi
AU - Abzianidze, Lasha
AU - Bos, Johan
N1 - Funding Information:
We thank our three anonymous reviewers for helpful suggestions.
Publisher Copyright:
© 2019 Association for Computational Linguistics
PY - 2019
Y1 - 2019
N2 - Large crowdsourced datasets are widely used for training and evaluating neural models on natural language inference (NLI). Despite these efforts, neural models have a hard time capturing logical inferences, including those licensed by phrase replacements, so-called monotonicity reasoning. Since no large dataset has been developed for monotonicity reasoning, it is still unclear whether the main obstacle is the size of datasets or the model architectures themselves. To investigate this issue, we introduce a new dataset, called HELP, for handling entailments with lexical and logical phenomena. We add it to training data for the state-of-the-art neural models and evaluate them on test sets for monotonicity phenomena. The results showed that our data augmentation improved the overall accuracy. We also find that the improvement is better on monotonicity inferences with lexical replacements than on downward inferences with disjunction and modification. This suggests that some types of inferences can be improved by our data augmentation while others are immune to it.
AB - Large crowdsourced datasets are widely used for training and evaluating neural models on natural language inference (NLI). Despite these efforts, neural models have a hard time capturing logical inferences, including those licensed by phrase replacements, so-called monotonicity reasoning. Since no large dataset has been developed for monotonicity reasoning, it is still unclear whether the main obstacle is the size of datasets or the model architectures themselves. To investigate this issue, we introduce a new dataset, called HELP, for handling entailments with lexical and logical phenomena. We add it to training data for the state-of-the-art neural models and evaluate them on test sets for monotonicity phenomena. The results showed that our data augmentation improved the overall accuracy. We also find that the improvement is better on monotonicity inferences with lexical replacements than on downward inferences with disjunction and modification. This suggests that some types of inferences can be improved by our data augmentation while others are immune to it.
UR - http://www.scopus.com/inward/record.url?scp=85094391653&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85094391653&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85094391653
T3 - *SEM@NAACL-HLT 2019 - 8th Joint Conference on Lexical and Computational Semantics
SP - 250
EP - 255
BT - *SEM@NAACL-HLT 2019 - 8th Joint Conference on Lexical and Computational Semantics
PB - Association for Computational Linguistics (ACL)
Y2 - 6 June 2019 through 7 June 2019
ER -