Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension

Type: Article

Publication Date: 2020-11-12

Citations: 114

DOI: https://doi.org/10.1162/tacl_a_00338

Abstract

Innovations in annotation methodology have been a catalyst for Reading Comprehension (RC) datasets and models. One recent trend to challenge current RC models is to involve a model in the annotation process: humans create questions adversarially, such that the model fails to answer them correctly. In this work we investigate this annotation methodology and apply it in three different settings, collecting a total of 36,000 samples with progressively stronger models in the annotation loop. This allows us to explore questions such as the reproducibility of the adversarial effect, transfer from data collected with varying model-in-the-loop strengths, and generalisation to data collected without a model. We find that training on adversarially collected samples leads to strong generalisation to non-adversarially collected datasets, yet with progressive performance deterioration with increasingly stronger models-in-the-loop. Furthermore, we find that stronger models can still learn from datasets collected with substantially weaker models-in-the-loop. When trained on data collected with a BiDAF model in the loop, RoBERTa achieves 39.9F1 on questions that it cannot answer when trained on SQuAD - only marginally lower than when trained on data collected using RoBERTa itself (41.0F1).

Locations

  • Transactions of the Association for Computational Linguistics - View - PDF
  • UCL Discovery (University College London) - View - PDF
  • arXiv (Cornell University) - View - PDF
  • DataCite API - View

Similar Works

Action Title Year Authors
+ Collecting high-quality adversarial data for machine reading comprehension tasks with humans and models in the loop 2022 Damian Yukio Romero Diaz
Magdalena Anioł
John Culnan
+ Collecting high-quality adversarial data for machine reading comprehension tasks with humans and models in the loop 2022 Damian Yukio Romero Diaz
Magdalena Anioł
John Culnan
+ Adversarial Examples for Evaluating Reading Comprehension Systems 2017 Robin Jia
Percy Liang
+ PDF Chat Adversarial Examples for Evaluating Reading Comprehension Systems 2017 Robin Jia
Percy Liang
+ MRQA 2019 Shared Task: Evaluating Generalization in Reading Comprehension 2019 Adam Fisch
Alon Talmor
Robin Jia
Minjoon Seo
Eunsol Choi
Danqi Chen
+ MRQA 2019 Shared Task: Evaluating Generalization in Reading Comprehension 2019 Adam Fisch
Alon Talmor
Robin Jia
Minjoon Seo
Eunsol Choi
Danqi Chen
+ Frustratingly Poor Performance of Reading Comprehension Models on Non-adversarial Examples. 2019 Soham Parikh
Ananya B. Sai
Preksha Nema
Mitesh M. Khapra
+ The Impacts of Unanswerable Questions on the Robustness of Machine Reading Comprehension Models 2023 Son Quoc Tran
Phong Nguyen-Thuan Do
Uyen Le
Matt Kretchmar
+ Improving Machine Reading Comprehension via Adversarial Training 2019 Ziqing Yang
Yiming Cui
Wanxiang Che
Ting Liu
Shijin Wang
Guoping Hu
+ Improving Machine Reading Comprehension via Adversarial Training. 2019 Ziqing Yang
Yiming Cui
Wanxiang Che
Ting Liu
Shijin Wang
Guoping Hu
+ The Impacts of Unanswerable Questions on the Robustness of Machine Reading Comprehension Models 2023 Son Quoc Tran
Phong Nguyen-Thuan
Uyen Le
Matt Kretchmar
+ Frustratingly Poor Performance of Reading Comprehension Models on Non-adversarial Examples 2019 Soham Parikh
Ananya B. Sai
Preksha Nema
Mitesh M. Khapra
+ Using Adversarial Attacks to Reveal the Statistical Bias in Machine Reading Comprehension Models 2021 Jieyu Lin
Jiajie Zou
Nai Ding
+ Benchmarking Robustness of Machine Reading Comprehension Models 2020 Chenglei Si
Ziqing Yang
Yiming Cui
Wentao Ma
Ting Liu
Shijin Wang
+ Power in Numbers: Robust reading comprehension by finetuning with four adversarial sentences per example 2024 Ariel Marcus
+ Benchmarking Robustness of Machine Reading Comprehension Models 2021 Chenglei Si
Ziqing Yang
Yiming Cui
Wentao Ma
Ting Liu
Shijin Wang
+ Undersensitivity in Neural Reading Comprehension 2020 Johannes Welbl
Pasquale Minervini
Max Bartolo
Pontus Stenetorp
Sebastian Riedel
+ Robust Machine Comprehension Models via Adversarial Training 2018 Yicheng Wang
Mohit Bansal
+ Adversarial Augmentation Policy Search for Domain and Cross-Lingual Generalization in Reading Comprehension 2020 Adyasha Maharana
Mohit Bansal
+ Adversarial Augmentation Policy Search for Domain and Cross-Lingual Generalization in Reading Comprehension 2020 Adyasha Maharana
Mohit Bansal

Works That Cite This (80)

Action Title Year Authors
+ PDF Chat A survey of human-in-the-loop for machine learning 2022 Xingjiao Wu
Luwei Xiao
Yixuan Sun
Junhang Zhang
Tianlong Ma
Liang He
+ PDF Chat <i>Did Aristotle Use a Laptop?</i>A Question Answering Benchmark with Implicit Reasoning Strategies 2021 Mor Geva
Daniel Khashabi
Elad Segal
Tushar Khot
Dan Roth
Jonathan Berant
+ Collecting high-quality adversarial data for machine reading comprehension tasks with humans and models in the loop 2022 Damian Yukio Romero Diaz
Magdalena Anioł
John Culnan
+ Dynatask: A Framework for Creating Dynamic AI Benchmark Tasks 2022 Tristan Thrush
Kushal Tirumala
Anmol Gupta
Max Bartolo
Pedro RodrĂ­guez
Tariq Kane
William Gaviria Rojas
Peter Mattson
Adina Williams
Douwe Kiela
+ Let’s Synthesize Step by Step: Iterative Dataset Synthesis with Large Language Models by Extrapolating Errors from Small Models 2023 Ruida Wang
Wangchunshu Zhou
Mrinmaya Sachan
+ PDF Chat Learning What Makes a Difference from Counterfactual Examples and Gradient Supervision 2020 Damien Teney
Ehsan Abbasnedjad
Anton van den Hengel
+ Did Aristotle Use a Laptop? A Question Answering Benchmark with Implicit Reasoning Strategies 2021 Mor Geva
Daniel Khashabi
Elad Segal
Tushar Khot
Dan Roth
Jonathan Berant
+ Mind the instructions: a holistic evaluation of consistency and interactions in prompt-based learning 2023 Lucas Weber
Elia Bruni
Dieuwke Hupkes
+ PDF Chat ZEROTOP: Zero-Shot Task-Oriented Semantic Parsing using Large Language Models 2023 Dheeraj Mekala
Jason Wolfe
Subhro Roy
+ Human-Adversarial Visual Question Answering 2021 Sasha Sheng
Amanpreet Singh
Vedanuj Goswami
Jose Alberto Lopez Magana
Wojciech Galuba
Devi Parikh
Douwe Kiela