SQuAD is a line of question-answering datasets created by Stanford. The first incarnation is published in Rajpurkar et al. (2016) and quickly became popular. However, results on this dataset quickly surpass human performance with the application of what Percy Liang has called "cheap tricks". Adversarial SQuAD (Jia and Liang, 2017) and SQuAD 2.0 (Rajpurkar et al. 2018) are created to evaluate for higher inference skills.
Open-source software packages Edit
- ↑ Rajpurkar, P., Zhang, J., Lopyrev, K., & Liang, P. (2016). SQuAD: 100,000+ Questions for Machine Comprehension of Text. In EMNLP 2016 (pp. 2383–2392). http://doi.org/10.18653/v1/D16-1264
- ↑ Jia, R., & Liang, P. (2017). Adversarial Examples for Evaluating Reading Comprehension Systems. EMNLP 2017, 2021–2031. Retrieved from http://arxiv.org/abs/1707.07328
- ↑ Rajpurkar, P., Jia, R., & Liang, P. (2018). Know What You Don’t Know: Unanswerable Questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (pp. 784–789). Association for Computational Linguistics.