Natural Language Understanding Wiki
(refs)
Tag: Visual edit
 
(add link)
Tag: Visual edit
 
(One intermediate revision by the same user not shown)
Line 1: Line 1:
SQuAD is a line of question-answering datasets created by Stanford. The first incarnation is published in Rajpurkar et al. (2016)<ref>Rajpurkar, P., Zhang, J., Lopyrev, K., & Liang, P. (2016). SQuAD: 100,000+ Questions for Machine Comprehension of Text. In ''EMNLP 2016'' (pp. 2383–2392). http://doi.org/10.18653/v1/D16-1264</ref> and quickly became popular. However, results on this dataset quickly surpass human performance with the application of what Percy Liang has [https://newgeneralization.github.io/slides/PercyLiang.pdf called] "cheap tricks". Adversarial SQuAD (Jia and Liang, 2017<ref>Jia, R., & Liang, P. (2017). Adversarial Examples for Evaluating Reading Comprehension Systems. ''EMNLP 2017'', 2021–2031. Retrieved from http://arxiv.org/abs/1707.07328</ref>) and SQuAD 2.0 (Rajpurkar et al. 2018<ref>Rajpurkar, P., Jia, R., & Liang, P. (2018). Know What You Don’t Know: Unanswerable Questions for SQuAD. In ''Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)'' (pp. 784–789). Association for Computational Linguistics.</ref>) are created to evaluate for higher inference skills.
+
[https://rajpurkar.github.io/SQuAD-explorer/ SQuAD] is a line of question-answering datasets created by Stanford. The first incarnation is published in Rajpurkar et al. (2016)<ref>Rajpurkar, P., Zhang, J., Lopyrev, K., & Liang, P. (2016). SQuAD: 100,000+ Questions for Machine Comprehension of Text. In ''EMNLP 2016'' (pp. 2383–2392). http://doi.org/10.18653/v1/D16-1264</ref> and quickly became popular. However, results on this dataset quickly surpass human performance with the application of what Percy Liang has [https://newgeneralization.github.io/slides/PercyLiang.pdf called] "cheap tricks". Adversarial SQuAD (Jia and Liang, 2017<ref>Jia, R., & Liang, P. (2017). Adversarial Examples for Evaluating Reading Comprehension Systems. ''EMNLP 2017'', 2021–2031. Retrieved from http://arxiv.org/abs/1707.07328</ref>) and SQuAD 2.0 (Rajpurkar et al. 2018<ref>Rajpurkar, P., Jia, R., & Liang, P. (2018). Know What You Don’t Know: Unanswerable Questions for SQuAD. In ''Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)'' (pp. 784–789). Association for Computational Linguistics.</ref>) are created to evaluate for higher inference skills.
  +
  +
== Open-source software packages ==
  +
* [https://allennlp.org/models AllenNLP]
  +
* [https://github.com/jojonki/BiDAF BiDAF]
   
 
== References ==
 
== References ==

Latest revision as of 09:44, 11 August 2018

SQuAD is a line of question-answering datasets created by Stanford. The first incarnation is published in Rajpurkar et al. (2016)[1] and quickly became popular. However, results on this dataset quickly surpass human performance with the application of what Percy Liang has called "cheap tricks". Adversarial SQuAD (Jia and Liang, 2017[2]) and SQuAD 2.0 (Rajpurkar et al. 2018[3]) are created to evaluate for higher inference skills.

Open-source software packages[]

References[]

  1. Rajpurkar, P., Zhang, J., Lopyrev, K., & Liang, P. (2016). SQuAD: 100,000+ Questions for Machine Comprehension of Text. In EMNLP 2016 (pp. 2383–2392). http://doi.org/10.18653/v1/D16-1264
  2. Jia, R., & Liang, P. (2017). Adversarial Examples for Evaluating Reading Comprehension Systems. EMNLP 2017, 2021–2031. Retrieved from http://arxiv.org/abs/1707.07328
  3. Rajpurkar, P., Jia, R., & Liang, P. (2018). Know What You Don’t Know: Unanswerable Questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (pp. 784–789). Association for Computational Linguistics.