參考文獻 |
[1] D. Chen, A. Fisch, J. Weston, and A. Bordes, “Reading Wikipedia to Answer Open-Domain Questions,” arXiv:1704.00051 [cs], Apr. 2017, Accessed: Dec. 22, 2020. [Online]. Available: http://arxiv.org/abs/1704.00051.
[2] W. Yang et al., “End-to-End Open-Domain Question Answering with BERTserini,” Proceedings of the 2019 Conference of the North, pp. 72–77, 2019, doi: 10.18653/v1/N19-4013.
[3] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,” arXiv:1810.04805 [cs], May 2019, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1810.04805.
[4] M. Seo, A. Kembhavi, A. Farhadi, and H. Hajishirzi, “Bidirectional Attention Flow for Machine Comprehension,” arXiv:1611.01603 [cs], Jun. 2018, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1611.01603.
[5] A. W. Yu et al., “QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension,” arXiv:1804.09541 [cs], Apr. 2018, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1804.09541.
[6] A. Vaswani et al., “Attention Is All You Need,” arXiv:1706.03762 [cs], Dec. 2017, Accessed: Jan. 05, 2021. [Online]. Available: http://arxiv.org/abs/1706.03762.
[7] D. Bahdanau, K. Cho, and Y. Bengio, “Neural Machine Translation by Jointly Learning to Align and Translate,” arXiv:1409.0473 [cs, stat], May 2016, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1409.0473.
[8] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, “Improving Language Understanding by Generative Pre-Training,” p. 12.
[9] T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient Estimation of Word Representations in Vector Space,” arXiv:1301.3781 [cs], Sep. 2013, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1301.3781.
[10] Y. Jernite, S. R. Bowman, and D. Sontag, “Discourse-Based Objectives for Fast Unsupervised Sentence Representation Learning,” arXiv:1705.00557 [cs, stat], Apr. 2017, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1705.00557.
[11] L. Logeswaran and H. Lee, “AN EFFICIENT FRAMEWORK FOR LEARNING SENTENCE REPRESENTATIONS,” p. 16.
[12] A. Williams, N. Nangia, and S. Bowman, “A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference,” in Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), New Orleans, Louisiana, 2018, pp. 1112–1122, doi: 10.18653/v1/N18-1101.
[13] L. Sharma, L. Graesser, N. Nangia, and U. Evci, “Natural Language Understanding with the Quora Question Pairs Dataset,” p. 10.
[14] R. Socher et al., “Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank,” p. 12.
[15] A. Warstadt, A. Singh, and S. R. Bowman, “Neural Network Acceptability Judgments,” arXiv:1805.12471 [cs], Oct. 2019, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1805.12471.
[16] P. Rajpurkar, J. Zhang, K. Lopyrev, and P. Liang, “SQuAD: 100,000+ Questions for Machine Comprehension of Text,” arXiv:1606.05250 [cs], Oct. 2016, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1606.05250.
[17] P. Rajpurkar, R. Jia, and P. Liang, “Know What You Don’t Know: Unanswerable Questions for SQuAD,” arXiv:1806.03822 [cs], Jun. 2018, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1806.03822.
[18] E. F. T. K. Sang and F. De Meulder, “Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition,” arXiv:cs/0306050, Jun. 2003, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/cs/0306050.
[19] Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut, “ALBERT: A Lite BERT for Self-supervised Learning of Language Representations,” arXiv:1909.11942 [cs], Feb. 2020, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1909.11942.
[20] M. Joshi, D. Chen, Y. Liu, D. S. Weld, L. Zettlemoyer, and O. Levy, “SpanBERT: Improving Pre-training by Representing and Predicting Spans,” arXiv:1907.10529 [cs], Jan. 2020, Accessed: May 27, 2020. [Online]. Available: http://arxiv.org/abs/1907.10529.
[21] V. Sanh, L. Debut, J. Chaumond, and T. Wolf, “DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter,” arXiv:1910.01108 [cs], Feb. 2020, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1910.01108.
[22] Z. Zhang, X. Han, Z. Liu, X. Jiang, M. Sun, and Q. Liu, “ERNIE: Enhanced Language Representation with Informative Entities,” in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, Jul. 2019, pp. 1441–1451, doi: 10.18653/v1/P19-1139.
[23] thunlp/PLMpapers. THUNLP, 2021.
[24] M. E. Peters et al., “Deep contextualized word representations,” arXiv:1802.05365 [cs], Mar. 2018, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1802.05365.
[25] K. Bollacker, C. Evans, P. Paritosh, T. Sturge, and J. Taylor, “Freebase: a collaboratively created graph database for structuring human knowledge,” in In SIGMOD Conference, 2008, pp. 1247–1250.
[26] S. Auer, C. Bizer, G. Kobilarov, J. Lehmann, R. Cyganiak, and Z. Ives, “DBpedia: A Nucleus for a Web of Open Data,” in The Semantic Web, vol. 4825, K. Aberer, K.-S. Choi, N. Noy, D. Allemang, K.-I. Lee, L. Nixon, J. Golbeck, P. Mika, D. Maynard, R. Mizoguchi, G. Schreiber, and P. Cudré-Mauroux, Eds. Berlin, Heidelberg: Springer Berlin Heidelberg, 2007, pp. 722–735.
[27] D. Ferrucci et al., “Building Watson: An Overview of the DeepQA Project,” AIMag, vol. 31, no. 3, Art. no. 3, Jul. 2010, doi: 10.1609/aimag.v31i3.2303.
[28] K. M. Hermann et al., “Teaching Machines to Read and Comprehend,” arXiv:1506.03340 [cs], Nov. 2015, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1506.03340.
[29] J. Pennington, R. Socher, and C. Manning, “GloVe: Global Vectors for Word Representation,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, Oct. 2014, pp. 1532–1543, doi: 10.3115/v1/D14-1162.
[30] P. Baudisˇ, “YodaQA: A Modular Question Answering System Pipeline,” p. 8, 2015.
[31] J. Berant, A. Chou, R. Frostig, and P. Liang, “Semantic Parsing on Freebase from Question-Answer Pairs,” in Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, Washington, USA, Oct. 2013, pp. 1533–1544, Accessed: May 03, 2021. [Online]. Available: https://www.aclweb.org/anthology/D13-1160.
[32] A. Miller, A. Fisch, J. Dodge, A.-H. Karimi, A. Bordes, and J. Weston, “Key-Value Memory Networks for Directly Reading Documents,” arXiv:1606.03126 [cs], Oct. 2016, Accessed: May 03, 2021. [Online]. Available: http://arxiv.org/abs/1606.03126.
[33] V. Karpukhin et al., “Dense Passage Retrieval for Open-Domain Question Answering,” arXiv:2004.04906 [cs], Sep. 2020, Accessed: Dec. 22, 2020. [Online]. Available: http://arxiv.org/abs/2004.04906.
[34] facebookresearch/faiss. Facebook Research, 2021.
[35] Y. Liu et al., “RoBERTa: A Robustly Optimized BERT Pretraining Approach,” arXiv:1907.11692 [cs], Jul. 2019, Accessed: May 03, 2021. [Online]. Available: http://arxiv.org/abs/1907.11692.
[36] C. C. Shao, T. Liu, Y. Lai, Y. Tseng, and S. Tsai, “DRCD: a Chinese Machine Reading Comprehension Dataset,” p. 5.
[37] Y. Cui et al., “A Span-Extraction Dataset for Chinese Machine Reading Comprehension,” Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 5882–5888, 2019, doi: 10.18653/v1/D19-1600.
[38] P. Li et al., “Dataset and Neural Recurrent Sequence Labeling Model for Open-Domain Factoid Question Answering,” arXiv:1607.06275 [cs], Sep. 2016, Accessed: May 03, 2021. [Online]. Available: http://arxiv.org/abs/1607.06275.
[39] K. Lee, M.-W. Chang, and K. Toutanova, “Latent Retrieval for Weakly Supervised Open Domain Question Answering,” arXiv:1906.00300 [cs], Jun. 2019, Accessed: Apr. 29, 2021. [Online]. Available: http://arxiv.org/abs/1906.00300.
[40] Y. Cui et al., “Pre-Training with Whole Word Masking for Chinese BERT,” arXiv:1906.08101 [cs], Oct. 2019, Accessed: May 03, 2021. [Online]. Available: http://arxiv.org/abs/1906.08101.
[41] G. Attardi, attardi/wikiextractor. 2021.
[42] C. Kuo, BYVoid/OpenCC. 2021.
[43] G. Häring, ghaering/pysqlite. 2021.
[44] S. Junyi, fxsjy/jieba. 2021.
[45] C. Raffel et al., “Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer,” arXiv:1910.10683 [cs, stat], Jul. 2020, Accessed: May 03, 2021. [Online]. Available: http://arxiv.org/abs/1910.10683. |