參考文獻 |
中文部分:
蔡汶霖(2018)以詞向量模型增進基於遞歸神經網路之中文文字摘要系統效能。
麥嘉芳 (2019) 基於注意力機制之詞向量中文萃取式摘要研究。
英文部分:
Bahdanau, D., Cho, K. & Bengio, Y. (2016) Neural Machine Translation by Jointly Learning to Align and Translate. arXiv:1409.0473 [cs, stat]. [Online] Available from: http://arxiv.org/abs/1409.0473 [Accessed: 26 March 2020].
Cao, Z., Li, W., Li, S., Wei, F., et al. (2016) AttSum: Joint Learning of Focusing and Summarization with Neural Attention. arXiv:1604.00125 [cs]. [Online] Available from: http://arxiv.org/abs/1604.00125 [Accessed: 25 January 2020].
Chen, Y.-C. & Bansal, M. (2018) Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting. arXiv:1805.11080 [cs]. [Online] Available from: http://arxiv.org/abs/1805.11080 [Accessed: 9 December 2019].
Chung, J., Gulcehre, C., Cho, K. & Bengio, Y. (2014) Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv:1412.3555 [cs]. [Online] Available from: http://arxiv.org/abs/1412.3555 [Accessed: 20 February 2020].
Elman, J.L. (1990) Finding Structure in Time. Cognitive Science. [Online] 14 (2), 179–211. Available from: doi:10.1207/s15516709cog1402_1.
Hochreiter, S. & Schmidhuber, J. (1997) Long Short-Term Memory. Neural Computation 9(8).
Hsieh, Y.-L., Liu, S.-H., Chen, K.-Y., Wang, H.-M., et al. (2016) 運用序列到序列生成架構於重寫式自動摘要(Exploiting Sequence-to-Sequence Generation Framework for Automatic Abstractive Summarization)[In Chinese]. In: Proceedings of the 28th Conference on Computational Linguistics and Speech Processing (ROCLING 2016). [Online]. October 2016 Tainan, Taiwan, The Association for Computational Linguistics and Chinese Language Processing (ACLCLP). pp. 115–128. Available from: https://www.aclweb.org/anthology/O16-1012 [Accessed: 26 March 2020].
Hu, B., Chen, Q. & Zhu, F. (2016) LCSTS: A Large Scale Chinese Short Text Summarization Dataset. arXiv:1506.05865 [cs]. [Online] Available from: http://arxiv.org/abs/1506.05865 [Accessed: 26 March 2020].
Joulin, A., Grave, E., Bojanowski, P., Nickel, M., et al. (2017) Fast Linear Model for Knowledge Graph Embeddings. arXiv:1710.10881 [cs, stat]. [Online] Available from: http://arxiv.org/abs/1710.10881 [Accessed: 6 January 2020].
LeCun, Y., Bengio, Y. & Hinton, G. (2015) Deep learning. Nature. [Online] 521 (7553), 436–444. Available from: doi:10.1038/nature14539.
Lin, C.-Y. (2004) ROUGE: A Package for Automatic Evaluation of Summaries. In: Text Summarization Branches Out. [Online]. July 2004 Barcelona, Spain, Association for Computational Linguistics. pp. 74–81. Available from: https://www.aclweb.org/anthology/W04-1013 [Accessed: 5 March 2020].
Liu, P.J., Saleh, M., Pot, E., Goodrich, B., et al. (2018) Generating Wikipedia by Summarizing Long Sequences. arXiv:1801.10198 [cs]. [Online] Available from: http://arxiv.org/abs/1801.10198 [Accessed: 26 March 2020].
Liu, Y. (2019) Fine-tune BERT for Extractive Summarization. arXiv:1903.10318 [cs]. [Online] Available from: http://arxiv.org/abs/1903.10318 [Accessed: 26 March 2020].
Liu, Y. & Lapata, M. (2019) Text Summarization with Pretrained Encoders. arXiv:1908.08345 [cs]. [Online] Available from: http://arxiv.org/abs/1908.08345 [Accessed: 21 May 2020].
McDonald, D.M. & Chen, H. (2006) Summary in context: Searching versus browsing. ACM Transactions on Information Systems (TOIS). [Online] 24 (1), 111–141. Available from: doi:10.1145/1125857.1125861.
McKeown, K. & Radev, D.R. (1995) Generating summaries of multiple news articles. In: Proceedings of the 18th annual international ACM SIGIR conference on Research and development in information retrieval - SIGIR ’95. [Online]. 1995 Seattle, Washington, United States, ACM Press. pp. 74–82. Available from: doi:10.1145/215206.215334 [Accessed: 22 May 2020].
Mikolov, T., Chen, K., Corrado, G. & Dean, J. (2013) Efficient Estimation of Word Representations in Vector Space. arXiv:1301.3781 [cs]. [Online] Available from: http://arxiv.org/abs/1301.3781 [Accessed: 6 January 2020].
Page, L., Brin, S., Motwani, R. & Winograd, T. (1999) The PageRank Citation Ranking: Bringing Order to the Web. [Online]. 11 November 1999. Available from: http://ilpubs.stanford.edu:8090/422/ [Accessed: 25 January 2020].
Radev, D.R., Hovy, E. & McKeown, K. (2002) Introduction to the Special Issue on Summarization. Computational Linguistics. [Online] 28 (4), 399–408. Available from: doi:10.1162/089120102762671927.
Ramos, J. (2003) Using TF-IDF to Determine Word Relevance in Document Queries. In Proceedings of the first instructional conference on machine learning. 242133–142.
Sanh, V., Wolf, T. & Ruder, S. (2019) A Hierarchical Multi-Task Approach for Learning Embeddings from Semantic Tasks. Proceedings of the AAAI Conference on Artificial Intelligence. [Online] 33 (01), 6949–6956. Available from: doi:10.1609/aaai.v33i01.33016949.
Schuster, M. & Paliwal, K.K. (1997) Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing. [Online] 45 (11), 2673–2681. Available from: doi:10.1109/78.650093.
Stribling, J., Aguayo, D. & Krohn, M. (2005) Rooter: A Methodology for the Typical Unification of Access Points and Redundancy. 4.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., et al. (2017) Attention is All you Need. In: I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, et al. (eds.). Advances in Neural Information Processing Systems 30. [Online]. Curran Associates, Inc. pp. 5998–6008. Available from: http://papers.nips.cc/paper/7181-attention-is-all-you-need.pdf [Accessed: 25 January 2020].
Yan, Y., Qi, W., Gong, Y., Liu, D., et al. (2020) ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training. arXiv:2001.04063 [cs]. [Online] Available from: http://arxiv.org/abs/2001.04063 [Accessed: 30 March 2020].
網路部分:
Olah, C. (2015) Understanding LSTM Networks -- colah’s blog. [Online]. 27 August 2015. Available from: https://colah.github.io/posts/2015-08-Understanding-LSTMs/ [Accessed: 27 March 2020]. |