. Bahdanau, D. et al. (2014). Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473.
. Baxendale, P. B. (1958). Machine-made index for technical literature—an experiment. IBM Journal of research and development, 2(4), 354-361.
. Bengio, Y. et al. (2003). A neural probabilistic language model. Journal of machine learning research, 3(Feb), 1137-1155.
. Bojanowski, P. et al. (2016). Enriching word vectors with subword information. arXiv preprint arXiv:1607.04606.
. BYVoid et al. (2017). OpenCC. GitHub repository. Retrieved from https://github.com/BYVoid/OpenCC
. Chopra, S. et al. (2016). Abstractive sentence summarization with attentive recurrent neural networks. Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.
. Collobert, R. et al. (2011). Natural language processing (almost) from scratch. Journal of machine learning research, 12(Aug), 2493-2537.
. Conroy, J. M. & O′leary, D. P. (2001). Text summarization via hidden markov models. Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval.
. Das, D. & Martins, A. F. (2007). A survey on automatic text summarization. Literature Survey for the Language and Statistics II course at CMU, 4, 192-195.
. Deerwester, S. et al. (1990). Indexing by latent semantic analysis. Journal of the American society for information science, 41(6), 391-407.
. Deng, L. (2014). A tutorial survey of architectures, algorithms, and applications for deep learning. APSIPA Transactions on Signal and Information Processing, 3.
. Edmundson, H. P. (1969). New methods in automatic extracting. Journal of the ACM (JACM), 16(2), 264-285.
. Elman, J. L. (1990). Finding structure in time. Cognitive science, 14(2), 179-211.
. fxsjy et al. (2018). jieba. GitHub repository. Retrieved from https://github.com/fxsjy/jieba
. Graves, A. & Jaitly, N. (2014). Towards end-to-end speech recognition with recurrent neural networks. International Conference on Machine Learning.
. Gu, J. et al. (2016). Incorporating copying mechanism in sequence-to-sequence learning. arXiv preprint arXiv:1603.06393.
. Hassan, H. et al. (2018). Achieving Human Parity on Automatic Chinese to English News Translation. arXiv preprint arXiv:1803.05567.
. Hinton, G. E. (1986). Learning distributed representations of concepts. Proceedings of the eighth annual conference of the cognitive science society.
. Hochreiter, S. & Schmidhuber, J. (1997). Long short-term memory. Neural computation, 9(8), 1735-1780.
. Hu, B. et al. (2015). Lcsts: A large scale chinese short text summarization dataset. arXiv preprint arXiv:1506.05865.
. Huang, E. H. et al. (2012). Improving word representations via global context and multiple word prototypes. Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1.
. Joulin, A. et al. (2016). Bag of tricks for efficient text classification. arXiv preprint arXiv:1607.01759.
. Karpathy, A. & Fei-Fei, L. (2015). Deep visual-semantic alignments for generating image descriptions. Proceedings of the IEEE conference on computer vision and pattern recognition.
. Klein, G. et al. (2017). Opennmt: Open-source toolkit for neural machine translation. arXiv preprint arXiv:1701.02810.
. Kupiec, J. et al. (1995). A trainable document summarizer. Proceedings of the 18th annual international ACM SIGIR conference on Research and development in information retrieval.
. Lai, S. et al. (2016). How to generate a good word embedding. IEEE Intelligent Systems, 31(6), 5-14.
. Lin, C.-Y. (1999). Training a selection function for extraction. Proceedings of the eighth international conference on Information and knowledge management.
. Lin, C.-Y. (2004). Rouge: A package for automatic evaluation of summaries. Text Summarization Branches Out.
. Lin, C.-Y. & Hovy, E. (1997). Identifying topics by position. Proceedings of the fifth conference on Applied natural language processing.
. Lipton, Z. C. et al. (2015). A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:1506.00019.
. Luhn, H. P. (1958). The automatic creation of literature abstracts. IBM Journal of research and development, 2(2), 159-165.
. Mesnil, G. et al. (2013). Investigation of recurrent-neural-network architectures and learning methods for spoken language understanding. Interspeech.
. Mikolov, T. et al. (2013a). Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.
. Mikolov, T. et al. (2010). Recurrent neural network based language model. Eleventh Annual Conference of the International Speech Communication Association.
. Mikolov, T. et al. (2013b). Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems.
. Mnih, A. & Hinton, G. (2007). Three new graphical models for statistical language modelling. Proceedings of the 24th international conference on Machine learning.
. Nallapati, R. et al. (2016). Abstractive text summarization using sequence-to-sequence rnns and beyond. arXiv preprint arXiv:1602.06023.
. Olah, C. (2015). Understanding LSTM Networks. Retrieved from http://colah.github.io/posts/2015-08-Understanding-LSTMs/
. Osborne, M. (2002). Using maximum entropy for sentence extraction. Proceedings of the ACL-02 Workshop on Automatic Summarization-Volume 4.
. Pennington, J. et al. (2014). Glove: Global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP).
. Radev, D. R. et al. (2002). Introduction to the special issue on summarization. Computational linguistics, 28(4), 399-408.
. Schuster, M. & Paliwal, K. K. (1997). Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, 45(11), 2673-2681.
. Sutskever, I. et al. (2014). Sequence to sequence learning with neural networks. Advances in neural information processing systems.
. Svore, K. et al. (2007). Enhancing single-document summarization by combining RankNet and third-party sources. Proceedings of the 2007 joint conference on empirical methods in natural language processing and computational natural language learning (EMNLP-CoNLL).
. Turian, J. et al. (2010). Word representations: a simple and general method for semi-supervised learning. Proceedings of the 48th annual meeting of the association for computational linguistics.
. Wang, P. et al. (2015a). A unified tagging solution: Bidirectional LSTM recurrent neural network with word embedding. arXiv preprint arXiv:1511.00215.
. Wang, P. et al. (2015b). Word embedding for recurrent neural network based TTS synthesis. Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on.
. Zuckerberg, M. (2017, June 27). As of this morning, the Facebook community is now officially 2 billion people! We′re making progress connecting the world, and now let′s bring the world closer together. It′s an honor to be on this journey with you [Facebook Status Update]. Retrieved from https://www.facebook.com/zuck/posts/10103831654565331
. 張昇暉. (2017). 中文文件串流之摘要擷取研究. (碩士論文), 國立中央大學.