博碩士論文 103522603 詳細資訊




以作者查詢圖書館館藏 以作者查詢臺灣博碩士 以作者查詢全國書目 勘誤回報 、線上人數:42 、訪客IP:3.139.80.82
姓名 施庫瑪(Sipun Kumar Pradhan)  查詢紙本館藏   畢業系所 資訊工程學系
論文名稱
(A Rapid Deep Learning Model for Goal-Oriented Dialog)
相關論文
★ 整合GRAFCET虛擬機器的智慧型控制器開發平台★ 分散式工業電子看板網路系統設計與實作
★ 設計與實作一個基於雙攝影機視覺系統的雙點觸控螢幕★ 智慧型機器人的嵌入式計算平台
★ 一個即時移動物偵測與追蹤的嵌入式系統★ 一個固態硬碟的多處理器架構與分散式控制演算法
★ 基於立體視覺手勢辨識的人機互動系統★ 整合仿生智慧行為控制的機器人系統晶片設計
★ 嵌入式無線影像感測網路的設計與實作★ 以雙核心處理器為基礎之車牌辨識系統
★ 基於立體視覺的連續三維手勢辨識★ 微型、超低功耗無線感測網路控制器設計與硬體實作
★ 串流影像之即時人臉偵測、追蹤與辨識─嵌入式系統設計★ 一個快速立體視覺系統的嵌入式硬體設計
★ 即時連續影像接合系統設計與實作★ 基於雙核心平台的嵌入式步態辨識系統
檔案 [Endnote RIS 格式]    [Bibtex 格式]    [相關文章]   [文章引用]   [完整記錄]   [館藏目錄]   [檢視]  [下載]
  1. 本電子論文使用權限為同意立即開放。
  2. 已達開放權限電子全文僅授權使用者為學術研究之目的,進行個人非營利性質之檢索、閱讀、列印。
  3. 請遵守中華民國著作權法之相關規定,切勿任意重製、散佈、改作、轉貼、播送,以免觸法。

摘要(中) 摘要

  開放領域問答系統(QA)系統,旨在沒有領域限制的範圍下,提供由自然語言提出問題中的確切答案。本研究的目標在於開發一個學習模式,可以自動產生新的特徵,而無需重新訓練,本系統特別在於解決結構和意義的多重開域QA任務。該框架的主要優點是,它只需要的少量的特徵選取工程以及領域,並能達到同時匹配或超過目前的先進成果。此外,它可以很容易地被訓練成符合任何一種開放域QA的使用。
  我們研究了一類新的學習模式稱為記憶神經網絡。記憶神經網絡的原因與推理具有長期記憶組件的結合,多個記憶神經網絡能學會共同使用彼此資源。長期存儲器可被讀取和寫入,我們使用它進行預測的目標,本論文應用在問題的應答(QA),其中長期記憶成為整個動態資料庫的基礎,並且以文字的形式輸出結果。我們提出基於端至端神經網絡的系統,可以達到我們所希望的目標,並學習執行特殊的操作。在本論文的最後,本系統將與多種不同的基準數據集做比較,並討論未來的工作。
摘要(英) Open-domain Question Answering (QA) systems aim at providing the exact answer(s) to questions formulated in natural language, without restriction of domain. My research goal in this thesis is to develop learning models that can automatically induce new facts without having to be re-trained, in particular its structure and meaning in order to solve multiple Open-domain QA tasks. The main advantage of this framework is that it requires little feature engineering and domain specificity whilst matching or surpassing state-of-the-art results. Furthermore, it can easily be trained to be used with any kind of Open-domain QA.

I investigate a new class of learning models called memory neural networks. Memory neural networks reason with inference components combined with a long-term memory component; they learn how to use these jointly. The long-term memory can be read and written to, with the goal of using it for prediction. I investigate these models in the context of question answering (QA) where the long-term memory effectively acts as a (dynamic) knowledge base, and the output is a textual response. Finally, I show that an end-to-end dialog system based on memory neural networks can reach promising and learn to perform non-trivial operations. I confirm those results by comparing my system to various well-crafted baseline Datasets and future work is discussed.
關鍵字(中) ★ 問題問答
★ 記憶神經網路
★ 長期記憶元件
關鍵字(英) ★ Question Answering
★ Memory neural networks
★ Long-term memory component
論文目次 Contents
Chapter 1 Introduction 1
1.1 Overview 1
Out-Of-order access: 1
Long-term dependency: 2
Unordered set: 3
1.2 Motivation 3
1.3 Brief Literature 4
1.4 Contributions 6
1.5 Thesis Organization 7
Chapter 2 Deep Learning Background 9
2.1 Deep Learning and Artificial Intelligence 9
2.2 Why Deep Learning? 9
2.2.1 Learning Representations 9
2.2.2 Distributed Representations 9
2.2.3 Learning Multiple Levels of Inference 10
2.3 Neural Networks: Definitions and Basics 11
2.3.1 Word Vector Representations 14
2.4 Recurrent Neural Network 15
2.4.1 Adaptive Context Features. 18
2.4.2 Forward Pass 19
2.4.3 Backward Pass 20
2.5 Memory Networks 21
2.5.1 Long Short Term Memory 21
2.5.2 The LSTM Architecture 23
2.5.3 Influence of Preprocessing 25
2.5.4 Gradient Calculation 26
2.5.5 Architectural Enhancements 27
2.5.6 LSTM Equations 27
2.6 Hashing Function 30
Chapter 3 Memory Neural Network 32
3.1 Memory Network Implementation 36
3.1.1 Memory Neural Network Model 36
3.1.2 Training a Memory Neural Network 39
3.1.3 Word Sequences as Input 40
3.1.4 Efficient Memory Via Hashing 41
3.1.5 Modelling Write Time 42
3.1.6 Modelling Previously Unseen Words 44
3.1.7 Exact Matches And Unseen Words 45
Chapter 4 Implementation 46
4.1 Single Layer 47
4.1.1 Memory Representation: 47
4.1.2 Generating the Final Prediction: 48
4.2 Multiple Layers 49
4.3 Synthetic Question and Answering Experiments 52
4.3.1 Sentence Representation: 53
4.3.2 Temporal Encoding: 54
4.3.3 Learning Time Invariance by Injecting Random Noise: 54
Chapter 5 Experimental result 55
5.1 Dataset 55
5.2 Preprocessing. 62
5.3 Baselines 63
5.4 Results 64
5.5 QA With Previously Unseen Words 69
5.6 Combining Simulated Data and Large-Scale QA 69
5.7 Language Modeling Experiments 69
Chapter 6 Conclusion 73
6.1 Main Contributions 73
6.2 Future work 74
REFERENCE………………………………………………………………..75
參考文獻 References

[1] M. Richardson, C. J. Burges, and E. Renshaw, "MCTest: A Challenge Dataset for the Open-Domain Machine Comprehension of Text," in EMNLP, 2013, p. 4.
[2] J. Berant, A. Chou, R. Frostig, and P. Liang, "Semantic Parsing on Freebase from Question-Answer Pairs," in EMNLP, 2013, p. 6.
[3] A. Fader, L. Zettlemoyer, and O. Etzioni, "Open question answering over curated and extracted knowledge bases," in Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, 2014, pp. 1156-1165.
[4] B. Bakker, "Reinforcement Learning with Long Short-Term Memory," in NIPS, 2001, pp. 1475-1482.
[5] X. Yao, J. Berant, and B. Van Durme, "Freebase QA: Information Extraction or Semantic Parsing?," ACL 2014, p. 82, 2014.
[6] J. Berant, V. Srikumar, P.-C. Chen, A. Vander Linden, B. Harding, B. Huang, et al., "Modeling Biological Processes for Reading Comprehension," in EMNLP, 2014.
[7] H. J. Levesque, E. Davis, and L. Morgenstern, "The Winograd schema challenge," in AAAI Spring Symposium: Logical Formalizations of Commonsense Reasoning, 2011, p. 47.
[8] P. Liang, "Lambda dependency-based compositional semantics," arXiv preprint arXiv:1309.4408, 2013.
[9] P. Liang, M. I. Jordan, and D. Klein, "Learning dependency-based compositional semantics," Computational Linguistics, vol. 39, pp. 389-446, 2013.
[10] C. D. Manning and H. Schütze, Foundations of statistical natural language processing vol. 999: MIT Press, 1999.
[11] R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu, and P. Kuksa, "Natural language processing (almost) from scratch," Journal of Machine Learning Research, vol. 12, pp. 2493-2537, 2011.
[12] E. H. Huang, R. Socher, C. D. Manning, and A. Y. Ng, "Improving word representations via global context and multiple word prototypes," in Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, 2012, pp. 873-882.
[13] T. Mikolov, W.-t. Yih, and G. Zweig, "Linguistic Regularities in Continuous Space Word Representations," in HLT-NAACL, 2013, pp. 746-751.
[14] T. Luong, R. Socher, and C. D. Manning, "Better Word Representations with Recursive Neural Networks for Morphology," in CoNLL, 2013, pp. 104-113.
[15] J. Pennington, R. Socher, and C. D. Manning, "Glove: Global Vectors for Word Representation," in EMNLP, 2014, pp. 1532-43.
[16] T. Mikolov, M. Karafiát, L. Burget, J. Cernocký, and S. Khudanpur, "Recurrent neural network based language model," in Interspeech, 2010, p. 3.
[17] S. Hochreiter and J. Schmidhuber, "Long short-term memory," Neural computation, vol. 9, pp. 1735-1780, 1997.
[18] A. Graves, "Supervised Sequence Labelling with Recurrent Neural Networks," Ph.D Thesis, Technical University of Munich, 2008.
[19] W. Zaremba and I. Sutskever, "Learning to execute," arXiv preprint arXiv:1410.4615, 2014.
[20] R. Socher, "Recursive Deep Learning for Natural Language Processing and Computer Vision," Citeseer, 2014.
[21] J. Goodman, "Classes for fast maximum entropy training," in Acoustics, Speech, and Signal Processing, 2001. Proceedings.(ICASSP′01). 2001 IEEE International Conference on, 2001, pp. 561-564.
[22] F. A. Gers and E. Schmidhuber, "LSTM recurrent networks learn simple context-free and context-sensitive languages," IEEE Transactions on Neural Networks, vol. 12, pp. 1333-1340, 2001.
[23] F. A. Gers, N. N. Schraudolph, and J. Schmidhuber, "Learning precise timing with LSTM recurrent networks," Journal of machine learning research, vol. 3, pp. 115-143, 2002.
[24] S. Hochreiter, M. Heusel, and K. Obermayer, "Fast model-based protein homology detection without alignment," Bioinformatics, vol. 23, pp. 1728-1736, 2007.
[25] J. Chen and N. S. Chaudhari, "Protein secondary structure prediction with bidirectional lstm networks," in International Joint Conference on Neural Networks: Post-Conference Workshop on Computational Intelligence Approaches for the Analysis of Bio-data (CI-BIO)(August 2005), 2005.
[26] D. Eck and J. Schmidhuber, "Finding temporal structure in music: Blues improvisation with LSTM recurrent networks," in Neural Networks for Signal Processing, 2002. Proceedings of the 2002 12th IEEE Workshop on, 2002, pp. 747-756.
[27] A. Graves and J. Schmidhuber, "Framewise phoneme classification with bidirectional LSTM and other neural network architectures," Neural Networks, vol. 18, pp. 602-610, 2005.
[28] A. Graves, S. Fernández, F. Gomez, and J. Schmidhuber, "Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks," in Proceedings of the 23rd international conference on Machine learning, 2006, pp. 369-376.
[29] M. Liwicki, A. Graves, H. Bunke, and J. Schmidhuber, "A novel approach to on-line handwriting recognition based on bidirectional long short-term memory networks," in Proc. 9th Int. Conf. on Document Analysis and Recognition, 2007, pp. 367-371.
[30] A. Graves, M. Liwicki, H. Bunke, J. Schmidhuber, and S. Fernández, "Unconstrained on-line handwriting recognition with recurrent neural networks," in Advances in Neural Information Processing Systems, 2008, pp. 577-584.
[31] F. A. Gers, J. Schmidhuber, and F. Cummins, "Learning to forget: Continual prediction with LSTM," Neural computation, vol. 12, pp. 2451-2471, 2000.
[32] A. Graves, S. Fernández, and J. Schmidhuber, "Bidirectional LSTM networks for improved phoneme classification and recognition," in International Conference on Artificial Neural Networks, 2005, pp. 799-804.
[33] S. Sukhbaatar, J. Weston, and R. Fergus, "End-to-end memory networks," in Advances in neural information processing systems, 2015, pp. 2440-2448.
[34] J. Weston, S. Bengio, and N. Usunier, "Wsabie: Scaling up to large vocabulary image annotation," 2011.
[35] John Tolkien and R. Reuel, "The Fellowship of the Ring. George Allen & Unwin," 1954.
[36] J. Weston, S. Chopra, and A. Bordes, "Memory networks," arXiv preprint arXiv:1410.3916, 2014.
[37] J. Weston, A. Bordes, S. Chopra, A. M. Rush, B. van Merriënboer, A. Joulin, et al., "Towards ai-complete question answering: A set of prerequisite toy tasks," arXiv preprint arXiv:1502.05698, 2015.
[38] M. P. Marcus, M. A. Marcinkiewicz, and B. Santorini, "Building a large annotated corpus of English: The Penn Treebank," Computational linguistics, vol. 19, pp. 313-330, 1993.
[39] T. Mikolov, A. Joulin, S. Chopra, M. Mathieu, and M. A. Ranzato, "Learning longer memory in recurrent neural networks," arXiv preprint arXiv:1412.7753, 2014.
指導教授 陳慶瀚(Ching-Han Chen) 審核日期 2016-8-18
推文 facebook   plurk   twitter   funp   google   live   udn   HD   myshare   reddit   netvibes   friend   youpush   delicious   baidu   
網路書籤 Google bookmarks   del.icio.us   hemidemi   myshare   

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明