中大機構典藏-NCU Institutional Repository-提供博碩士論文、考古題、期刊論文、研究計畫等下載:Item 987654321/79589
English  |  正體中文  |  简体中文  |  Items with full text/Total items : 80990/80990 (100%)
Visitors : 41262146      Online Users : 84
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version


    Please use this identifier to cite or link to this item: http://ir.lib.ncu.edu.tw/handle/987654321/79589


    Title: 實作於微控制器的深度神經網路聲音事件辨識;A Deep Neural Network for Sound Event Recognition Implemented in Microcontroller
    Authors: 劉振宏;Liu, Chen-Hung
    Contributors: 資訊工程學系在職專班
    Keywords: 深度神經網路;聲音事件辨識;微控制器;量化;深度學習;DS-CNN;DS-CNN;MCU;quantization
    Date: 2019-01-30
    Issue Date: 2019-04-02 15:04:14 (UTC+8)
    Publisher: 國立中央大學
    Abstract: 典型的深度神經網路需要使用大量記憶體和高速浮點數計算性能,難以應用在硬體資源極少的微控制器嵌入式平台。深度神經網路可以成功的應用在聲音事件辨識,但為了能夠在微控制器平台實作深度聲音事件辨識應用,本研究提出一個量化策略,用以壓縮深度神經網路模型,以便在辨識性能和硬體資源需求之間進行最佳化。本研究採用了DS-CNN的架構去建構聲音事件辨識神經網路模型,擷取聲音的MFCC作為特徵來訓練辨識模型,透過我們的量化程序,將量化過後的權重參數置入ARM Cortex-M7微控制器進行驗證。在PC平台訓練完成的神經網路模型可以達到82%的辨識率,經過量化和移植到MCU平台後,在維持相同的0.2秒的辨識速度條件下,辨識率降低至60%。證實此方法的確可將PC上訓練後的深度神經網路模型移植到MCU平台運行,且仍然維持可接受的辨識性能和辨識率。本研究成果可將深度學習AI技術推廣至眾多低硬體資源需求的應用。;Typical deep neural networks require the use of considerable memories and high-speed floating-point arithmetic; hence, it is difficult to apply it to microcontroller-embedded platforms with limited hardware resources. Deep neural networks can be successfully applied in recognizing sound events. To facilitate the implementation of microcontroller platforms in deep sound event recognition, this study proposed a quantization strategy to compress deep neural networks and optimize the recognition performance and hardware resource needs. This study adopted the depthwise separable convolutional neural network (DS-CNN) structure to establish the neural network model for sound event recognition. Mel-frequency cepstral coefficients (MFCC) that extract sound were used as the features to train recognition models. Through the quantization process, the quantized weight parameters were input into an ARM Cortex-M7 microcontroller to facilitate verification. The neural network model that completed training on a personal computer platform reached a recognition rate of 82%. After the model was quantized and transferred to a microcontroller unit, the recognition rate dropped to 60% with the recognition speed remaining at 0.2 second. The result verified that the proposed method can enable the deep neural network model training on a personal computer to be transferred to microcontroller units while maintaining acceptable recognition performance and recognition rates. The results can extend the deep learning artificial intelligence technologies to numerous applications with low requirement of hardware resources.
    Appears in Collections:[Executive Master of Computer Science and Information Engineering] Electronic Thesis & Dissertation

    Files in This Item:

    File Description SizeFormat
    index.html0KbHTML226View/Open


    All items in NCUIR are protected by copyright, with all rights reserved.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明