English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 78852/78852 (100%)
造訪人次 : 35423972      線上人數 : 1186
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋


    請使用永久網址來引用或連結此文件: http://ir.lib.ncu.edu.tw/handle/987654321/92500


    題名: 基於標靶訓練策略與強預測器的神經網路架構搜索方法;HTTP-NAS:Highly Targeted Training Strategy with Strong Predictors-based Method for Neural Architecture Search
    作者: 張毓修;Chang, Yu-Hsiu
    貢獻者: 資訊工程學系
    關鍵詞: 神經網路架構搜索;Neural Architecture Search
    日期: 2023-07-13
    上傳時間: 2023-10-04 16:03:09 (UTC+8)
    出版者: 國立中央大學
    摘要: 正如我們所知,設計神經網路架構需要大量的手工努力。因此促進了神經網路架構搜索(NAS)的發展。但訓練和驗證每個候選架構需要大量的時間,因此如何在最少的時間成本下找到效能最好的神經網路架構就是NAS領域很重要的衡量指標。最近研究者會採用迭代式訓練策略(例如BRP-NAS, WeakNAS)或者結合zero-cost(例如:ProxyBO)讓訓練過程中盡量挑選高效能的架構來訓練預測器,事實也證明在相同預算下會強於隨機挑選訓練架構訓練出來的預測器,這因此激發我們做出進一步猜想:在迭代式訓練策略中,如果相同訓練預算下只保留一部分高分架構來訓練預測器,會不會比全部訓練預算都拿來訓練的預測器還要強?我們對此做了一系列的實驗並且驗證了此猜想,而且效果非常卓越,因此我們將此發現結合迭代式訓練策略,提出了Highly Targeted Training Strategy(HTTS)。在預測器架構方面,我們針對Predictor-based NAS領域中基於雙向圖形卷積網路(Bi-GCN)的強預測器架構進行分析和優化。在本論文中,我們提出了更強力的預測器:Fully-BiGCN,其大幅加強了預測器對每層特徵的重視,使用Fully-BiGCN預測器搭配HTTS,我們發展出NAS新方法:HTTP-NAS。跟目前Predictor-based NAS領域的SOTA(WeakNAS)相比,HTTP-NAS取得了很好的效果,以NAS-Bench-201當作Benchmark,分別只需要WeakNAS的27.1% (CIFAR10), 49.0% (CIFAR100), 51.75% (ImageNet16-120)的訓練預算,預測器就可以找到全局最佳架構。;As we know, the design of a neural network architecture requires a significant amount of manual effort. It hence spurs the development of Neural Architecture Search (NAS). However, the training and evaluation of each candidate′s architecture requires tremendous amount of time. Thus, finding the best-performing neural network architecture with minimal computation cost is a crucial event in the NAS research. Recently, researchers adopt iterative training strategies (e.g., BRP-NAS, WeakNAS) or combine them with zero-cost approaches (e.g., ProxyBO) to train predictors to select high-performance architectures during the training process. It has been observed that these methods outperform random sample-based training architectures under the same cost. It hence leads to a hypothesis: If we train a predictor by retaining only a subset of high-score architectures within the same training budget, will it be more robust than a predictor trained with the entire training? We have conducted a series of experiments to validate this hypothesis and found significant improvements. Combining this discovery with the iterative training strategy, we proposed the Highly Targeted Training Strategy (HTTS). In terms of predictor architecture, we analyze and optimize the strong predictor architecture based on the Bidirectional Graph Convolutional Network (Bi-GCN) in the field of Predictor-based NAS. In this thesis, we propose a more powerful predictor called Fully-BiGCN which can significantly enhance the emphasis of the predictor on each layer′s features. Using the Fully-BiGCN predictor with HTTS, a new NAS method called HTTP-NAS is developed. HTTP-NAS achieves remarkable results comparing with the state-of-the-art in Predictor-based NAS (WeakNAS),. Using NAS-Bench-201 as the benchmark, HTTP-NAS required only 27.1% (CIFAR10), 49.0% (CIFAR100), and 51.75% (ImageNet16-120) of training cost of WeakNAS in finding the globally optimal architecture.
    顯示於類別:[資訊工程研究所] 博碩士論文

    文件中的檔案:

    檔案 描述 大小格式瀏覽次數
    index.html0KbHTML19檢視/開啟


    在NCUIR中所有的資料項目都受到原著作權保護.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明