中大機構典藏-NCU Institutional Repository-提供博碩士論文、考古題、期刊論文、研究計畫等下載:Item 987654321/98285
English  |  正體中文  |  简体中文  |  Items with full text/Total items : 83776/83776 (100%)
Visitors : 59503912      Online Users : 676
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version


    Please use this identifier to cite or link to this item: https://ir.lib.ncu.edu.tw/handle/987654321/98285


    Title: The Trade-off between Long-tailed Classification Accuracy and Model Size
    Authors: 張世何;Chang, Shih-He
    Contributors: 資訊工程學系
    Keywords: Long-Tailed Recognition;Pruning;長尾辨識;剪枝
    Date: 2025-07-21
    Issue Date: 2025-10-17 12:34:58 (UTC+8)
    Publisher: 國立中央大學
    Abstract: 在真實世界的長尾數據集上訓練深度神經網路是一項極具挑戰性的任務,因為資料不平衡會嚴重損害模型在稀有尾部類別上的表現 。本論文旨在解決分類準確度與模型大小之間的關鍵權衡問題 。我們利用多樣化的輔助訓練器 (Auxiliary Trainers) 來生成多個候選遮罩,採用非均勻分層剪枝策略 (Layered Pruning Strategy) 以保留深層網路中的關鍵參數,並透過加權遮罩合併 (Weighted Mask Merging) 來整合出最佳的剪枝決策 。實驗結果顯示,在 CIFAR-100-LT 與 ImageNet-LT 數據集上,我們的模型在分別經過 30% 與 35% 的剪枝後,其尾部類別的準確率依然優於未經壓縮的完整基線模型 。我們的研究確立了一個策略性設計的剪枝流程不僅是模型壓縮的工具,更是一種能有效提升模型在尾部類別上泛化能力的機制,從而打造出在真實世界場景中兼具效率與準確性的模型 。
    ;Training deep neural networks on long-tailed real-world datasets is challenging, as data imbalance harms performance in rare tail classes. This thesis addresses the critical trade-off between classification accuracy and model size. We leverage diverse Auxiliary Trainers to generate multiple candidate masks, apply a non-uniform Layer Pruning Strategy to preserve critical parameters in deeper layers, and consolidate the best pruning decisions via Weighted Mask Merging. On CIFAR-100-LT and ImageNet-LT, our models, even when heavily pruned to 30\% and 35\% sparsity, respectively, maintain superior tail class accuracy over their full-sized baselines. Our work establishes that a strategically designed pruning process is not merely a tool for model compression but a potent mechanism for enhancing generalization on tail classes, yielding models that are simultaneously more efficient and accurate for real-world scenarios.
    Appears in Collections:[Graduate Institute of Computer Science and Information Engineering] Electronic Thesis & Dissertation

    Files in This Item:

    File Description SizeFormat
    index.html0KbHTML14View/Open


    All items in NCUIR are protected by copyright, with all rights reserved.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明