博碩士論文 108522050 完整後設資料紀錄

DC 欄位 語言
DC.contributor資訊工程學系zh_TW
DC.creator賴映如zh_TW
DC.creatorYing-Ru Laien_US
dc.date.accessioned2022-4-7T07:39:07Z
dc.date.available2022-4-7T07:39:07Z
dc.date.issued2022
dc.identifier.urihttp://ir.lib.ncu.edu.tw:88/thesis/view_etd.asp?URN=108522050
dc.contributor.department資訊工程學系zh_TW
DC.description國立中央大學zh_TW
DC.descriptionNational Central Universityen_US
dc.description.abstract視障者在日常生活常常會面臨許多的困難,尤其在衣著方面,他 們除了難以自行得知衣物布料的顏色、花紋和款式等資訊之外,要如 何適切地掌握衣物間互相的搭配更是個挑戰。因此,如何使視障者能 在無旁人輔助的狀態下,瞭解目前全身衣物的布料資訊,並對其色彩 與花紋搭配合適程度有所認知,避免因缺乏資訊而穿著突兀或不適當, 是個重要的研究課題。因此,本論文結合深度學習與影像處理技術來 開發一套盲人穿搭輔助系統,幫助視障者了解選用的衣物資訊與搭配合宜度。 本篇論文系統主要包含以下四個模組: (1) 骨架偵測模組: 使用 OpenPose 模型偵測與提醒拍攝是否為完整正面全身照、(2) 肢體偵測 與切割處理模組: 透過CDCLHuman-Part-Segmentation模型取得肢體 遮罩,並依據相對應欲判斷衣物部位做聚焦切割處理、(3) 衣物特徵 辨識模組: 以深度學習技術訓練模型辨識衣物資訊與布料樣式、(4) 布料色系與飽和度辨識模組: 透過影像處理識別布料的顏色資訊。綜 合上述四部分之模組功能,在取得衣物特徵、布料色系與花紋樣式等資訊後,系統透過語音提示的方式來輔助視障者能夠獨立了解他們選用的各衣物特徵,並給予穿搭參考建議,希望使他們能對於衣物搭配 組合後的效果更有概念和依據,同時保有自己對搭配美感的彈性。 根據系統實驗結果顯示,三種衣物特徵以及布料款式辨識率皆在 95%以上,色系分類準確率達 87%,搭配結果建議與市調相符度約達 77.6%,由此可證明本系統具備一定程度之可用性。zh_TW
dc.description.abstractThe visually impaired people face many difficulties in daily life, especially in clothing identification and matching. In addition to having difficulty in independently knowing the color, pattern, style and other information of clothing fabrics, it is also a challenge for them to match each other well. Therefore, how to let the visually impaired understand the fabric information what they are wearing, and the suitability of the color and pattern of their clothes without the help of others, so as to avoid conflicts or unexpected clothes matching caused by insufficient information, is an important research topic. This research uses deep learning and image processing technology to develop an outfit improvement system in order to help the visually impaired understand the information about their clothes and improve the outfit collocation. The deep-learning-based outfit improvement system developed in this paper mainly includes: (1) Skeleton detection: Use the OpenPose model to detect and remind the user whether the photo is a frontal full-body photo, (2) Human body part detection and segmentation processing: Obtain the human body mask through the CDCL Human-Part-Segmentation model, and perform focus cutting processing according to the corresponding parts of the clothing that need to be judged, (3) Clothing feature recognition: Use deep learning technology to train the model to recognize clothing information and fabric styles, (4) Recognition of fabric color system and saturation: Recognize the color information of fabric through image processing. Through the functions of the above four parts, after obtaining clothing features and fabric color system and pattern style information, voice prompts will be used to assist the visually impaired to independently understand the features of each clothing they choose, and to give non-absolute wear reference suggestions to help them have a more conceptual and basis for the effect of clothing collocation and combination, while maintaining their flexibility in the beauty of collocation. According to the system experiment results, the accracy of three kinds of clothing features and fabric styles has all reached more than 95%, the color classification accuracy rate is 87%, and the match between the recommended matching results and the taste of the general public is 77.6%, which can prove that the system has a certain degree of usability.en_US
DC.subject深度學習zh_TW
DC.subject影像辨識zh_TW
DC.subject色彩配色zh_TW
DC.subject視障者zh_TW
DC.subjectdeep learningen_US
DC.subjectimage recognitionen_US
DC.subjectcolor combinationen_US
DC.subjectvisually impaireden_US
DC.title基於深度學習之盲人穿搭輔助系統zh_TW
dc.language.isozh-TWzh-TW
DC.titleA Deep-learning-based Outfit Improvement System for Visually Impaired Personsen_US
DC.type博碩士論文zh_TW
DC.typethesisen_US
DC.publisherNational Central Universityen_US

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明