在醫學影像分析領域中,深度學習模型的效能高度依賴大量且高 品質的標註資料,然而醫學標註往往面臨高成本與高專業門檻的挑戰。 為降低對人工標註的依賴,本研究提出一套應用於2D 醫學影像之自 監督式對比學習架構,改良自3D VoCo(Volume Contrastive Learning Framework)而來,並整合序列建模技術,提升模型於腹部創傷分類任務 中的辨識能力。 本研究探討改良自3D VoCo 的2D 自監督對比學習方法於腹部CT 影像分類任務中的應用。透過在公開腹部資料集進行切片級對比預訓練, 學習切片間的語意結構,再將主幹架構遷移至RSNA 2023 資料集,執行 多器官與單器官損傷分類。下游模型結合CNN-LSTM 架構以捕捉切片 序列關聯,並透過多組消融實驗驗證對比策略之效益。 實驗結果顯示,在多器官分類任務中也有不錯的成效,在資料標註有 限的醫學場景下,能有效捕捉空間語意關聯並提升分類性能,證實VoCo 框架在腹部CT 分析領域具備實務可行性與應用潛力,並提出了未來改 進的方向,以進一步提升模型的實用性和泛化能力。這些結果表明2D VoCo 方法在醫療影像結合深度學習領域具有廣泛的應用前景和強大的 擴展能力。;In the field of medical image analysis, the performance of deep learning models heavily depends on large-scale, high-quality annotated datasets. However, medical annotations often face high costs and require specialized expertise. To reduce reliance on manual labeling, this study proposes a selfsupervised contrastive learning framework tailored for 2D medical imaging, adapted from the 3D Volume Contrastive Learning Framework (VoCo), and integrates sequence modeling techniques to enhance performance in abdominal trauma classification tasks. This study explores the application of the improved 2D VoCo method on abdominal CT image classification. By conducting slice-level contrastive pretraining on publicly available abdominal datasets, the model learns semantic structures across slices and transfers the pretrained backbone to the RSNA 2023 dataset for downstream multi-organ and single-organ injury classification tasks. The downstream model adopts a CNN-LSTM architecture to capture spatial-temporal correlations across slices, and a series of ablation studies are conducted to validate the effectiveness of the proposed contrastive strategy. Experimental results show that the proposed approach achieves promising performance even in multi-organ classification settings. Under limited annotation scenarios, the method effectively captures spatial-semantic dependencies and improves classification accuracy. These findings demonstrate the practical feasibility and application potential of the VoCo framework for abdominal CT analysis, and suggest directions for further improvement to enhance model generalizability and utility. Overall, the 2D VoCo method exhibits strong potential and scalability for medical image analysis in combination with deep learning.