簡易檢索 / 詳目顯示

研究生: 楊佩姗
Yang, Pei-Shan
論文名稱: 基於改良YoloV4模型在肺部電腦斷層掃描影像上進行亞實質肺結節檢測
Subsolid Nodules Detection in 2D Lung CT Images Using Modified YoloV4 Model
指導教授: 郭淑美
Guo, Shu-Mei
共同指導: 連震杰
Lien, Jenn-Jier James
張超群
Chang, Chao-Chun
學位類別: 碩士
Master
系所名稱: 電機資訊學院 - 資訊工程學系
Department of Computer Science and Information Engineering
論文出版年: 2021
畢業學年度: 109
語文別: 英文
論文頁數: 87
中文關鍵詞: 肺結節亞實質結節肺癌電腦斷層掃描影像影像偵測醫學影像處理
外文關鍵詞: Lung Nodule, Subsolid Nodule, Lung Cancer, Computed Tomography, Image Detection, Medical Image Processing
相關次數: 點閱:172下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 在台灣,肺癌(Lung Cancer)的罹患率是十大癌症中的第二名,並且連續十二年佔據癌症死亡率之首。肺癌初期不易察覺,但經醫學證明,肺癌篩檢可有效降低死亡率及提高存活率。為了協助醫師在電腦斷層掃瞄影像(Low-Dose Computed Tomography, LDCT)中找出肺結節,本論文改善YoloV4,以完成偵測肺結節(Lung Nodule)的任務。由骨幹(Backbone)網路-CSPDarkNet53進行不同大小的特徵擷取;接著經過頸部(Neck)架構-FPN、PAN更進一步擷取特徵同時與骨幹網路做資訊融合,保留經過多卷積層而失去的原始資訊;最後在頭部(Head)架構-Detector根據特徵資訊找出肺結節。另外,我們的貢獻在於,因應醫學影像的特性,如: 解析度較高、目標較小等,且更精細的資訊對於醫學影像極為重要,因此,我們在原本架構上增加了一條連結,連通最底層的C1、C2保留原始資訊,並加上空間金字塔池化模組(Spatial Pyramid Pooling Module, SPP Module)增強原始資訊。完成偵測網路後,我們還提出了使用形態學(Morphological Processing)影像處理基於冠狀切面(Coronal-View)影像去除水平切面(Axial-View)非肺部影像,來減少肺外誤判(False Positive),以及基於3D資訊減少血管誤判。最後,針對主要偵測目標,召回率(Recall)為81.0%和準確率(Precision)為87.4%。

    In Taiwan, the probability of having lung cancer ranks second among the top ten cancers, and the death rate is the first for 12 consecutive years. Lung cancer is not easy to detect at the beginning, but we can do lung cancer examination to reduce death rate and increase survival rate. To help doctors to find lung nodules in Low-Dose Computed Tomography (LDCT) images, we improve YoloV4 to complete the task of detecting lung nodules. We use CSPDarkNet53 as the backbone to extract features in different sizes. Then, using FPN and PAN as the neck to extract more features and do the feature fusion from the backbone to keep the original information which lost in multiple convolutional layers. Finally, we detect lung nodules at the head. Due to the characteristics of medical images, such as higher resolution and smaller targets, we add a connection from lower layers (C1 and C2) to keep the original information and use a spatial pyramid pooling module to enhance the information. After the detection network, we remove axial-view non-lung slices based on accumulated coronal-view slices using morphological processing to reduce the false positive at outside of lung and reduce false positive based on blood vessel remove in 3D. Finally, for the main detection target, we got recall rate: 81.0% and precision rate: 87.4%.

    CHAPTER 1 INTRODUCTION 1 1.1 MOTIVATION 1 1.2 RELATED WORK 8 1.3 GLOBAL FRAMEWORK 10 1.4 CONTRIBUTION 11 CHAPTER 2 BACKGROUND 14 2.1 3 VIEWS OF CT IMAGE 14 2.2 FEATURE OF LUNG NODULE 17 2.3 2D LUNG NODULE CATEGORY 19 CHAPTER 3 2D NODULES DETECTION USING MODIFIED YOLOV4 21 3.1 REMOVE AXIAL-VIEW NON-LUNG SLICES BASED ON ACCUMULATED CORONAL-VIEW SLICES USING MORPHOLOGICAL PROCESSING 30 3.2 2D NODULE DETECTION USING MODIFIED YOLOV4 35 3.2.1 MODIFIED: BACKBONE 35 3.2.2 MODIFIED YOLOV4: NECK 41 3.3 MODIFIED YOLOV4: HEAD AND LOSS FUNCTION 46 3.3.1 MODIFIED YOLOV4: HEAD 46 3.3.2 MODIFIED YOLOV4: LOSS FUNCTION 51 3.4 REDUCE FALSE POSITIVE BASED ON BLOOD VESSEL IN 3D 53 CHAPTER 4 EXPERIMENT RESULT 56 4.1 DATA COLLECTION AND METRIC 56 4.1.1 DATA COLLECTION 56 4.1.2 METRIC 61 4.2 EXPERIMENTAL RESULTS: 2D NODULES DETECTION USING MODIFIED YOLOV4 67 4.2.1 2D NODULES DETECTION USING MODIFIED YOLOV4 67 4.2.2 EXPERIMENTAL RESULTS: REDUCE FALSE POSITIVE BASED ON BLOOD VESSEL IN 3D 69 4.2.3 EXPERIMENTAL RESULTS: EXAMPLE FIGURE 71 4.3 EXPERIMENTAL RESULTS: USING DIFFERENT CONFIDENCE SCORE 79 4.4 OTHER EXPERIMENT 82 4.4.1 ONE-STAGE DETECTOR: YOLOV4 VS RETINANET 82 4.4.2 NUMBER OF DETECTORS 82 CHAPTER 5 CONCLUSION AND FUTURE WORK 85 REFERENCE 86

    [1] A. Bochkovskiy, C.Y. Wang, and H.Y.M. Liao, “YOLOv4: Optimal Speed and Accuracy of Object Detection,” In arXiv, vol. 2004.10934, 2020.
    [2] K. Chen, J. Pang, J. Wang, Y. Xiong, X. Li, S. Sun, W. Feng, Z. Liu, J. Shi, W. Ouyang, C.C. Loy, and D. Lin, “Hybrid Task Cascade for Instance Segmentation,” In IEEE Conference on Computer Vision and Pattern Recognition, pp. 4974-4983, 2019.
    [3] S.H. Gao, M.M. Cheng, K. Zhao, X.Y. Zhang, M.H. Yang, and P. Torr, “Res2Net: A New Multi-scale Backbone Architecture,” In IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 43, No. 2, pp. 652-662, 2021.
    [4] K. He, X. Zhang, S. Ren, and J. Sun, “Deep Residual Learning for Image Recognition,” In Conference on Computer Vision and Pattern Recognition, Vol. 1, pp. 770-778, 2016.
    [5] K. He, X. Zhang, S. Ren, and J. Sun, “Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition,” In IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 37, No. 9, pp. 1904-1916, 2015.
    [6] J. Hu, L. Shen, S. Albanie, G. Sun, and E. Wu, “Squeeze-and-Excitation Networks,” In Conference on Computer Vision and Pattern Recognition, pp. 7132-7141, 2018.
    [7] P.F. Jaeger, S.A.A. Kohl, S. Bickelhaupt, F. Isensee, T.A. Kuder, H. Schlemmer, and K.H. Maier-Hein, “Retina U-Net: Embarrassingly Simple Exploitation of Segmentation Supervision for Medical Object Detection,” In arXiv, vol. 1811.08661, 2018.
    [8] Y.L. Li, and S.J. Wang, “HAR-Net: Joint Learning of Hybrid Attention for Single-stage Object Detection,” In arXiv, vol. 1904.11141, 2019.
    [9] T.Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie, “Feature Pyramid Networks for Object Detection,” In Conference on Computer Vision and Pattern Recognition, Vol. 1, pp. 936-944, 2017.
    [10] T.Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollár, “Focal Loss for Dense Object Detection,” In IEEE International Conference on Computer Vision, Vol. 1, pp. 2999-3007, 2017.
    [11] M. Liu, Y. Tan, and L. Chen, “Pneumonia detection based on deep neural network Retinanet,” In International Conference on Image, Video Processing and Artificial Intelligence, Vol. 11321, pp. 113210F, 2019.
    [12] S. Liu, L. Qi, H.F. Qin, J.P. Shi, and J.Y. Jia, “Path Aggregation Network for Instance Segmentation,” In IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759-8768, 2018.
    [13] J. Pang, K. Chen, J. Shi, H. Feng, W. Ouyang, and D. Lin, “Libra R-CNN: Towards Balanced Learning for Object Detection,” In IEEE Conference on Computer Vision and Pattern Recognition, pp. 821-830, 2019.
    [14] J. Redmon, and A. Farhadi, “YOLOv3: An Incremental Improvement,” In arXiv, vol. 1804.02767, 2018.
    [15] S.Q. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” In International Conference on Neural Information Processing Systems, Vol. 1, pp. 91-99, 2015.
    [16] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” In International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 234-241, 2015.
    [17] J. Schlemper, O. Oktay, M. Schaap, M. Heinrich, B. Kainz, B. Glocker, and D. Rueckert, “Attention Gated Networks: Learning to Leverage Salient Regions in Medical Images,” In arXiv, vol. 1808.08114, 2018.
    [18] S.Y. Sun, J.M. Pang, J.P. Shi, S. Yi, and W.L. Ouyang, “FishNet: A Versatile Backbone for Image, Region, and Pixel Level Prediction,” In International Conference on Neural Information Processing Systems, pp. 762-772, 2018.
    [19] M. Tan, and Q.V. Le, “MixConv: Mixed Depthwise Convolutional Kernels,” In British Machine Vision Conference, pp. 1-13, 2019.
    [20] W. Zhu, C. Liu, W. Fan, and X. Xie, “DeepLung: Deep 3D Dual Path Nets for Automated Pulmonary Nodule Detection and Classification,” In IEEE Workshop on Applications of Computer Vision, pp. 673-681, 2018.
    [21] M. Zlocha, Q. Dou, and B. Glocker, “Improving RetinaNet for CT Lesion Detection with Dense Masks from Weak RECIST Labels,” In International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 402-410, 2019.

    下載圖示
    2026-01-26公開
    QR CODE