簡易檢索 / 詳目顯示

研究生: 許鈺祥
Hsu, Yu-Hsiang
論文名稱: 基於YOLO深度學習應用於火災及人員安全檢測-以建築物及車輛為例
Fire and Personnel Safety Detection Based on YOLO Deep Learning-Using Building and Vehicle as an Example
指導教授: 蔡青志
Tsai, Shing-Chih
學位類別: 碩士
Master
系所名稱: 管理學院 - 工業與資訊管理學系碩士在職專班
Department of Industrial and Information Management (on the job class)
論文出版年: 2023
畢業學年度: 111
語文別: 中文
論文頁數: 69
中文關鍵詞: 火災偵測人員偵測車輛偵測建築物偵測深度學習物件辨識YOLO
外文關鍵詞: Fire Detection, Personnel Detection, Vehicle Detection, Building Detection, Deep Learning, Object Detection, YOLO
相關次數: 點閱:229下載:54
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 火災屬於意外或人為發生的災害,以致於有不少人員受傷甚至是死亡,在現實生活中有很多偵測火災發生時的設備,如偵煙探測器、火警自動警報設備、緊急廣播設備、住宅警報器、火警受信總機等等,即使如此偵測火災的發生仍然會有死角漏洞等缺陷,無法提早發出通報以減少人員生命財產的受損。
      本論文使用YOLO (You Only Look Once)V4深度學習模型方法,此為一種影像物件辨識的方法,除了上述所列之設備外,也可利用影像辨識來偵測現場是否有火災,本文主要偵測的物件以都市的建築物和車輛為主,2種均為常見火災發生的案件,檢測建築物內外及火燒車輛的附近是否發生火災及人員在場,除了上述火災發生的預警通報外,也能協助消防隊判斷火災發生及人員的位置,以提升救援的效率。
    實驗結果得知在適用於火災場景時,YOLOv4與YOLOv4 CSP具有較佳的檢測效果,YOLOv4的最佳mAP為85.67%,YOLOv4 CSP的最佳mAP為82%,YOLOv4 Tiny的最佳mAP僅69.27%,且有許多誤判及辨識的情況,因此除非在硬體條件比較嚴苛的情況下才考慮使用,否則理應以YOLOv4與YOLOv4 CSP為主要的辨識算法。

    Fire is an accidental or human-caused disaster that results in injuries or even deaths for many people. There are many devices that detect fires in real life, such as smoke detectors, automatic fire alarm devices, emergency broadcast equipment, home alarms, fire alarm master stations, etc. However, even with such devices, there are still dead angles and other defects that prevent early notification to reduce the loss of lives and property.
      This thesis uses the YOLO (You Only Look Once) V4 deep learning model method, which is an image object recognition method. In addition to the above-mentioned devices, image recognition can also be used to detect fires on-site. The main objects detected in this paper are urban buildings and vehicles, which are both common cases of fire.
      The detection checks whether there is a fire and whether there are people inside and outside the buildings and near the burning vehicles, as well as assisting the fire department in determining the location of the fire and the people to improve the efficiency of rescue.
      Experimental results show that YOLOv4 and YOLOv4 CSP have better detection results when applied to fire scenes. YOLOv4 has a best mAP of 85.67%, YOLOv4 CSP has a best mAP of 82%, and YOLOv4 Tiny has a best mAP of only 69.27%, with many misjudgments and recognition cases. Therefore, unless under hardware conditions that are more stringent, YOLOv4 and YOLOv4 CSP should be the main recognition algorithms, rather than YOLOv4 Tiny.

    摘要 I EXTENDED ABSTRACT II 目錄 VII 表目錄 IX 圖目錄 X 第一章 緒論 1  1.1 研究背景與動機 1  1.2 研究目的 2  1.3 研究範圍與限制 3  1.4 論文架構 3 第二章 文獻探討 4  2.1 火災檢測 4  2.2 深度學習(Deep Learning) 5  2.3 物件影像辨識相關技術 6   2.3.1 CNN 8   2.3.2 Faster RCNN 10   2.3.3 YOLOV1 12   2.3.4 YOLOV2 15   2.3.5 YOLOV3 17   2.3.6 YOLOV4 20   2.3.7 Scaled YOLOV4 26  2.4 小結 29 第三章 研究方法 30  3.1 研究流程與架構 30  3.2 資料收集與標記 31   3.2.1 YOLO資料標記格式 32  3.3 物件辨識演算法 34   3.3.1 YOLOV4 34   3.3.2 CSP-ized YOLOv4 35   3.3.3 YOLO V4 tiny 35  3.4 評估方法 35  3.6 小結 38 第四章 實驗結果與分析 39  4.1實驗環境 39  4.2訓練階段及參數設置 40   4.2.1 樣本數與相關實驗資料準備 40   4.2.2 實驗參數設置 41   4.2.3 訓練過程 42  4.3訓練與辨識結果 43   4.3.1 訓練結果 43   4.3.2 圖片物件辨識結果 45   4.3.3 影片物件辨識結果 51  4.4小結 56 第五章 結論與未來發展 57  5.1結論 57  5.2未來發展 58 參考文獻 59 附錄一 主要程式碼 63

    [1] 內政部消防署全球資訊網(2022)。火災統計。2022年9月15日,取自: https://www.nfa.gov.tw/cht/index.php?code=list&ids=220
    [2] 林東清(2018)。資訊管理e化企業的核心競爭能力七版。臺北市:智勝文化。
    [3] 李謦伊(2021)。YOLO演進—3—YOLOv4詳細介紹。2021年4月10日,取自:https://medium.com/ching-i/yolo演進-3-yolov4詳細介紹-5ab2490754ef
    [4] Bochkovskiy, A., Wang, C.-Y., & Hong, Y. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934.
    [5] He, K., Zhang, X., Ren, S., & Sun, J. (2015). Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 37(9), 1904-1916.
    [6] He, K., Zhang, X., Ren, S., & Sun, J. (2015). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778.
    [7] Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. Science, 313(5786), 504-507.
    [8] Hu, H., Wang, G., Zhang, Q., Wang, J., Fang, J., & Zhang, Y. (2009). Design wireless multi-sensor fire detection and alarm system based on ARM. In 2009 9th International Conference on Electronic Measurement & Instruments, pp. 3-285.
    [9] LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278-2324.
    [10] Lee, Y., Hwang, J.-w., Lee, S., Bae, Y., & Park, J. (2019). An energy and GPU-computation efficient backbone network for real-time object detection. IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 2019-June, 752–760. https://doi.org/10.1109/CVPRW.2019.00103
    [11] Lin, T.-Y., Doll'ar, P., Girshick, R., He, K., Hariharan, B., & Belongie, S. (2017). Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117-2125.
    [12] Liu, S., Qi, L., Qin, H., Shi, J., & Jia, J. (2018). Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759-8768.
    [13] Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C., & Berg, A. (2020). Ssd: Single shot multibox detector. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I 14, pp. 21-37. Springer International Publishing.
    [14] Mohd Razmi, S., Saad, N., & Asirvadam, V. S. (2010). Vision-based flame detection: Motion detection & fire analysis. In 2010 IEEE Student Conference on Research and Development (SCOReD), pp. 187-191.
    [15] Mwedzi, N. A., Nwulu, N. I., & Gbadamosi, S. L. (2019). Machine learning applications for fire detection in a residential building. In 2019 IEEE 6th International Conference on Engineering Technologies and Applied Sciences (ICETAS), pp. 1-4.
    [16] Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779-788.
    [17] Redmon, J., & Farhadi, A. (2017). YOLO9000: better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779-788.
    [18] Redmon, J., & Farhadi, A. (2018). Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767.
    [19] Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks. arXiv preprint arXiv:1506.01497.
    [20] Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., & Savarese, S. (2019). Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 658-666.
    [21] Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
    [22] Solorzano, A., Fonollosa, J., Fernandez, L., Eichmann, J., & Marco, S. (2017). Fire detection using a gas sensor array with sensor fusion algorithms. In 2017 ISOCS/IEEE International Symposium on Olfaction and Electronic nose (ISOEN), pp. 1-3.
    [23] Soviany, P., & Ionescu, R. T. (2018). Optimizing the trade-off between single-stage and two-stage deep object detectors using image difficulty prediction. In 2018 20th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC), pp. 209-214.
    [24] Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., & Rabinovich, A. (2014). Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1-9.
    [25] Wang, C.-Y., Bochkovskiy, A., & Liao, H.-Y. M. (2021). Scaled-yolov4: Scaling cross stage partial network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13029-13038.
    [26] Wang, C.-Y., Hong, Y., Yeh, I. H., Wu, Y.-H., Chen, P.-Y., & Hsieh, J.-W. (2019). CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 390-391.
    [27] Wu, S., & Zhang, L. (2018). Using popular object detection methods for real time forest fire detection. In 2018 11th International Symposium on Computational Intelligence and Design (ISCID), pp. 280-284.
    [28] Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., & Ren, D. (2020). Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, 34(7), 12993-13000.

    下載圖示 校內:立即公開
    校外:立即公開
    QR CODE