簡易檢索 / 詳目顯示

研究生: 郭泰德
Guo, Tai-De
論文名稱: 應用高精地圖於邊緣設備上之交通號誌辨識
Application of High-Definition Maps for Traffic Light Recognition on Edge Devices
指導教授: 莊智清
Juang, Jyh-Ching
學位類別: 碩士
Master
系所名稱: 電機資訊學院 - 電機工程學系
Department of Electrical Engineering
論文出版年: 2022
畢業學年度: 110
語文別: 英文
論文頁數: 76
中文關鍵詞: 交通號誌辨識自動駕駛高精地圖感興趣區域
外文關鍵詞: Traffic Light Recognition, Automated Driving, HD maps, ROI
相關次數: 點閱:109下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 本文旨在解決自動駕駛車輛的交通號誌辨識任務。我們採用基於視覺的方法中的深度學習方法來執行交通號誌辨識。但是自動駕駛車輛上的邊緣設備計算資源有限,如果將整幅圖像作為模型輸入,推理時間成本會很高。因此,我們利用高精地圖來提取感興趣區域作為模型輸入。這個方法的另一個優點是可以通過高精地圖的先驗知識來去除圖片中不相關的背景減少物體檢測器產生的誤報。此外,我們提出了一種具有數據關聯的感興趣區域提取策略,即使在定位失敗時也能繼續執行交通號誌辨識任務。我們還構建了狀態機以提供平滑的紅綠燈狀態。實驗證明本文提出的方法可以在保持預測性能的同時降低推理時間的成本。

    In this thesis, we focus on solving the task of traffic light recognition which is one of
    the important perception tasks for an automated driving vehicle. We adopt the deep learning methods among vision-based methods to perform the traffic light recognition. The computing resources of edge devices on an automated driving vehicle are limited so that the inference time cost will be high if the entire image is used as model input. Therefore, we leverage the high-definition maps (HD maps) to extract the region of interest (ROI) as model input. Another advantage is that the false positives generated by object detectors can be effectively reduced by prior knowledge. In addition, we propose an ROI extraction strategy with data association to provide the ROI even when localization fails. We also build the state
    machine to provide smooth traffic light states. The experiments show that the proposed method maintains the prediction performance and reduces inference time costs at the same time.

    摘要 I Abstract II Acknowledgments III Contents IV List of Tables VI List of Figures VII List of Abbreviations IX Chapter 1 Introduction 1 1.1 Motivation 1 1.2 Related Works 2 1.3 Contributions 4 Chapter 2 System Overview 5 2.1 System Architecture 5 2.2 HD Maps 7 2.3 Coordinate System and Transformation 14 2.3.1 Coordinate System 14 2.3.2 Transformation 19 Chapter 3 Methods 24 3.1 Traffic Light Projection 24 3.2 ROI Extraction Strategies 32 3.2.1 Region Proposal 32 3.2.2 Default Region Proposal 35 3.2.3 Data Association 36 3.3 Model Selection 38 3.4 State Machine 42 Chapter 4 Experiments 45 4.1 Experimental Setup 46 4.1.1 Hardware Platform 46 4.1.2 Software Platform 51 4.1.3 Training Settings 54 4.1.4 System Organization 56 4.2 Experiment Results 58 4.2.1 ROI Extraction 58 4.2.2 ROI Extraction Strategy Comparison 63 4.2.3 Inference on Edge Devices 69 4.3 Discussion 71 Chapter 5 Conclusion 73 5.1 Conclusions 73 5.2 Future Works 73 Reference 75

    [1] A. Bochkovskiy, C. Wang and H. Liao, “YOLOv4: optimal speed and accuracy of object detection,” arXiv preprint arXiv:2004.10934, 2020
    [2] A. Almagambetov, S. Velipasalar and A. Baitassova, “Mobile standards-based traffic light detection in assistive devices for individuals with color-vision deficiency,” IEEE Transactions on Intelligent Transportation Systems, vol. 16, issue 3, 2015
    [3] Z. Chen, Q. Shi and X. Huang, “Automatic detection of traffic lights using support vector machine,” in IEEE Intelligent Vehicles Symposium (IV), 2015
    [4] K. Behrendt, L. Novak and R. Botros, “A deep learning approach to traffic lights: detection, tracking, and classification,” in IEEE International Conference on Robotics and Automation (ICRA), 2017
    [5] N. Fairfield, C. Urmson and R. Botros, “Traffic light mapping and detection,” in IEEE International Conference on Robotics and Automation, 2011
    [6] C. Janga, S. Choa, S. Jeong et al, “Traffic light recognition exploiting map and localization at every stage,” Expert Systems with Applications, vol. 88, pp. 290–304, 2017
    [7] M. Hirabayashi, A. Sujiwo, A. Monrroy et al, “Traffic light recognition using high-definition map features,” Robotics and Autonomous Systems, vol. 111, pp. 62–72, 2019
    [8] S. Han, J. Pool, J. Tran et al, “Learning both weights and connections for efficient neural network,” Advances in neural information processing systems, pp. 1135–1143, 2015
    [9] M. Zhu and S. Gupta. “To prune, or not to prune: exploring the efficacy of pruning for model compression,” in International Conference on Learning Representations, 2018
    [10] S. Gao, F. Huang, W. Cai et al, “Network pruning via performance maximization,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021
    [11] N. Jouppi, C. Young, N. Patil et al, “In-datacenter performance analysis of a tensor processing unit,” in ACM/IEEE 44th Annual International Symposium on Computer Architecture, 2017
    [12] NVIDIA, “NVIDIA Tesla V100 GPU architecture,” Accessed: Jul. 10, 2022. [Online]. Available: http://images.nvidia.com/content/volta-architecture/pdf/volta-architecture-whitepaper.pdf
    [13] C. Zhu, K. Huang, S. Yang et al, “An efficient hardware accelerator for structured sparse convolutional neural networks on FPGAs,” IEEE Transactions on Very Large Scale Integration Systems, vol. 28, issue 9, 2020
    [14] M. O. TAŞ, H. S. YAVUZ and A. YAZICI, “Updating HD-Maps for autonomous transfer vehicles in smart factories,” in International Conference on Control Engineering & Information Technology (CEIT), 2018
    [15] Y. Kang and A. Magdy, “HiDaM: A unified data model for high-definition (HD) map data,” in International Conference on Data Engineering Workshops (ICDEW), 2020
    [16] R. Liu, J. Wang and B. Zhang, “High definition map for automated driving: overview and analysis,” The Journal of Navigation, vol. 73, no. 2, pp. 324–341, 2020
    [17] W. N. Tun, S. Kim, J.-W. Lee et al, “Open-source tool of vector map for path planning in autoware autonomous driving software,” in IEEE International Conference on Big Data and Smart Computing, 2019
    [18] T. Foote, "tf: The transform library," in IEEE Conference on Technologies for Practical Robot Applications (TePRA), 2013
    [19] Z. Zhang, “A flexible new technique for camera calibration,” IEEE Transactions on pattern analysis and machine intelligence, vol. 22, no. 11, pp. 1330–1334, 2000.
    [20] Opencv, “Camera calibration and 3D reconstruction,” Accessed: Jul. 3, 2022. [Online]. Available: https://docs.opencv.org/2.4/modules/calib3d/doc/camera_calibration_and_3d_reconstruction.html?highlight=calibratecamera
    [21] MathWorks, “What is camera calibration,” Accessed: Jul. 3, 2022. [Online]. Available: https://www.mathworks.com/help/vision/ug/camera-calibration.html#bu0niag
    [22] Kuhn HW, “The hungarian method for the assignment problem,” Princeton University, Naval Research Logistics Quarterly 2, pp. 83–97,1955
    [23] R. Girshick, J. Donahue, T. Darrell et al, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in IEEE Conference on Computer Vision and Pattern Recognition, 2014
    [24] J. Redmon, S. Divvala, R. Girshick et al, “You only look once: unified, real-time object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2016
    [25] ROS, “ROS.org | Wiki: Documentation,” Accessed: Jun. 3, 2022. [Online]. Available: http://wiki.ros.org/
    [26] Autoware, “What is Autoware,” Accessed: Jun. 4, 2022. [Online]. Available: https://github.com/autowarefoundation/autoware/tree/autoware-ai
    [27] F. Munir, S. Azam, A. Sheri et al, “Where am I: localization and 3D maps for autonomous vehicles,” VEHITS, 2019

    無法下載圖示 校內:2027-08-15公開
    校外:2027-08-15公開
    電子論文尚未授權公開,紙本請查館藏目錄
    QR CODE