| 研究生: |
葉家瑋 Yeh, Chia-Wei |
|---|---|
| 論文名稱: |
使用深度學習結構光於3D物體檢測系統與其系統穩定性及加速 3D Object Inspection System Using Deep Learning-Based Structured Light, and System Stability and Speedup |
| 指導教授: |
連震杰
Lien, Jenn-Jier 郭淑美 Guo, Shu-Mei |
| 學位類別: |
碩士 Master |
| 系所名稱: |
電機資訊學院 - 資訊工程學系 Department of Computer Science and Information Engineering |
| 論文出版年: | 2020 |
| 畢業學年度: | 108 |
| 語文別: | 英文 |
| 論文頁數: | 116 |
| 中文關鍵詞: | 三角測量 、相位移 、結構光 、立體視覺 、深度學習 、3D物體檢測 |
| 外文關鍵詞: | Triangulation, Phase-shifting, Structured light, Stereo Vision, Deep Learning, 3D Object Inspection |
| 相關次數: | 點閱:118 下載:1 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
目前結構光系統被廣泛應用於機器人視覺、工業量測、3D人臉辨識等領域中。本研究主要是藉由結構光立體視覺系統來量測物體並產生具有X,Y,Z軸三維資訊的點雲,再透過點雲來進行3D物體檢測,在產生點雲之前必須要先建立兩支相機各自的內部校正參數,再透過立體視覺校正找出兩相機之間的外部校正參數,經過結構光所建立的相位移編碼條紋找出兩相機影像平面上的對應關係,之後就可以透過兩支相機間的校正參數輔以三角測量法回推出待測物體表面在三維世界坐標系底下的X,Y,Z軸三維資訊。由於結構光系統的演算法非常耗時,本研究也針對整個系統加速,並針對在點雲成像結果不穩定這方面也做了改進。最後也導入了深度學習的模型進入我們的系統中取代了我們一對一尋找視差值的部分。實驗結果證明,所提出的方法能達到XY軸精度為0.028mm,Z軸精度達到0.01mm,整體執行時間由33s加速到7s,而在最後使用了深度學習的方式Z軸精度達到1mm。
At present, structured light systems are widely used in the fields of robot vision, industrial measurement, and 3D face recognition. This research mainly uses a structured light stereo vision system to measure objects and generate a point cloud with three-dimensional information on X, Y, and Z axes, and then use the point cloud to do 3D object inspection. Before generating the point cloud, the calibration parameters of two cameras must be created. Two cameras need to find their own intrinsic parameters first. Then we can find out the extrinsic parameters between the two cameras through stereo vision calibration. The corresponding relationship on the image plane of the two cameras can be found through the phase-shifting patterns created by structured light, and then we can use the calibration parameters and triangulation method to get the three-dimensional information of the surface of the object in the three-dimensional world coordinate. Since the algorithm of the structured light system is very time-consuming, this research also aims at the acceleration of the whole system, and also improves on the unstable point cloud results. Finally, the deep learning model was imported into our system to replace our one-to-one search for disparity value. Experimental results prove that the proposed method can achieve XY-axis accuracy of 0.028mm, Z-axis accuracy of 0.01mm, and the overall execution time accelerated from 33s to 7s. In the end, deep learning is used to achieve Z-axis accuracy of 1mm.
[1] G. Bradski and A. Kaehler, Learning OpenCV: Computer Vision with the OpenCV Library, O’Reilly Media, 2008.
[2] A. Fusiello, E. Trucco, and A. Verri, "A Compact Algorithm for Rectification of Stereo Pairs," Machine Vision and Applications, Vol. 12, No. 1, pp. 16-22, 2000.
[3] M. Gupta, A. Agrawal, A. Veeraraghavan and S.G. Narasimhan, "Structured Light 3D Scanning in the Presence of Global Illumination," IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2011.
[4] K. Herakleous, and C. Poullis, "3DUNDERWORLD-SLS: an Open-source Structured-light Scanning System for Rapid Geometry Acquisition," arXiv preprint arXiv:1406.6595, 2014.
[5] D. Lanman, and G. Taubin, "Build Your Own 3D Scanner: 3D Photography for Beginners," ACM SIGGRAPH 2009 Courses, 2009.
[6] D. Li, H. Zhao, and H. Jiang, “Fast Phase-based Stereo Matching Method for 3D Shape Measurement,” International Symposium on Optomechatronic Technologies (ISOT), 2010.
[7] P.F. Luo, Y.J. Chao, and M.A. Sutton, "Application of Stereo Vision to Three-dimensional Deformation Analyses in Fracture Experiments," Optical Engineering, Vol. 33, No. 3. pp. 981-991, 1994.
[8] J.S. Massa, G.S. Buller, A.C. Walker, S. Cova, M. Umasuthan, and A.M. Wallace, "Time-of-flight Optical Ranging System Based on Time-correlated Single-photon Counting," Applied Optics, Vol. 37, No. 31, pp. 7298-7304, 1998.
[9] D. Moreno, K. Son, and G. Taubin, "Embedded Phase Shifting: Robust Phase Shifting with Embedded Signals," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015.
[10] C. Reich, R. Ritter, and J. Thesing. "White light heterodyne principle for 3D-measurement." Sensors, Sensor Systems, and Sensor Data Processing. Vol. 3100. International Society for Optics and Photonics, 1997.
[11] J. Salvi, S. Fernandez, T. Pribanic, and X. Llado, "A State of the Art in Structured Light Patterns for Surface Profilometry," Pattern Recognition, Vol. 43, No. 8, pp. 2666-2680, 2010.
[12] J. Salvi, J. Pages, and J. Batlle, "Pattern Codification Strategies in Structured Light Systems," Pattern Recognition, Vol. 37, No .4, pp. 827-849, 2004.
[13] S.M. Seitz, and C.R. Dyer, "View Morphing," Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, 1996.
[14] S. Zhang, "Digital Multiple Wavelength Phase Shifting Algorithm," International Society for Optics and Photonics, 2009.
[15] Z. Zhang, "A Flexible New Technique for Camera Calibration," IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 22, 2000.
[16] Z. Zhang, "A Flexible New Technique for Camera Calibration," IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 22, 2000.
[17] Li, F., et al. "Depth acquisition with the combination of structured light and deep learning stereo matching." Signal Process. Image Commun. 75, 111–117 (2019).
[18] Kniaz, V. V. "FringeMatchNet: effective stereo matching onboard of mobile structured light 3D scanner." Optics for Arts, Architecture, and Archaeology VII, 11058 152 –160 International Society for Optics and Photonics, SPIE(2019).
[19] Q. Du, R. Liu, B. Guan, Y. Pan, S. Sun. " Stereo-Matching network for structured light." IEEE Signal Process. Lett., 26 (1) (2019), pp. 164-168.
[20] G. Yang, J. Manela, M. Happold, and D. Ramanan. Hierarchical deep stereo matching on high resolution images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5515–5524, 2019.
[21] F. Zhang, V. Prisacariu, R. Yang, and P. H. Torr, “Ga-net: Guided aggregation net for end-to-end stereo matching,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 185–194.
[22] K. He, X. Zhang, S. Ren, and J. Sun. "Deep residual learning for image recognition." Computer Vision and Pattern Recognition (CVPR), 2016 IEEE Conference on. IEEE, 2016.
[23] A. Geiger, P. Lenz, and R. Urtasun. “Are we ready for autonomous driving? the kitti vision benchmark suite.” Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on. IEEE, 2012.
[24] J. Zbontar and Y. LeCun. Stereo matching by training a convolutional neural network to compare image patches. Journal of Machine Learning Research, 17(1-32):2, 2016.
[25] O. Ronneberger, P. Fischer, and T. Brox. U-Net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 234–241. Springer, 2015.
[26] J. Pang, W. Sun, J. S. Ren, C. Yang, and Q. Yan. Cascade residual learning: A two-stage convolutional neural network for stereo matching. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017
[27] A. Newell, K. Yang, and J. Deng. Stacked hourglass networks for human pose estimation. In European Conference on Computer Vision, pages 483–499. Springer, 2016.
[28] N. Mayer, E. Ilg, P. Hausser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox. A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016
[29] J.-R. Chang and Y.-S. Chen. Pyramid stereo matching network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5410–5418, 2018.
[30] A. Kendall, H. Martirosyan, S. Dasgupta, P. Henry, R. Kennedy, A. Bachrach, and A. Bry. End-to-end learning of geometry and context for deep stereo regression. In Proceedings of the IEEE International Conference on Computer Vision, pages 66–75, 2017.
[31] Y. Wu and K. He. Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV), pages 3–19, 2018.
[32] C. Godard, O. Mac Aodha, and G. J. Brostow. Unsupervised monocular depth estimation with left-right consistency. Computer Vision and Pattern Recognition (CVPR), 2017 IEEE Conference on. IEEE, 2017.
[33] X. Guo, K. Yang, W. Yang, X. Wang, and H. Li. Group-wise correlation stereo network. Computer Vision and Pattern Recognition (CVPR), 2019 IEEE Conference on. IEEE, 2019.