簡易檢索 / 詳目顯示

研究生: 鄒承翰
Tsou, Cheng-Han
論文名稱: 改良型ESRGAN模型之全方面大型機台面板辨識系統
The improved ESRGAN model of the comprehensive mainframe panel recognition system
指導教授: 王駿發
Wang, Jhing-Fa
學位類別: 碩士
Master
系所名稱: 電機資訊學院 - 電機工程學系
Department of Electrical Engineering
論文出版年: 2023
畢業學年度: 111
語文別: 英文
論文頁數: 58
中文關鍵詞: 大型機台面板文件萃取ESRGAN圖像超解析度
外文關鍵詞: mainframe panels, file extraction, ESRGAN, image super-resolution
相關次數: 點閱:35下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 傳統大型自動化生產設備由海外原廠進口機具,透過操作面板進行參數調整,
    進行自動化生產製造,而大型面板參數與台灣原物料環境與商品容錯性差異有些許
    差異,往往需由人工紀錄操作,建立警示參數標準。
    目前市場中唯一可以進行傳統機台操作紀錄擷取方式,為透過人力進行系統面
    板資訊紀錄,往往費時費力,並且人力成本過高,對產線人員來說工作負擔過於龐
    大。希望透過影像辨識,由外部不破壞機台狀態下,進行相關數據擷取分析改善傳
    統機台無法數據雲端化的技術瓶頸與障礙。
    本論文透過改良ESRGAN網路對大型機台面板進行影像細節紋理還原,進行相
    關大型機台面板數據擷取分析並結合google OCR 將面板影像轉為參數數值,將擷
    取結果結合數據分析,提供更精準大型機台面板參數標準資訊。同時在不同干擾源
    狀態下,仍能進行數據萃取分析並完成自動化輸入圖片輸出偵測的txt文件,有效協
    助產線人員記錄機台面板參數,降低產線人員的工作負擔。

    Traditional mainframe panel automatic production equipment is imported from overseas original factories, and the parameters are adjusted through the operation panel for automatic production and manufacturing. However, the parameters of mainframe panels are slightly different from those of Taiwan’s raw material environment and product fault tolerance, and often need to be manually recorded. Warning parameter standard.
    At present, the only method in the market that can capture the operation records of traditional machines is to manually record the system panel information, which is often time-consuming and laborious, and the labor cost is too high, which is too heavy for the production line personnel. It is hoped that through image recognition, relevant data acquisition and analysis can be performed from the outside without destroying the machine, to improve the technical bottlenecks and obstacles that cannot be cloud-based for traditional machines.
    This research improves the ESRGAN network to restore the image details and textures of mainframe panels, conducts relevant mainframe panel data acquisition and analysis, combines google OCR to convert panel images into parameter values, and combines the captured results with data analysis to provide more accurate mainframe panel parameter standard information. At the same time, under different interference source conditions, data extraction and analysis can still be carried out and txt files for automatic input image output detection can be completed, which effectively assists production line personnel to record machine panel parameters and reduces the workload of production line personnel.

    中文摘要 i Abstract ii Acknowledgments iv Contents v List of Tables vii List of Figures viii 1 Introduction 1 1.1 Background 1 1.2 Motivation 2 1.3 Objectives 2 1.4 Dissertation Organization 3 2 Related Work 5 2.1 Image super-resolution (ISR)5 2.2 GAN 10 2.3 GAN architecture combined with image super-resolution tasks . . . . .12 2.4 Optical Character Recognition (OCR) 14 3 The Proposed System 16 3.1 System overview 16 3.1.1 Degradation Processing 17 3.1.2 Data Pre-processing 18 3.1.3 Image Pre-processing 19 3.1.4 Frame Selection Module 19 3.1.5 Text Recognition Module 19 3.1.6 Data Post-processing 19 3.2 Database 20 3.3 Pre-processing 21 3.3.1 Dataset Processing 21 3.3.2 Data Splitting 22 3.4 Image Super-Resolution Module Architecture 22 3.4.1 ESRGAN Module 22 3.4.2 Improved Degradation Module 30 3.5 Image Pre-processing 32 3.6 Frame Selection Module 32 3.7 Text Recognition Module 35 3.8 Data Post-processing 37 4 Experimental Results 38 4.1 Experimental Environment 38 4.2 Experiment details 38 4.2.1 Experiment Result 39 5 Conclusions 50 5.1 Conclusions 50 5.2 Contribution 50 5.3 Future Works 51 References 52

    [1] Bicubic interpolation. Available: https://en.wikipedia.org/wiki/Bicubic_ interpolation.
    [2] JD Van Ouwerkerk. Image super-resolution survey. Image and vision Computing, 24(10):1039–1052, 2006.
    [3] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and super-resolution. In European conference on computer vision, pages 694–711. Springer, 2016.
    [4] Yulun Zhang, Kunpeng Li, Kai Li, Lichen Wang, Bineng Zhong, and Yun Fu.
    Image super-resolution using very deep residual channel attention networks. In Proceedings of the European conference on computer vision (ECCV), pages 286– 301, 2018.
    [5] Cheng Ma, Yongming Rao, Yean Cheng, Ce Chen, Jiwen Lu, and Jie Zhou.
    Structure-preserving super resolution with gradient guidance. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 7769– 7778, 2020.
    [6] Jian Sun, Zongben Xu, and Heung-Yeung Shum. Gradient profile prior and its applications in image super-resolution and enhancement. IEEE Transactions on Image Processing, 20(6):1529–1542, 2010.
    [7] Jia-Bin Huang, Abhishek Singh, and Narendra Ahuja. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5197–5206, 2015.
    [8] Mehdi SM Sajjadi, Bernhard Scholkopf, and Michael Hirsch. Enhancenet: Single image super-resolution through automated texture synthesis. In Proceedings of the IEEE international conference on computer vision, pages 4491–4500, 2017.
    [9] Yulun Zhang, Yapeng Tian, Yu Kong, Bineng Zhong, and Yun Fu. Residual dense network for image super-resolution. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2472–2481, 2018.
    [10] Jin Xiao, Hongwei Yong, and Lei Zhang. Degradation model learning for real-world single image super-resolution. In Proceedings of the Asian Conference on Computer Vision, 2020.
    [11] Thomas K¨ohler, Michel B¨atz, Farzad Naderi, Andr´e Kaup, Andreas Maier, and Christian Riess. Toward bridging the simulated-to-real gap: Benchmarking super-resolution on real data. IEEE transactions on pattern analysis and machine intelligence, 42(11):2944–2959, 2019.
    [12] Chih-Yuan Yang, Chao Ma, and Ming-Hsuan Yang. Single-image super-resolution: A benchmark. In European conference on computer vision, pages 372–386. Springer, 2014.
    [13] Philippe Th´evenaz, Thierry Blu, and Michael Unser. Image interpolation and resampling. Handbook of medical imaging, processing and analysis, 1(1):393–420, 2000.
    [14] Chao Dong, Chen Change Loy, Kaiming He, and Xiaoou Tang. Learning a deep convolutional network for image super-resolution. In European conference on computer vision, pages 184–199. Springer, 2014.
    [15] Chao Dong, Chen Change Loy, and Xiaoou Tang. Accelerating the super-resolution convolutional neural network. In European conference on computer vision, pages 391–407. Springer, 2016.
    [16] Wenzhe Shi, Jose Caballero, Ferenc Husz´ar, Johannes Totz, Andrew P Aitken, Rob Bishop, Daniel Rueckert, and Zehan Wang. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1874–1883, 2016.
    [17] Tong Tong, Gen Li, Xiejie Liu, and Qinquan Gao. Image super-resolution using dense skip connections. In Proceedings of the IEEE international conference on computer vision, pages 4799–4807, 2017.
    [18] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, DavidWarde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks.
    Communications of the ACM, 63(11):139–144, 2020.
    [19] Alireza Makhzani, Jonathon Shlens, Navdeep Jaitly, Ian Goodfellow, and Brendan Frey. Adversarial autoencoders. arXiv preprint arXiv:1511.05644, 2015.
    [20] Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4401–4410, 2019.
    [21] Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196, 2017.
    [22] Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale gan training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096, 2018.
    [23] Christian Ledig, Lucas Theis, Ferenc Husz´ar, Jose Caballero, Andrew Cunningham, Alejandro Acosta, Andrew Aitken, Alykhan Tejani, Johannes Totz, Zehan Wang, et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4681–4690, 2017.
    [24] Xintao Wang, Ke Yu, Shixiang Wu, Jinjin Gu, Yihao Liu, Chao Dong, Yu Qiao, and Chen Change Loy. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European conference on computer vision (ECCV) workshops, pages 0–0, 2018.
    [25] Anran Liu, Yihao Liu, Jinjin Gu, Yu Qiao, and Chao Dong. Blind image super-resolution: A survey and beyond. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
    [26] Jie Liang, Hui Zeng, and Lei Zhang. Efficient and degradation-adaptive network for real-world image super-resolution. arXiv preprint arXiv:2203.14216, 2022.
    [27] Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, and Kyoung Mu Lee. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pages 136–144, 2017.
    [28] Seungjun Nah, Tae Hyun Kim, and Kyoung Mu Lee. Deep multi-scale convolutional neural network for dynamic scene deblurring. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3883–3891, 2017.
    [29] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
    [30] Ke Zhang, Miao Sun, Tony X Han, Xingfang Yuan, Liru Guo, and Tao Liu. Residual networks of residual networks: Multilevel residual networks. IEEE Transactions on Circuits and Systems for Video Technology, 28(6):1303–1314, 2017.
    [31] Alexia Jolicoeur-Martineau. The relativistic discriminator: a key element missing from standard gan. arXiv preprint arXiv:1807.00734, 2018.
    [32] Kai Zhang, Wangmeng Zuo, and Lei Zhang. Learning a single convolutional super-resolution network for multiple degradations. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3262–3271, 2018.
    [33] Jinjin Gu, Hannan Lu, Wangmeng Zuo, and Chao Dong. Blind super-resolution with iterative kernel correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1604–1613, 2019.
    [34] Yan Huang, Shang Li, Liang Wang, Tieniu Tan, et al. Unfolding the alternating optimization for blind super resolution. Advances in Neural Information Processing Systems, 33:5632–5643, 2020.
    [35] Zhengxiong Luo, Yan Huang, Shang Li, Liang Wang, and Tieniu Tan. Endtoend alternating optimization for blind super resolution. arXiv preprint arXiv:2105.06878, 2021.
    [36] Longguang Wang, Yingqian Wang, Xiaoyu Dong, Qingyu Xu, Jungang Yang, Wei An, and Yulan Guo. Unsupervised degradation representation learning for blind super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10581–10590, 2021.
    [37] Assaf Shocher, Nadav Cohen, and Michal Irani. “zero-shot” super-resolution using deep internal learning. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3118–3126, 2018.
    [38] Victor Cornillere, Abdelaziz Djelouah, Wang Yifan, Olga Sorkine-Hornung, and Christopher Schroers. Blind image super-resolution with spatially variant degradations. ACM Transactions on Graphics (TOG), 38(6):1–13, 2019.
    [39] Jianrui Cai, Hui Zeng, Hongwei Yong, Zisheng Cao, and Lei Zhang. Toward real-world single image super-resolution: A new benchmark and a new model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3086–3095, 2019.
    [40] Pengxu Wei, Hannan Lu, Radu Timofte, Liang Lin, Wangmeng Zuo, Zhihong Pan, Baopu Li, Teng Xi, Yanwen Fan, Gang Zhang, et al. Aim 2020 challenge on real image super-resolution: Methods and results. In European Conference on Computer Vision, pages 392–422. Springer, 2020.
    [41] Yuan Yuan, Siyuan Liu, Jiawei Zhang, Yongbing Zhang, Chao Dong, and Liang Lin. Unsupervised image super-resolution using cycle-in-cycle generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 701–710, 2018.
    [42] Andreas Lugmayr, Martin Danelljan, and Radu Timofte. Unsupervised learning for real-world super-resolution. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), pages 3408–3416. IEEE, 2019.
    [43] ZhouWang, Alan C Bovik, Hamid R Sheikh, and Eero P Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4):600–612, 2004.
    [44] Chao Ma, Chih-Yuan Yang, Xiaokang Yang, and Ming-Hsuan Yang. Learning a no-reference quality metric for single-image super-resolution. Computer Vision and Image Understanding, 158:1–16, 2017.
    [45] Anish Mittal, Rajiv Soundararajan, and Alan C Bovik. Making a “completely blind” image quality analyzer. IEEE Signal processing letters, 20(3):209–212, 2012.

    無法下載圖示 校內:2028-02-02公開
    校外:2028-02-02公開
    電子論文尚未授權公開,紙本請查館藏目錄
    QR CODE