簡易檢索 / 詳目顯示

研究生: 黃楷軒
Huang, Kai-Hsuan
論文名稱: 一個基於循環生成對抗網路的影像超解析演算法
An Image Super-Resolution Algorithm Based on Cycle Generative Adversarial Networks
指導教授: 戴顯權
Tai, Shen-Chuan
學位類別: 碩士
Master
系所名稱: 電機資訊學院 - 電腦與通信工程研究所
Institute of Computer & Communication Engineering
論文出版年: 2019
畢業學年度: 107
語文別: 英文
論文頁數: 53
中文關鍵詞: 超解析生成對抗網路卷積神經網路
外文關鍵詞: super-resolution, generative adversarial network, convolutional neural
相關次數: 點閱:89下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 影像超解析技術的主要目的是利用低解析度影像產生高解析度影像,並盡可能的保留影像資訊、維持自然的紋理樣貌。在本論文中提出了一個基於循環生成對抗網路的影像超解析演算法。此方法中有兩組生成對抗網路,其中一組著重在由低解析度影像生成高解析度影像,另一組則是用來增加資料多樣性,以強化整體網路的效能。經過訓練之後,得到一個低解析度影像與高解析度影像的非線性映射關係。
    實驗結果顯示,本論文提出的影像超解析演算法,相較於其他方法,在客觀影像評估標準以及主觀的影像品質上皆有較好的表現。

    The main purpose of super-resolution technology is to generate high-resolution (HR) images from low-resolution (LR) images, and preserve image information as much as possible to maintain natural texture.
    In this Thesis, a super-resolution algorithm based on the cycle generative adversarial network is proposed. There are two generative adversarial networks, one of which focuses on generating high-resolution images from low-resolution images, and another is used to increase data diversity to enhance performance. A nonlinear mapping relationship between low-resolution images and high-resolution images is calculated after training.
    The experimental results show that the proposed approach has better performance than other methods on subjective visual quality and objective measurement

    Contents iv List of Tables vi List of Figures vii Chapter 1 Introduction 1 1.1 Overview of Super-Resolution 1 Chapter 2 Background and Related Works 4 2.1 Interpolation Method 4 2.1.1 Bicubic iInterpolation 4 2.1.2 Lanczos Interpolation 6 2.2 Neural Networks 7 2.3 Convolutional Neural Network 15 2.3 VGG-16 Network 19 2.4 Generative Adversarial Networks 21 2.5 Cycle-Consistent Adversarial Network 23 Chapter 3 The Proposed Algorithm 25 3.1 Proposed Network Architecture 27 3.2 Generator Network 31 3.3 Discriminator Network 31 3.4 Loss Function 33 Chapter 4 Experimental Results 38 4.1 Experimental Dataset 38 4.2 Parameter and Experimental Setting 41 4.3 Experimental Results of Simulated Images 42 4.4 Ablation Experimental Result 48 Chapter 5 Conclusion and Future Work 49 5.1 Conclusion 49 5.2 Future Work 49 References 50

    [1] R. Y. Tsai and T. S. Huang, "Multiframe image restoration and registration," in Advance Computer Visual and Image Processing, Greenwich, JAI Press, pp.317-339, 1984.
    [2] H. S. Hou, and C. Andrews, “Cubic splines for image interpolation and digital filtering,” IEEE Transactions on acoustics, speech, and signal processing, vol.26 no.6, pp. 508-519, 1978.
    [3] C. E. Duchon, "Lanczos filtering in one and two dimensions." Journal of applied meteorology vol.18, no.8 pp.1016-1022, 1979.
    [4] S. Dai, M. Han, T. Wu and Y. Gong, "Bilateral back-projection for single image super resolution," IEEE International Conference on Multimedia and Expo, pp. 1039-1042. ,2007.
    [5] W. Dong, L. Zhang, G. Shi, and X. Wu, "Nonlocal back-projection for adaptive image enlargement," IEEE International Conference on Image Processing. Vols 1-6, pp.349-352, 2009.
    [6] T. C. Wee, and J. S. Hyun, "A novel fast fractal super resolution technique." IEEE Transactions on Consumer Electronics. vol 56, no.3, pp.1537-1541, 2010.
    [7] W. T. Freeman, T. R. Jones, and E. C. Pasztor, “Learning low-level vision,” International journal of computer vision. Vol.40, no.1, pp.24-47, 2000.
    [8] H. Chang, D. Y. Yeung and Y. Xiong, "Super-resolution through neighbor embedding," Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Vol. 1, pp.275-282, 2004.
    [9] S. H. Hong, R. H. Park, S. Yang, and J. Y. Kim, "Image interpolation using interpolative classified vector quantization," Image and Vision Computing, vol.2, no.2, pp.109-122, 2016.
    [10] C. Dong, C. C. Loy, K. He, and X. Tang, "Learning a deep convolutional network for image super-resolution,” European conference on computer vision. pp.184-199, 2014.
    [11] W. Shi et al, "Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network," Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1874-1883, 2016.
    [12] C. Ledig, et al. "Photo-realistic single image super-resolution using a generative adversarial network," Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4681-4690, 2017.
    [13] K. Fukushima. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biological cybernetics, 1980.
    [14] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel. Backpropagation applied to handwritten zip code recognition. Neural computation, 1989.
    [15] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations (ICLR), 2015.
    [16] A. Krizhevsky, I. Sutskever, and G. E. Hinton, "Imagenet classification with deep convolutional neural networks," in Advances in neural information processing systems, pp. 1097-1105, 2012.
    [17] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. “Generative adversarial nets,” In Neural Information Processing Systems (NIPS), 2014.
    [18] Z. Yi, H. Zhang, T. Gong, Tan, and M. Gong. “Dualgan: Unsupervised dual learning for image-to-image translation,” In IEEE International Conference on Computer Vision (ICCV), 2017.
    [19] J.-Y. Zhu, R. Zhang, D. Pathak, T. Darrell, A. A. Efros, O. Wang, and E. Shechtman. Toward multimodal image-to-image translation. In Advances in Neural Information Processing Systems (NIPS), 2017.
    [20] H. Zhang, V. A. Sindagi, and V. M. Patel. Image de-raining using a conditional generative adversarial network. arXiv preprint arXiv:1701.05957, 2017.
    [21] J. Zhu, T. Park, P. Isola, and A. A. Efros. “Unpaired image-to-image translation using cycle-consistent adversarial networks,” In International Conference on Computer Vision (ICCV), to appear, 2017.
    [22] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. In IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
    [23] A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, W. Wang, and R. Webb. “Learning from simulated and unsupervised images through adversarial training,” The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
    [24] A. Odena, C. Olah, and J. Shlens. “Conditional Image Synthesis With Auxiliary Classifier GANs,” ArXiv e-prints, Oct. 2016.
    [25] O. Ronneberger, P. Fischer, and T. Brox, "U-net: Convolutional networks for biomedical image segmentation," International Conference on Medical image computing and computer-assisted intervention. pp. 234-241, 2015.
    [26] J. Johnson, A. Alahi, and L. Fei-Fei. “Perceptual losses for real-time style transfer and super-resolution,” In Proceedings of the European Conference on Computer Vision (ECCV), pages 694–711. Springer, 2016.
    [27] Guilin Liu, Fitsum A Reda, Kevin J Shih, Ting-Chun Wang, Andrew Tao, and Bryan Catanzaro. Image inpainting for irregular holes using partial convolutions. arXiv preprint arXiv:1804.07723, 2018.
    [28] Lin, Tsung-Yi, et al. "Microsoft coco: Common objects in context," European conference on computer vision. Springer, Cham, 2014.
    [29] M. Bevilacqua, A. Roumy, C. Guillemot, and M. L. Alberi-Morel. “Low-complexity single-image super-resolution based on nonnegative neighbor embedding,” BMVC, 2012.
    [30] R. Zeyde, M. Elad, and M. Protter. “On single image scale-up using sparse-representations,” In Curves and Surfaces, pages 711–730. Springer, 2012.
    [31] tegg89’s github: https://github.com/tegg89/SRCNN-Tensorflow
    [32] JGuillaumin’s github: https://github.com/JGuillaumin/SuperResGAN-keras

    無法下載圖示 校內:2024-06-30公開
    校外:不公開
    電子論文尚未授權公開,紙本請查館藏目錄
    QR CODE