| 研究生: |
黃楷軒 Huang, Kai-Hsuan |
|---|---|
| 論文名稱: |
一個基於循環生成對抗網路的影像超解析演算法 An Image Super-Resolution Algorithm Based on Cycle Generative Adversarial Networks |
| 指導教授: |
戴顯權
Tai, Shen-Chuan |
| 學位類別: |
碩士 Master |
| 系所名稱: |
電機資訊學院 - 電腦與通信工程研究所 Institute of Computer & Communication Engineering |
| 論文出版年: | 2019 |
| 畢業學年度: | 107 |
| 語文別: | 英文 |
| 論文頁數: | 53 |
| 中文關鍵詞: | 超解析 、生成對抗網路 、卷積神經網路 |
| 外文關鍵詞: | super-resolution, generative adversarial network, convolutional neural |
| 相關次數: | 點閱:89 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
影像超解析技術的主要目的是利用低解析度影像產生高解析度影像,並盡可能的保留影像資訊、維持自然的紋理樣貌。在本論文中提出了一個基於循環生成對抗網路的影像超解析演算法。此方法中有兩組生成對抗網路,其中一組著重在由低解析度影像生成高解析度影像,另一組則是用來增加資料多樣性,以強化整體網路的效能。經過訓練之後,得到一個低解析度影像與高解析度影像的非線性映射關係。
實驗結果顯示,本論文提出的影像超解析演算法,相較於其他方法,在客觀影像評估標準以及主觀的影像品質上皆有較好的表現。
The main purpose of super-resolution technology is to generate high-resolution (HR) images from low-resolution (LR) images, and preserve image information as much as possible to maintain natural texture.
In this Thesis, a super-resolution algorithm based on the cycle generative adversarial network is proposed. There are two generative adversarial networks, one of which focuses on generating high-resolution images from low-resolution images, and another is used to increase data diversity to enhance performance. A nonlinear mapping relationship between low-resolution images and high-resolution images is calculated after training.
The experimental results show that the proposed approach has better performance than other methods on subjective visual quality and objective measurement
[1] R. Y. Tsai and T. S. Huang, "Multiframe image restoration and registration," in Advance Computer Visual and Image Processing, Greenwich, JAI Press, pp.317-339, 1984.
[2] H. S. Hou, and C. Andrews, “Cubic splines for image interpolation and digital filtering,” IEEE Transactions on acoustics, speech, and signal processing, vol.26 no.6, pp. 508-519, 1978.
[3] C. E. Duchon, "Lanczos filtering in one and two dimensions." Journal of applied meteorology vol.18, no.8 pp.1016-1022, 1979.
[4] S. Dai, M. Han, T. Wu and Y. Gong, "Bilateral back-projection for single image super resolution," IEEE International Conference on Multimedia and Expo, pp. 1039-1042. ,2007.
[5] W. Dong, L. Zhang, G. Shi, and X. Wu, "Nonlocal back-projection for adaptive image enlargement," IEEE International Conference on Image Processing. Vols 1-6, pp.349-352, 2009.
[6] T. C. Wee, and J. S. Hyun, "A novel fast fractal super resolution technique." IEEE Transactions on Consumer Electronics. vol 56, no.3, pp.1537-1541, 2010.
[7] W. T. Freeman, T. R. Jones, and E. C. Pasztor, “Learning low-level vision,” International journal of computer vision. Vol.40, no.1, pp.24-47, 2000.
[8] H. Chang, D. Y. Yeung and Y. Xiong, "Super-resolution through neighbor embedding," Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Vol. 1, pp.275-282, 2004.
[9] S. H. Hong, R. H. Park, S. Yang, and J. Y. Kim, "Image interpolation using interpolative classified vector quantization," Image and Vision Computing, vol.2, no.2, pp.109-122, 2016.
[10] C. Dong, C. C. Loy, K. He, and X. Tang, "Learning a deep convolutional network for image super-resolution,” European conference on computer vision. pp.184-199, 2014.
[11] W. Shi et al, "Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network," Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1874-1883, 2016.
[12] C. Ledig, et al. "Photo-realistic single image super-resolution using a generative adversarial network," Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4681-4690, 2017.
[13] K. Fukushima. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biological cybernetics, 1980.
[14] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel. Backpropagation applied to handwritten zip code recognition. Neural computation, 1989.
[15] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations (ICLR), 2015.
[16] A. Krizhevsky, I. Sutskever, and G. E. Hinton, "Imagenet classification with deep convolutional neural networks," in Advances in neural information processing systems, pp. 1097-1105, 2012.
[17] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. “Generative adversarial nets,” In Neural Information Processing Systems (NIPS), 2014.
[18] Z. Yi, H. Zhang, T. Gong, Tan, and M. Gong. “Dualgan: Unsupervised dual learning for image-to-image translation,” In IEEE International Conference on Computer Vision (ICCV), 2017.
[19] J.-Y. Zhu, R. Zhang, D. Pathak, T. Darrell, A. A. Efros, O. Wang, and E. Shechtman. Toward multimodal image-to-image translation. In Advances in Neural Information Processing Systems (NIPS), 2017.
[20] H. Zhang, V. A. Sindagi, and V. M. Patel. Image de-raining using a conditional generative adversarial network. arXiv preprint arXiv:1701.05957, 2017.
[21] J. Zhu, T. Park, P. Isola, and A. A. Efros. “Unpaired image-to-image translation using cycle-consistent adversarial networks,” In International Conference on Computer Vision (ICCV), to appear, 2017.
[22] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. In IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[23] A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, W. Wang, and R. Webb. “Learning from simulated and unsupervised images through adversarial training,” The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
[24] A. Odena, C. Olah, and J. Shlens. “Conditional Image Synthesis With Auxiliary Classifier GANs,” ArXiv e-prints, Oct. 2016.
[25] O. Ronneberger, P. Fischer, and T. Brox, "U-net: Convolutional networks for biomedical image segmentation," International Conference on Medical image computing and computer-assisted intervention. pp. 234-241, 2015.
[26] J. Johnson, A. Alahi, and L. Fei-Fei. “Perceptual losses for real-time style transfer and super-resolution,” In Proceedings of the European Conference on Computer Vision (ECCV), pages 694–711. Springer, 2016.
[27] Guilin Liu, Fitsum A Reda, Kevin J Shih, Ting-Chun Wang, Andrew Tao, and Bryan Catanzaro. Image inpainting for irregular holes using partial convolutions. arXiv preprint arXiv:1804.07723, 2018.
[28] Lin, Tsung-Yi, et al. "Microsoft coco: Common objects in context," European conference on computer vision. Springer, Cham, 2014.
[29] M. Bevilacqua, A. Roumy, C. Guillemot, and M. L. Alberi-Morel. “Low-complexity single-image super-resolution based on nonnegative neighbor embedding,” BMVC, 2012.
[30] R. Zeyde, M. Elad, and M. Protter. “On single image scale-up using sparse-representations,” In Curves and Surfaces, pages 711–730. Springer, 2012.
[31] tegg89’s github: https://github.com/tegg89/SRCNN-Tensorflow
[32] JGuillaumin’s github: https://github.com/JGuillaumin/SuperResGAN-keras
校內:2024-06-30公開