| 研究生: |
陳行正 Chen, Hsing-Cheng |
|---|---|
| 論文名稱: |
基於亮度不變前景擷取方法和時空體積切割之監視影片摘要 Surveillance Video Synopsis Using Illumination Invariant Foreground Extraction and Volumetric Cut |
| 指導教授: |
詹寶珠
Chung, Pau-Choo |
| 學位類別: |
碩士 Master |
| 系所名稱: |
電機資訊學院 - 電腦與通信工程研究所 Institute of Computer & Communication Engineering |
| 論文出版年: | 2011 |
| 畢業學年度: | 99 |
| 語文別: | 英文 |
| 論文頁數: | 53 |
| 中文關鍵詞: | 影片監控 、影片總結 、影片摘要 |
| 外文關鍵詞: | video surveillance, video summarization, video synopsis |
| 相關次數: | 點閱:62 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
為了監控環境,我們需要連續不間斷地儲存監視影片,而龐大的影片儲存空間和如何有效率地瀏覽長時間影片便成為很重要的問題。為了解決這些問題,我們提出了一種新的監視影片摘要方法,是在即時的前提之下,將前景物體依照時間先後順序重新排列位置與其出現的時間。而為了擷取出前景物體,我們提出了基於亮度不變的前景擷取方法,以及時空的體積切割方法。相較於傳統的影片摘要,我們的方法可以直接應用在線上的串流監視影片,並且在不必篩選過整段影片的前提下就能產生摘要影片。
In order to monitor environments, continuously storing surveillance videos is required. Enormous storage of videos and how to browse long-time videos become important problems. To solve these problems, we propose a novel surveillance video synopsis method to chronologically rearrange positions and appearing frames of foreground objects in real-time. To extract foreground objects, illumination invariant foreground extraction and spatial-temporal volumetric cut are proposed. Comparing with traditional synopsis approaches, our method can directly be applied to online streaming surveillance videos and generate synopsis videos without screening the whole videos.
[1] R. Feris, Y.-L. Tian and A. Hampapur, “Capturing People in Surveillance Video,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, CVPR’07, pp. 1-8, Jun. 2007.
[2] A. Hampapur, L. Brown, R. Feris, A. Senior, C.-F. Shu, Y.-L. Tian, Y. Zhai and M. Lu, “Searching Surveillance Video,” in Proceedings of IEEE Conference on Advanced Video and Signal Based Surveillance, pp. 75-80, Sep. 2007.
[3] Y.-T. Chen and C.-S. Chen, “Fast Human Detection Using a Novel Boosted Cascading Structure With Meta Stages,” IEEE Transactions on Image Processing, vol. 17, no. 8, pp. 1452-1464, 2008.
[4] C.-R. Huang, P.-C. Chung, K.-W. Lin and S.-C. Tseng, “Wheelchair Detection Using Cascaded Decision Tree,” IEEE Transactions on Information Technology in Biomedicine, vol. 14, no. 2, pp. 292- 300, 2010.
[5] W.-H. Cheng, Y.-Y. Chuang, Y.-T. Lin, C.-C. Hsieh, S.-Y. Fang, B.-Y. Chen and J.-L. Wu, “Semantic Analysis for Automatic Event Recognition and Segmentation of Wedding Ceremony Videos,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 18, no. 11, pp. 1639-1650, Nov. 2008.
[6] C. Piciarelli, C. Micheloni and G. L. Foresti, “Trajectory-Based Anomalous Event Detection,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 18, no. 11, pp. 1544-1554, Nov. 2008.
[7] G. Zhu, C. Xu, Q. Huang, Y. Rui, S. Jiang, W. Gao and H. Yao, "Event Tactic Analysis Based on Broadcast Sports Video," IEEE Transactions on Multimedia, vol.11, no.1, pp.49-67, Jan. 2009
[8] L. Liu, Z. Li and E. J. Delp, “Efficient and Low-Complexity Surveillance Video Compression Using Backward-Channel Aware Wyner-Ziv Video Coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 19, no. 4, pp. 453-465, Apr. 2009.
[9] R. V. Babu and A. Makur, “Object-Based Surveillance Video Compression Using Foreground Motion Compensation,” in Proceedings of 9th International Conference on Control, Automation, Robotics and Vision, ICARCV’06, pp. 1-6, Dec. 2006.
[10] C.-R. Huang, H.-P. Lee and C.-S. Chen, “Shot Change Detection via Local Keypoint Matching,” IEEE Transactions on Multimedia, vol. 10, no.6, pp. 1097-1108, 2008.
[11] H. J. Zhang, J. Wu, D. Zhong and S. W. Smoliar, “An Integrated System for Content-Based Video Retrieval and Browsing,” Pattern Recognition, vol. 30, no. 4, pp. 643-658, 1997
[12] M. M. Yeung and B. Liu, “Efficient Matching and Clustering of Video Shots,” in Proceedings of International Conference on Image Processing, vol. 1, pp. 338–341, 1995.
[13] Y. Zhuang, Y. Rui, T. S. Huang and S. Mehrotra, “Adaptive Key Frame Extraction Using Unsupervised Clustering,” in Proceedings of International Conference on Image Processing, ICIP’98, vol. 1, pp. 866–870, 1998.
[14] R. L. Lagendijk, A. Hanjalic, M. Ceccarelli, M. Soletic and E. Persoon, “Visual Search in a Smash System,” in Proceedings of International Conference on Image Processing, ICIP’96, pp. 671-674, Sep. 1996.
[15] S. X. Ju, M. J. Black, S. Minneman and D. Kimber, “Summarization of Video-Taped Presentations: Automatic Analysis of Motion and Gestures,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 8, no. 5, pp. 686-696, Sep. 1998.
[16] C. Li, Y.-T. Wu, S.-S. Yu and T. Chen, “Motion-Focusing Key Frame Extraction and Video Summarization for Lane Surveillance System,” in Proceedings of IEEE International Conference on Image Processing, ICIP’09, pp. 4329-4332, 2009.
[17] Y.-F. Ma, X.-S. Hua, L. Lu and H.-J. Zhang, “A Generic Framework of User Attention Model and Its Application in Video Summarization,” IEEE Transactions on Multimedia, vol. 7, no. 5, pp. 907-919, 2005.
[18] Y. Fu, Y. Guo, Y. Zhu, F. Liu, C. Song and Z.-H. Zhou, “Multi-View Video Summarization,” IEEE Transactions on Multimedia, vol.12, no.7, pp.717-729, 2010.
[19] S. Pfeiffer, R. Lienhart, S. Fischer and W. Effelsberg, “Abstracting Digital Movies Automatically,” Journal of Visual Community and Image Representation, vol. 7, no. 4, pp. 345–353, Dec. 1996.
[20] A. Hanjalic and H. J. Zhang, “An Integrated Scheme for Automated Video Abstraction Based on Unsupervised Cluster-Validity Analysis,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 9, no. 8, pp. 1280-1289, 1999.
[21] A. M. Ferman and A. M. Tekalp, “Multiscale Content Extraction and Representation for Video Indexing,” in Proceedings of SPIE, vol. 3229, pp. 23-31, 1997.
[22] A. Stefanidis, P. Partsinevelos, P. Agouris and P. Doucette, “Summarizing Video Datasets in the Spatiotemporal Domain,” in Proceedings of 11th International Workshop on Database and Expert Systems Applications, pp. 906-912, Sep. 2000.
[23] H. Sundaram and S.-F. Chang, “Condensing Computable Scenes Using Visual Complexity and Film Syntax Analysis,” in Proceedings of IEEE International Conference on Multimedia Expo, ICME’01, pp. 389-392, 2001.
[24] Y. Ma, L. Lu, H. Zhang and M. Li, “A User Attention Model for Video Summarization,” in Proceedings of the Tenth ACM international Conference on Multimedia, ACMMM’02, pp. 533-542, 2002.
[25] Y. Li, S.-H. Lee, C.-H. Yeh and C.-C. J. Kuo, “Techniques for Movie Content Analysis and Skimming - Tutorial and Overview on Video Abstraction Techniques,” IEEE Signal Processing Magazine, vol. 23, no. 2, pp. 79-89, 2006.
[26] B. M. Wildemuth, G. Marchionini, M. Yang, G. Geisler, T. Wilkens, A. Hughes and R. Gruss, “How Fast is Too Fast? Evaluating Fast Forward Surrogates for Digital Video,” in Proceedings of the 3rd ACM/IEEE-CS Joint Conference on Digital Libraries, JCDL’03, pp.221-230, 2003.
[27] E. P. Bennett and L. McMillan, “Computational Time-Lapse Video,” ACM Transactions on Graphics, Special issue on ACM Siggraph 2007, vol. 26, no. 3, 2007.
[28] B. Chen and P. Sen. “Video Carving,” EUROGRAPHICS’08, 2008.
[29] Z. Li, P. Ishwar and J. Konrad, "Video Condensation by Ribbon Carving," IEEE Transactions on Image Processing, vol. 18, no. 11, pp. 2572-2583, Nov. 2009.
[30] A. Rav-Acha, Y. Pritch and S. Peleg, “Making a Long Video Short: Dynamic Video Synopsis,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, CVPR’06, pp. 435-441, 2006.
[31] Y. Pritch, A. Rav-Acha and S. Peleg, “Nonchronological Video Synopsis and Indexing”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, no. 11, pp. 1971-1984, 2008.
[32] M. Xu, S. Z. Li, B. Li, X.-T. Yuan and S.-M. Xiang, “A Set Theoretical Method for Video Synopsis,” in Proceedings of the 1st ACM international conference on Multimedia information retrieval, pp. 366-370 2008.
[33] U. Vural and Y. S. Akgul, “Eye-Gaze Based Real-Time Surveillance Video Synopsis,” Pattern Recognition Letters, vol. 30, no. 12, pp. 1151-1159, 2009.
[34] H. Kang, Y. Matsushita, X. Tang and X. Chen, “Space-Time Video Montage,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 1331-1338, 2006.
[35] V. Kolmogorov and R. Zabih, “What Energy Functions Can be Minimized via Graph Cuts?” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, no. 2, pp. 147-159, Feb. 2004.
[36] Y.-T. Chen, C.-S. Chen, C.-R. Huang and Y.-P. Hung, “Efficient Hierarchical Method for Background Subtraction,” Pattern Recognition, vol. 40, no. 10, pp. 2706-2715, 2007.
[37] C. R. Jung, “Efficient Background Subtraction and Shadow Removal for Monochromatic Video Sequences,” IEEE Transactions on Multimedia, vol.11, no.3, pp.571-577, 2009.
[38] K. Skifstad and R. Jain, “Illumination Independent Change Detection for Real World Image Sequences,” Computer Vision, Graphics, and Image Processing, vol. 46, no. 3, pp. 387–399, Jun. 1989.
[39] D. G. Lowe, "Distinctive Image Features from Scale Invariant Keypoints," International Journal of Computer Vision, vol. 20, no. 2, pp. 91-110, 2004.
[40] C.-R. Huang, C.-S. Chen, and P.-C. Chung, “Contrast Context Histogram - An Efficient Discriminating Local Descriptor for Object Recognition and Image Matching,” Pattern Recognition, vol.41, no.10, pp. 3071-3077, 2008.
[41] W. Hu, X. Zhou, M. Hu and S. Maybank, “Occlusion Reasoning for Tracking Multiple People,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 19, no. 1, pp. 114-121, Jan. 2009.
[42] N.D.F. Campbell, G. Vogiatzis, C. Hernández and R. Cipolla, “Automatic 3D Object Segmentation in Multiple Views Using Volumetric Graph-Cuts,” Image and Vision Computing, vol. 28, no. 1, pp. 14-25, Jan. 2010.
[43] D. G. Bailey, “An Efficient Euclidean Distance Transform,” International Workshop on Combinatorial Image Analysis, pp. 394-408, 2004.
校內:2021-12-31公開