簡易檢索 / 詳目顯示

研究生: 陳仕凱
Chen, Shih-Kai
論文名稱: 基於音樂氛圍的舞台燈光自動化調控模式
A methodology for stage lighting control based on music emotion feeling
指導教授: 蕭世文
Hsiao, Shih-Wen
學位類別: 碩士
Master
系所名稱: 規劃與設計學院 - 工業設計學系
Department of Industrial Design
論文出版年: 2015
畢業學年度: 103
語文別: 英文
論文頁數: 99
中文關鍵詞: 自動化舞台燈光調控模式音樂情感辨識情感與燈光調控色彩支持向量回歸(SVR)音樂段落辨識
外文關鍵詞: automatic stage-lighting regulation, music emotion recognition, lighting color regulation based on music emotion and genre, support vector regression (SVR), automatic music segment detection
相關次數: 點閱:213下載:12
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 傳統燈光調控需要經過專業訓練的燈光人員來進行操作,多數控台工作人員在演出前花大量的時間將燈光與音樂搭配的序列先製成MIDI檔,同樣的演出時間卻需要兩到三倍的時間來進行前置,可以說是相當的費工費時,因此,一種電腦輔助自動化舞台燈光調控模式確實是眾望所歸的。
    有鑑於音樂情感辨識的研究的成熟,以及類神經網路監督式學習機的廣泛發展和應用的基礎,漸漸對於音樂情感(氛圍)也能進行量化的描述及電腦模擬,為達成音樂情感辨識,以音樂特徵對映到賽耶情感平面(Thayer model)上,產生出線性量化的音樂情感描述值,本文收集來自於Musicovery網站點播率最高之2087首之歌曲20秒音樂片段擷取21種音樂特徵,利用主成分分析法(Principle component analysis;PCA)進行降維後,以支持向量機(Support vector machine;SVM)分類器進行特徵的交叉訓練以得到最準確之音樂特徵組合,利用支持向量回歸(Support vector regression;SVR)進行回歸訓練。另一方面,進行音樂情感燈光調控實驗並研究音樂情感與燈光調控色彩調控趨向,同樣利用支持向量回歸模擬以上結果,亦考慮音樂段落間情感氛圍和強度感受上的差異,依據音樂力度發展一套音樂段落辨識方法論,做為情感辨識和燈光亮度的指標,其後進一步地加入音樂風格和燈光色彩因素,建立一套符合音樂風格和情感,依據音樂段落進行舞台燈光調控的自動化系統。為驗證本文發展的自動化燈光調控模式,本研究邀請十名受測者進行驗證問卷,結果顯示本研究開發之燈光音樂自動化搭配,確實可以依據音樂風格和段落情感給予合適的燈光調控引發觀賞時更多的娛樂性。

    Traditionally, the stage-lighting regulation requires professionally trained technicians to operate. However, the contemporary requirements of higher-quality performance, making this work needs more preparation before the performance. Technicians or club DJ spends two to three more times before the show to make the lighting control sequence MIDI file to match the music. It is really waste of time. Thus, A methodology for automatic stage-lighting regulation would be helped.
    Music emotion recognition (MER) got much development these years, so as neural network algorithms. Music feeling has been able to be recognized and even been quantifiable by a supervised machine learning approach. In this paper, A variety of music signal features from 2087 song clips were captured and been selected the main features which are related to music emotion reflected to Thayer's emotion plane in order to produce a linear quantitative value describing music emotion. After that, the music emotion and color preferences of stage-lighting were studied. Using the experimental results trained a support vector regression (SVR) to construct simulations. To be more realistic, we developed an automatic music segment detected methodology based on music signal intensity to present different music strength and feeling of each segment. Furthermore, The factor of music genre has been studied, comprehensively develops an automatic stage-lighting based on feeling, genre and intensity of each segment of music.

    摘要 I SUMMARY II ACKNOWLEDGENTS III TABLE OF CONTENT IV LIST OF TABLES VI LIST OF FIGURES VII CHAPTER 1 INTRODUCTION 1 1.1 Motivation 1 1.1.1 Music emotion recognition 2 1.1.2 Lighting colors and emotional feeling 4 1.2 Purpose 5 1.3 Limitation 7 1.4 Research Framework 7 CHAPTER 2 LITERATURE 12 2.1 Emotion Model 12 2.2 Music Emotion Recognition 15 2.2.1 Music feature extraction 15 2.2.2 Supervised learning machine 16 2.3 Connection between lighting color and emotion 17 CHAPTER 3 THEORETICAL FRAMEWORK 20 3.1 Music Emotion Recognition 20 3.1.1 Audio analysis and processing 21 3.1.2 Music feature extraction 21 3.1.3 Feature dimensions reduction 24 3.1.4 Support vector machine (SVM) 25 3.1.5 Optimal kernel function parameter selection 28 3.1.6 Support vector regression (SVR) 29 3.2 Automatic Lighting Regulation Methodology 33 3.2.1 Automatic music segment detection 34 3.2.2 Audio peak and valley detection 37 3.2.3 Audio onset detection 38 CHAPTER 4 RESEARCH PROCEDURES 40 4.1 Music Emotion Recognition 40 4.1.1 Steps 40 4.1.2 Experiment music samples selecting 41 4.1.3 Music features extracting 42 4.1.4 Music feature dimensions reducing 43 4.1.5 Emotion related features selecting 44 4.1.6 Music emotion recognition SVR building 45 4.2 Lighting Color Regulation Experiment 47 4.2.1 Steps 47 4.2.2 Experiment music sample selecting 48 4.2.3 Experiment system 49 4.2.4 Experiment operation 54 4.2.5 Experiment result analysis 56 4.2.6 Lighting color regulation SVR building 59 4.2.7 Lighting color emotion map 63 4.3 Lighting Color Regulation With Music Genre 64 4.3.1 Steps 64 4.3.2 Experiment result analysis 65 4.3.3 Lighting color regulation adding music genre factor SVR building 68 4.3.4 Lighting color emotion map of each music genre 72 CHAPTER 5 MODE DISCUSSION 74 5.1 Automatic Lighting Regulation Program Structure 74 5.2 Automatic Lighting Regulation Program Operation 76 5.3 Music Genre Recognition 80 5.3.1 Music genre related feature selecting 81 5.3.2 Correlation testing 82 5.3.3 Similarity testing 82 5.4 Automatic Music Segment Detection And Lighting Brightness Regulation Approach 82 5.5 Music Emotion and Lighting Color Regulation 84 5.6 Lighting Regulation Simulation Video Construction 85 5.6.1 Image blending approach 86 5.7 Case Studies 88 CHAPTER 6 CONCLUSIONS AND RECOMMENDATIONS 90 REFERENCE 93

    Ajmera, J., McCowan, I., & Bourlard, H.. (2003). Speech/music segmentation using entropy and dynamism features in a HMM classification framework. Speech Communication, 40(3), 351-363. doi: Doi 10.1016/S0167-6393(02)00087-0
    Bigand, E., Vieillard, S., Madurell, F., Marozeau, J., & Dacquet, A.. (2005). Multidimensional scaling of emotional responses to music: The effect of musical expertise and of the duration of the excerpts. Cognition & Emotion, 19(8), 1113-1139.
    Caivano, J. L.. (1994). Color and Sound - Physical and Psychophysical Relations. Color Research and Application, 19(2), 126-133.
    Camps-Valls, G., Gómez-Chova, L., Calpe, J., Soria, E., Martín, J. D., Alonso, L., & Moreno, J.. (2004). Robust support vector method for hyperspectral data classification and knowledge discovery. IEEE Trans. Geosci. Remote Sens., 42(7), 1530-1542.
    Dhanalakshmi, P., Palanivel, S., & Ramalingam, V.. (2009). Classification of audio signals using SVM and RBFNN. Expert Systems with Applications, 36(3), 6069-6075. doi: DOI 10.1016/j.eswa.2008.06.126
    Duda, R. O., Hart, P. E., & Stork, D. G.. (2000). Pattern Recognition. New York: Wiley.
    Esmaili, S., Krishnan, S., & Raahemifar, K.. (2004). Content based audio classification and retrieval using joint time–frequency analysis. Paper presented at the IEEE international conference on acoustics, speech and signal processing.
    Farnsworth, P. R.. (1954). A study of the hevner adjective list. The Journal of Aesthetics and Art Criticism, 13(1), 97-103.
    Feng, Y., Zhuang, Y., & Pan, Y.. (2003). Popular music retrieval by detecting mood. Paper presented at the Proceedings of the 26th annual international ACM SIGIR conference on Research and development in information retrieval, Toronto, ON, Canada.
    Hsiao, S. W.. (1994). Fuzzy set theory on car-color design. Color Research & Application, 19(3), p.202-p.213.
    Hsiao, S. W., Chiu, F. Y., & Hsu, H. Y.. (2008). A computer-assisted colour selection system based on aesthetic measure for colour harmony and fuzzy logic theory. Color Research & Application, 33(5), p.411-p.423.
    Hsiao, S. W., Hsu, C. F., & Tang, K. W.. ( 2013). A consultation and simulation system for product color planning based on interactive genetic algorithms. Color Research & Application, 38(5), p.475-p.390.
    Huron, D.. (1992). The Ramp Archetype and the Maintenance of Passive Auditory Attention. Music Perception, 10(1), 83-92.
    Juslin, P. N.. (2000). Cue utilization in communication of emotion in music performance: relating performance to perception. Paper presented at the J. Exper. Psychol.: Human Percept. Perf..
    Juslin, P. N., & Laukka, P.. (2004). Expression, perception, and induction of musical emotions: A review and a questionnaire study of everyday listening. Journal of New Music Research, 33(3), 217-238. doi: Doi 10.1080/0929821042000317813
    Kaya, N., & Epps, H.. (2004). Relationship between Color and Emotion: A Study of College Students. College Student Journal, 38, 396-405.
    Kreurz, G.. (2000). Basic emotions in music. Paper presented at the Proc. 6th Int. Conf. Music Perception Cognition.
    Krumhansl, C. L.. (1997). An exploratory study of musical emotions and psychophysiology. Canadian Journal of Experimental Psychology-Revue Canadienne De Psychologie Experimentale, 51(4), 336-353. doi: Doi 10.1037/1196-1961.51.4.336
    Laurier, C., & Herrera, P.. (2007). Audio music mood classification using support vector machine. Paper presented at the Proceedings of the 8th International Conference on Music Information Retrieval, Vienna, Austria.
    Laurier, C., Meyers, O., Serr`a, J., Blech, M., & Herrera, P.. (2009). Music Mood Annotator Design and Integration.
    Lee, C. H., Shih, J. L., Yu, K. M., & Lin, H. S.. (2009). Automatic Music Genre Classification Based on Modulation Spectral Analysis of Spectral and Cepstral Features. IEEE Transactions on Multimedia, 11(4), 670-682. doi: Doi 10.1109/Tmm.2009.2017635
    Li, C. H.. (2010). An automatic method for selecting the parameter of the RBF kernel function to support vector machines. Honolulu, HI
    Li, D. G., Sethi, I. K., Dimitrova, N., & McGee, T.. (2001). Classification of general audio data for content-based retrieval. Pattern Recognition Letters, 22(5), 533-544. doi: Doi 10.1016/S0167-8655(00)00119-7
    Li, T., & Ogihara, M.. (2003). Detecting emotion in music. Paper presented at the Proceedings of the 4th International Conference on Music Information Retrieval, Baltimore, MD, USA.
    Li, T., & Ogihara, M.. (2004, 17-21 May). Content-based music similarity search and emotion detection. Paper presented at the Int. Conf. Acoust., Speech, Signal Process, Toulouse, France.
    Li, T., & Ogihara, M.. (2006). Toward intelligent music information retrieval. IEEE Transactions on Multimedia, 8(3), 564-574. doi: Doi 10.1109/Tmm.2006.870730
    Lie, L., Liu, D., & Zhang, H. J.. (2006). Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio Speech and Language Processing, 14(1), 5-18. doi: Doi 10.1109/Tsa.2005.860344
    Lindstrom, E., & Juslin, P. N.. (2003). Expressivity Comes From Within Your Soul: A Questionnaire Study of Student's Perception on Musical Expressivity. Research Studies in Music Education, 20, 23-47.
    Mammone, R. J., Zhang, X., & Ramachandran, R. P.. (1996). Robust speaker recognition: a feature-based approach. IEEE Signal Processing Magazine, 13(5), 1053-5888.
    Mandel, M. I., Poliner, G. E., & Ellis, D. P. W.. (2006). Support vector machine active learning for music retrieval. Multimedia Systems, 12(1), 3-13. doi: DOI 10.1007/s00530-006-0032-2
    Marks, L. E.. (1997). On colored-hearing Synesthesia: Crossmodal Translations of Sensory Dimension. Paper presented at the Classic and contemporary readings, In S. Baron-Cohen, J. E. Harrison, eds. Synesthesia.
    Miller, Mary C. (1997). Color for Interior Architecture (1st ed.). U.S.A. NY.
    Nagamachi, M. (1995). Kansei engineering: a new ergonomic consumer-oriented technology for product development. International Journal of Industrial Ergonomics, 15(1), 3-11.
    Nissen, L., Faulkner, R. D., & Faulkner, S.. (1999). Inside Today's Home. Hillsboro: Goodwill Books.
    Ou, L. C., Luo, M. R., Woodcock, A., & Wright, A.. (2004). A study of colour emotion and colour preference. Part I: Colour emotions for single colours. Color Research & Application, 29(3), p.232-p.240.
    Park, D. C.. (2009). Classification of audio signals using Fuzzy c-Means with divergence-based Kernel. Pattern Recognition Letters, 30(9), 794-798. doi: DOI 10.1016/j.patrec.2008.05.019
    Peretz, I., Gagnon, L., & Bouchard, B.. (1998). Music and emotion: perceptual determinants, immediacy, and isolation after brain damage. Cognition, 68(2), 111-141. doi: Doi 10.1016/S0010-0277(98)00043-2
    Picard, R. W., & Cosier, G.. (1997). Affective intelligence - the missing link?. Bt Technology Journal, 15(4), 150-161.
    Pridmore, R. W.. (1992). Music and Color - Relations in the Psychophysical Perspective. Color Research and Application, 17(1), 57-61. doi: DOI 10.1002/col.5080170110
    Roberts, Lawrence G. (1980). Machine perception of three-dimensional solids. New York: Garland Pub.
    Russell, J. A.. (1980). A Circumplex Model of Affect. Journal of Personality and Social Psychology, 39(6), 1161-1178. doi: Doi 10.1037/H0077714
    Sebba, R.. (1991). Structural Correspondence between Music and Color. Color Research and Application, 16(2), 81-88. doi: DOI 10.1002/col.5080160206
    Sen, A., & Srivastava, M.. (1990). Regression Analysis: Theory, Methods, and Applications. New York: Springer.
    Seo, C., Lee, K. Y., & Lee, J.. (2001). GMM based on local PCA for speaker identification. Electronics Letters, 37(24), 1486 - 1488.
    Shao, B., Wang, D. D., Li, T., & Ogihara, M.. (2009). Music Recommendation Based on Acoustic Features and User Access Patterns. IEEE Transactions on Audio Speech and Language Processing, 17(8), 1602-1611. doi: Doi 10.1109/Tasl.2009.2020893
    Shi, Y. Y., Zhu, X., Kim, H. G., & Eom, K. W.. (2006). A tempo feature via modulation spectrum analysis and its application to music emotion classification. Paper presented at the Proceedings of the IEEE International Conference on Multimedia and Expo, Toronto, Canada.
    Skowronek, J., McKinney, M. F., & van de Par, S.. (2007). Demonstrator for automatic music mood estimation. Paper presented at the The International Conference on Music Information Retrieval, Vienna, Austria.
    Smola, A. J., & Schölkopf, B.. (2004). A tutorial on support vector regression. Statist. Comput, 199-222.
    Solomatine, D. P., & Shrestha, D. L. (2004). AdaBoost.RT: A boosting algorithm for regression problems.. Paper presented at the Proc. IEEE Int. Joint Conf., Neural Netw.
    Sordo, M., Laurier, C., & Celma, O.. (2007). Annotating music collections: How content-based similarity helps to propagate labels. Paper presented at the Proceedings of the 8th International Conference on Music Information Retrieval, Vienna, Austria.
    Tay, Francis E. H., & Cao, L.. (2001). Application of support vector machines in financial time series forecasting. Omega, 29(4), 309-317.
    Thayer, R. E.. (1996). The Origin of Everyday Moods: Managing Energy, Tension, and Stress. Oxford: Oxford University Press.
    Tzanetakis, G., & Cook, P.. (2002). Musical genre classification of audio signals. Ieee Transactions on Speech and Audio Processing, 10(5), 293-302. doi: Doi 10.1109/Tsa.2002.800560
    Umapathy, K., Krishnan, S., & Jimaa, S.. (2005). Multigroup classification of audio signals using time-frequency parameters. IEEE Transactions on Multimedia, 7(2), 308-315. doi: Doi 10.1109/Tmm.2005.843363
    Valdez, P., & Mehrabian, A.. (1994). Effects of Color on Emotions. Journal of Experimental Psychology-General 123(4), 394-409. doi: Doi 10.1037/0096-3445.123.4.394
    Vapnik, V.. (1995). The nature of statistical learning theorySpringer. New York.
    Vapnik, V., Goldwich, S., & Smola, A.. (1997). Support vector method for function approximation, regression estimation, and signal processing: Cambride:MIT press.
    Velius, G.. (1988). Variants of cepstrum based speaker identity verification. Paper presented at the Acoustics, Speech, and Signal Processing New York, NY.
    VieIllard, S., Peretz, I., Gosselin, N., Khalfa, S., Gagnon, L., & Bouchard, B.. (2008). Happy, sad, scary and peaceful musical excerpts for research on emotions. Cognition & Emotion, 22(4), 720-752. doi: Doi 10.1080/02699930701503567
    Xu, C. S., Maddage, N. C., & Shao, X.. (2005). Automatic music classification and summarization. IEEE Transactions on Speech and Audio Processing, 13(3), 441-450. doi: Doi 10.1109/Tsa.2004.840939
    Yang, Y. H., Lin, Y. C., Su, Y. F., & Chen, H. H.. (2008). A regression approach to music emotion recognition. IEEE Transactions on Audio Speech and Language Processing, 16(2), 448-457. doi: Doi 10.1109/Tasl.2007.911513
    Zhu, X., Shi, Y. Y., Kim, H. G., & Eom, K. W.. (2006). An integrated music recommendation system. IEEE Transactions on Consumer Electronics, 52(3), 917-925.

    下載圖示 校內:2016-02-12公開
    校外:2018-02-12公開
    QR CODE