| 研究生: |
陳仕凱 Chen, Shih-Kai |
|---|---|
| 論文名稱: |
基於音樂氛圍的舞台燈光自動化調控模式 A methodology for stage lighting control based on music emotion feeling |
| 指導教授: |
蕭世文
Hsiao, Shih-Wen |
| 學位類別: |
碩士 Master |
| 系所名稱: |
規劃與設計學院 - 工業設計學系 Department of Industrial Design |
| 論文出版年: | 2015 |
| 畢業學年度: | 103 |
| 語文別: | 英文 |
| 論文頁數: | 99 |
| 中文關鍵詞: | 自動化舞台燈光調控模式 、音樂情感辨識 、情感與燈光調控色彩 、支持向量回歸(SVR) 、音樂段落辨識 |
| 外文關鍵詞: | automatic stage-lighting regulation, music emotion recognition, lighting color regulation based on music emotion and genre, support vector regression (SVR), automatic music segment detection |
| 相關次數: | 點閱:213 下載:12 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
傳統燈光調控需要經過專業訓練的燈光人員來進行操作,多數控台工作人員在演出前花大量的時間將燈光與音樂搭配的序列先製成MIDI檔,同樣的演出時間卻需要兩到三倍的時間來進行前置,可以說是相當的費工費時,因此,一種電腦輔助自動化舞台燈光調控模式確實是眾望所歸的。
有鑑於音樂情感辨識的研究的成熟,以及類神經網路監督式學習機的廣泛發展和應用的基礎,漸漸對於音樂情感(氛圍)也能進行量化的描述及電腦模擬,為達成音樂情感辨識,以音樂特徵對映到賽耶情感平面(Thayer model)上,產生出線性量化的音樂情感描述值,本文收集來自於Musicovery網站點播率最高之2087首之歌曲20秒音樂片段擷取21種音樂特徵,利用主成分分析法(Principle component analysis;PCA)進行降維後,以支持向量機(Support vector machine;SVM)分類器進行特徵的交叉訓練以得到最準確之音樂特徵組合,利用支持向量回歸(Support vector regression;SVR)進行回歸訓練。另一方面,進行音樂情感燈光調控實驗並研究音樂情感與燈光調控色彩調控趨向,同樣利用支持向量回歸模擬以上結果,亦考慮音樂段落間情感氛圍和強度感受上的差異,依據音樂力度發展一套音樂段落辨識方法論,做為情感辨識和燈光亮度的指標,其後進一步地加入音樂風格和燈光色彩因素,建立一套符合音樂風格和情感,依據音樂段落進行舞台燈光調控的自動化系統。為驗證本文發展的自動化燈光調控模式,本研究邀請十名受測者進行驗證問卷,結果顯示本研究開發之燈光音樂自動化搭配,確實可以依據音樂風格和段落情感給予合適的燈光調控引發觀賞時更多的娛樂性。
Traditionally, the stage-lighting regulation requires professionally trained technicians to operate. However, the contemporary requirements of higher-quality performance, making this work needs more preparation before the performance. Technicians or club DJ spends two to three more times before the show to make the lighting control sequence MIDI file to match the music. It is really waste of time. Thus, A methodology for automatic stage-lighting regulation would be helped.
Music emotion recognition (MER) got much development these years, so as neural network algorithms. Music feeling has been able to be recognized and even been quantifiable by a supervised machine learning approach. In this paper, A variety of music signal features from 2087 song clips were captured and been selected the main features which are related to music emotion reflected to Thayer's emotion plane in order to produce a linear quantitative value describing music emotion. After that, the music emotion and color preferences of stage-lighting were studied. Using the experimental results trained a support vector regression (SVR) to construct simulations. To be more realistic, we developed an automatic music segment detected methodology based on music signal intensity to present different music strength and feeling of each segment. Furthermore, The factor of music genre has been studied, comprehensively develops an automatic stage-lighting based on feeling, genre and intensity of each segment of music.
Ajmera, J., McCowan, I., & Bourlard, H.. (2003). Speech/music segmentation using entropy and dynamism features in a HMM classification framework. Speech Communication, 40(3), 351-363. doi: Doi 10.1016/S0167-6393(02)00087-0
Bigand, E., Vieillard, S., Madurell, F., Marozeau, J., & Dacquet, A.. (2005). Multidimensional scaling of emotional responses to music: The effect of musical expertise and of the duration of the excerpts. Cognition & Emotion, 19(8), 1113-1139.
Caivano, J. L.. (1994). Color and Sound - Physical and Psychophysical Relations. Color Research and Application, 19(2), 126-133.
Camps-Valls, G., Gómez-Chova, L., Calpe, J., Soria, E., Martín, J. D., Alonso, L., & Moreno, J.. (2004). Robust support vector method for hyperspectral data classification and knowledge discovery. IEEE Trans. Geosci. Remote Sens., 42(7), 1530-1542.
Dhanalakshmi, P., Palanivel, S., & Ramalingam, V.. (2009). Classification of audio signals using SVM and RBFNN. Expert Systems with Applications, 36(3), 6069-6075. doi: DOI 10.1016/j.eswa.2008.06.126
Duda, R. O., Hart, P. E., & Stork, D. G.. (2000). Pattern Recognition. New York: Wiley.
Esmaili, S., Krishnan, S., & Raahemifar, K.. (2004). Content based audio classification and retrieval using joint time–frequency analysis. Paper presented at the IEEE international conference on acoustics, speech and signal processing.
Farnsworth, P. R.. (1954). A study of the hevner adjective list. The Journal of Aesthetics and Art Criticism, 13(1), 97-103.
Feng, Y., Zhuang, Y., & Pan, Y.. (2003). Popular music retrieval by detecting mood. Paper presented at the Proceedings of the 26th annual international ACM SIGIR conference on Research and development in information retrieval, Toronto, ON, Canada.
Hsiao, S. W.. (1994). Fuzzy set theory on car-color design. Color Research & Application, 19(3), p.202-p.213.
Hsiao, S. W., Chiu, F. Y., & Hsu, H. Y.. (2008). A computer-assisted colour selection system based on aesthetic measure for colour harmony and fuzzy logic theory. Color Research & Application, 33(5), p.411-p.423.
Hsiao, S. W., Hsu, C. F., & Tang, K. W.. ( 2013). A consultation and simulation system for product color planning based on interactive genetic algorithms. Color Research & Application, 38(5), p.475-p.390.
Huron, D.. (1992). The Ramp Archetype and the Maintenance of Passive Auditory Attention. Music Perception, 10(1), 83-92.
Juslin, P. N.. (2000). Cue utilization in communication of emotion in music performance: relating performance to perception. Paper presented at the J. Exper. Psychol.: Human Percept. Perf..
Juslin, P. N., & Laukka, P.. (2004). Expression, perception, and induction of musical emotions: A review and a questionnaire study of everyday listening. Journal of New Music Research, 33(3), 217-238. doi: Doi 10.1080/0929821042000317813
Kaya, N., & Epps, H.. (2004). Relationship between Color and Emotion: A Study of College Students. College Student Journal, 38, 396-405.
Kreurz, G.. (2000). Basic emotions in music. Paper presented at the Proc. 6th Int. Conf. Music Perception Cognition.
Krumhansl, C. L.. (1997). An exploratory study of musical emotions and psychophysiology. Canadian Journal of Experimental Psychology-Revue Canadienne De Psychologie Experimentale, 51(4), 336-353. doi: Doi 10.1037/1196-1961.51.4.336
Laurier, C., & Herrera, P.. (2007). Audio music mood classification using support vector machine. Paper presented at the Proceedings of the 8th International Conference on Music Information Retrieval, Vienna, Austria.
Laurier, C., Meyers, O., Serr`a, J., Blech, M., & Herrera, P.. (2009). Music Mood Annotator Design and Integration.
Lee, C. H., Shih, J. L., Yu, K. M., & Lin, H. S.. (2009). Automatic Music Genre Classification Based on Modulation Spectral Analysis of Spectral and Cepstral Features. IEEE Transactions on Multimedia, 11(4), 670-682. doi: Doi 10.1109/Tmm.2009.2017635
Li, C. H.. (2010). An automatic method for selecting the parameter of the RBF kernel function to support vector machines. Honolulu, HI
Li, D. G., Sethi, I. K., Dimitrova, N., & McGee, T.. (2001). Classification of general audio data for content-based retrieval. Pattern Recognition Letters, 22(5), 533-544. doi: Doi 10.1016/S0167-8655(00)00119-7
Li, T., & Ogihara, M.. (2003). Detecting emotion in music. Paper presented at the Proceedings of the 4th International Conference on Music Information Retrieval, Baltimore, MD, USA.
Li, T., & Ogihara, M.. (2004, 17-21 May). Content-based music similarity search and emotion detection. Paper presented at the Int. Conf. Acoust., Speech, Signal Process, Toulouse, France.
Li, T., & Ogihara, M.. (2006). Toward intelligent music information retrieval. IEEE Transactions on Multimedia, 8(3), 564-574. doi: Doi 10.1109/Tmm.2006.870730
Lie, L., Liu, D., & Zhang, H. J.. (2006). Automatic mood detection and tracking of music audio signals. IEEE Transactions on Audio Speech and Language Processing, 14(1), 5-18. doi: Doi 10.1109/Tsa.2005.860344
Lindstrom, E., & Juslin, P. N.. (2003). Expressivity Comes From Within Your Soul: A Questionnaire Study of Student's Perception on Musical Expressivity. Research Studies in Music Education, 20, 23-47.
Mammone, R. J., Zhang, X., & Ramachandran, R. P.. (1996). Robust speaker recognition: a feature-based approach. IEEE Signal Processing Magazine, 13(5), 1053-5888.
Mandel, M. I., Poliner, G. E., & Ellis, D. P. W.. (2006). Support vector machine active learning for music retrieval. Multimedia Systems, 12(1), 3-13. doi: DOI 10.1007/s00530-006-0032-2
Marks, L. E.. (1997). On colored-hearing Synesthesia: Crossmodal Translations of Sensory Dimension. Paper presented at the Classic and contemporary readings, In S. Baron-Cohen, J. E. Harrison, eds. Synesthesia.
Miller, Mary C. (1997). Color for Interior Architecture (1st ed.). U.S.A. NY.
Nagamachi, M. (1995). Kansei engineering: a new ergonomic consumer-oriented technology for product development. International Journal of Industrial Ergonomics, 15(1), 3-11.
Nissen, L., Faulkner, R. D., & Faulkner, S.. (1999). Inside Today's Home. Hillsboro: Goodwill Books.
Ou, L. C., Luo, M. R., Woodcock, A., & Wright, A.. (2004). A study of colour emotion and colour preference. Part I: Colour emotions for single colours. Color Research & Application, 29(3), p.232-p.240.
Park, D. C.. (2009). Classification of audio signals using Fuzzy c-Means with divergence-based Kernel. Pattern Recognition Letters, 30(9), 794-798. doi: DOI 10.1016/j.patrec.2008.05.019
Peretz, I., Gagnon, L., & Bouchard, B.. (1998). Music and emotion: perceptual determinants, immediacy, and isolation after brain damage. Cognition, 68(2), 111-141. doi: Doi 10.1016/S0010-0277(98)00043-2
Picard, R. W., & Cosier, G.. (1997). Affective intelligence - the missing link?. Bt Technology Journal, 15(4), 150-161.
Pridmore, R. W.. (1992). Music and Color - Relations in the Psychophysical Perspective. Color Research and Application, 17(1), 57-61. doi: DOI 10.1002/col.5080170110
Roberts, Lawrence G. (1980). Machine perception of three-dimensional solids. New York: Garland Pub.
Russell, J. A.. (1980). A Circumplex Model of Affect. Journal of Personality and Social Psychology, 39(6), 1161-1178. doi: Doi 10.1037/H0077714
Sebba, R.. (1991). Structural Correspondence between Music and Color. Color Research and Application, 16(2), 81-88. doi: DOI 10.1002/col.5080160206
Sen, A., & Srivastava, M.. (1990). Regression Analysis: Theory, Methods, and Applications. New York: Springer.
Seo, C., Lee, K. Y., & Lee, J.. (2001). GMM based on local PCA for speaker identification. Electronics Letters, 37(24), 1486 - 1488.
Shao, B., Wang, D. D., Li, T., & Ogihara, M.. (2009). Music Recommendation Based on Acoustic Features and User Access Patterns. IEEE Transactions on Audio Speech and Language Processing, 17(8), 1602-1611. doi: Doi 10.1109/Tasl.2009.2020893
Shi, Y. Y., Zhu, X., Kim, H. G., & Eom, K. W.. (2006). A tempo feature via modulation spectrum analysis and its application to music emotion classification. Paper presented at the Proceedings of the IEEE International Conference on Multimedia and Expo, Toronto, Canada.
Skowronek, J., McKinney, M. F., & van de Par, S.. (2007). Demonstrator for automatic music mood estimation. Paper presented at the The International Conference on Music Information Retrieval, Vienna, Austria.
Smola, A. J., & Schölkopf, B.. (2004). A tutorial on support vector regression. Statist. Comput, 199-222.
Solomatine, D. P., & Shrestha, D. L. (2004). AdaBoost.RT: A boosting algorithm for regression problems.. Paper presented at the Proc. IEEE Int. Joint Conf., Neural Netw.
Sordo, M., Laurier, C., & Celma, O.. (2007). Annotating music collections: How content-based similarity helps to propagate labels. Paper presented at the Proceedings of the 8th International Conference on Music Information Retrieval, Vienna, Austria.
Tay, Francis E. H., & Cao, L.. (2001). Application of support vector machines in financial time series forecasting. Omega, 29(4), 309-317.
Thayer, R. E.. (1996). The Origin of Everyday Moods: Managing Energy, Tension, and Stress. Oxford: Oxford University Press.
Tzanetakis, G., & Cook, P.. (2002). Musical genre classification of audio signals. Ieee Transactions on Speech and Audio Processing, 10(5), 293-302. doi: Doi 10.1109/Tsa.2002.800560
Umapathy, K., Krishnan, S., & Jimaa, S.. (2005). Multigroup classification of audio signals using time-frequency parameters. IEEE Transactions on Multimedia, 7(2), 308-315. doi: Doi 10.1109/Tmm.2005.843363
Valdez, P., & Mehrabian, A.. (1994). Effects of Color on Emotions. Journal of Experimental Psychology-General 123(4), 394-409. doi: Doi 10.1037/0096-3445.123.4.394
Vapnik, V.. (1995). The nature of statistical learning theorySpringer. New York.
Vapnik, V., Goldwich, S., & Smola, A.. (1997). Support vector method for function approximation, regression estimation, and signal processing: Cambride:MIT press.
Velius, G.. (1988). Variants of cepstrum based speaker identity verification. Paper presented at the Acoustics, Speech, and Signal Processing New York, NY.
VieIllard, S., Peretz, I., Gosselin, N., Khalfa, S., Gagnon, L., & Bouchard, B.. (2008). Happy, sad, scary and peaceful musical excerpts for research on emotions. Cognition & Emotion, 22(4), 720-752. doi: Doi 10.1080/02699930701503567
Xu, C. S., Maddage, N. C., & Shao, X.. (2005). Automatic music classification and summarization. IEEE Transactions on Speech and Audio Processing, 13(3), 441-450. doi: Doi 10.1109/Tsa.2004.840939
Yang, Y. H., Lin, Y. C., Su, Y. F., & Chen, H. H.. (2008). A regression approach to music emotion recognition. IEEE Transactions on Audio Speech and Language Processing, 16(2), 448-457. doi: Doi 10.1109/Tasl.2007.911513
Zhu, X., Shi, Y. Y., Kim, H. G., & Eom, K. W.. (2006). An integrated music recommendation system. IEEE Transactions on Consumer Electronics, 52(3), 917-925.