簡易檢索 / 詳目顯示

研究生: 郭雅涵
Kuo, Ya-Han
論文名稱: 複音器樂錄音中獨立單音的TELPC再合成演算法
TELPC Based Re-synthesis Method for Isolated Notes of Polyphonic Instrument Music Recording
指導教授: 蘇文鈺
Su, Wen-Yu
學位類別: 碩士
Master
系所名稱: 電機資訊學院 - 資訊工程學系
Department of Computer Science and Information Engineering
論文出版年: 2013
畢業學年度: 101
語文別: 英文
論文頁數: 37
中文關鍵詞: 再合成TELPC
外文關鍵詞: Re-synthesis, TELPC
相關次數: 點閱:85下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 在這篇論文中,我們希望可以完美合成從複音音樂中擷取的單音,我們使用TE-LPC合成法,它比傳統語音所使用的LPC合成法更適用於音樂合成,因為LPC是針對formant做逼近,但TELPC是針對harmonic partials做逼近,由於音樂訊號中partials的分布情況會影響訊號的音色,故使用True Envelope來模擬,如此一來,可以使得合成音具有原本音樂訊號的音色。此外,在傳統的TELPC合成法使用的是固定的frame size及hop size這會使合成音發生振幅不穩的現象,故此篇論文使用適合該音樂訊號的frame size及hop size,藉此改善振幅不穩的問題。在此架構下,我們可以像創作家玩音樂,改變音樂的內容,甚至是改變原本音樂家的風格。

    In this thesis, I presented a flexible analysis/re-synthesis method for smoothly changing the properties of isolated notes in polyphonic instrumental music recordings. True Envelope Linear Predictive Coding (TELPC) method has been employed as the analysis/synthesis model in order to preserve the original timbre quality as much as possible due to its accurate spectral envelope estimation. I modified the conventional LPC analysis/synthesis processing by using pitch synchronous analysis frames to avoid the severe magnitude modulation problem. Smaller frames can thus be used to capture more local characteristics of the original signals to further improve the sound quality. In this framework, one can manipulate a sequence of isolated notes from commercially available polyphonic instrumental music recordings and interesting re-synthesized results are achieved.

    中文摘要 I Abstract II Content IV List of Figures V 1. Introduction 1 1.1 Motivation and Good 1 1.2 Outline 3 2. Background 4 2.1 LPC 4 2.2 True Envelope Estimation 8 2.3 PSOLA 13 2.4 Pitch Detection 13 3. Proposed Method 15 3.1 Note Extraction 15 3.2 Note re-synthesizer 16 3.2.1 Adaptive Frame Segmentation 17 3.2.2 Analysis Part 18 3.2.3 Transformation 20 3.2.4 Synthesis Part 22 4. Experiment 24 5. Conclusion and Future work 33 5.1 Conclusion 33 5.2 Future Work 33 Reference 35

    [1] M. A. Casey and A. Westner, “Separation of mixed audio sources by independent subspace analysis,” in Proc. ICMC, pp. 154-161, 2000.
    [2] C. Uhle, C. Dittmar, and T. Sporer, “Extraction of drum tracks from polyphonic music using independent subspace analysis,” in Proc. ICA, pp. 843-848, 2003.
    [3] D. D. Lee and H. S. Seung, “Learning the parts of objects by nonnegative matrix factorization,” Nature, vol. 401, pp. 788-791, 1999.
    [4] B. Wang and M. D. Plumbley, “Musical audio stream separation by non-negative matrix factorization,” in Proc. DMRN Summer Conference, Glasgow, 2005.
    [5] S. A. Abdallah and M. D. Plumbley, “Unsupervised analysis of polyphonic music using sparse coding,” IEEE Trans. Neural Networks, vol.17, no. 1, pp. 179-196, 2006.
    [6] T.-M. Wang, T.-C. Chen, Y.-L. Chen, Alvin W.Y. Su, “Time-dependent recursive regularization for sound source separation,” in Proc. of the 3rd International Conference on Audio, Language and Image Processing (ICALIP2012), Shanghai, China, Jul. 16-18, 2012.
    [7] M. Dolson, “The phase vocoder: A tutorial,” Computer Music J., vol. 10, no. 4, pp. 14-27, 1986.
    [8] A. Roebel and X. Rodet, “Efficient spectral envelope estimationand its application to pitch shifting and envelope preservation,” in Proc. DAFx, 2005.
    [9] F. Villavicencio, A. Röbel, and X.Rodet, “Improving LPC spectral envelope extraction of voiced speech by true-envelope estimation,” in Proc. of the ICASSP'06, France, 2006.
    [10] Imai, Satoshi, and Yoshiharu Abe. "Spectral envelope extraction by improved cepstral method." Journal of IEICE 62 : 217-223, 1979.
    [11] V. Villavicencio, A. Röbel, and X. Rodet, "Applying improved spectral modeling for high quality voice conversion," ICASSP2009, pp. 4285-4288, 2009.
    [12] J.J. Burred, A. Roebel, and T. Sikora, “Dynamic spectral envelope modeling for timbre analysis of musical in strumentsounds,” IEEE Trans. Audio, Speech and Lang. Proc., vol. 18, no. 3, March 2010.
    [13] M. Caerano and X. Rodet, “A source-filter model for musical instrument sound transformation,” ICASSP2012, pp. 137-140, 2012.
    [14] J. D. Markel and A. H. Gray Jr., Linear Prediction of Speech, Springer-Verlag, New York, 1976.
    [15] John Coltrane, McCoy, Tynser, Jimmy Garrison and Elvin Jones, Ballads, Compact disc. 1962.
    [16] Kodaly, Ravel, Maggio Ormezowski, Bianchi, Works for Violin and Cello, Compact disc. 1995.
    [17] M. V. Mathews, “Pitch synchronous analysis of voiced sounds,” Journal of the Acoustical Society of America, vol. 33, no. 2, 1961.
    [18] Y. Medan and Eyal Yair, “Pitch synchronous spectral analysis scheme for voiced speech”, IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 37, no. 9, September 1989.
    [19] H. Yang, S.N. Koh, and P. Sivaprakasapillai, “Pitch synchronous multi-band (PSMB) speech coding”. in Proceedings ICASSP-95, 1995.
    [20] H. Ding, I. Y. Soon, and C. K Yeo, “A DCT-based speech enhancement system with pitch synchronous analysis,” IEEE Trans. On Audio, Speech, and Lang. Proc., vol. 19, no. 8, pp. 2614-2623, Nov. 2011.
    [21] Ya-Han Kuo, Re-synthesizing Isolated Notes from Polyphonic Instrumental Music Recordings. URL available at: http://screamlab-ncku-2008.blogspot.tw/2013/04/music-files-of-telpc-based-time.html [accessed 11 July 2013]

    無法下載圖示 校內:2018-08-23公開
    校外:不公開
    電子論文尚未授權公開,紙本請查館藏目錄
    QR CODE