透過您的圖書館登入
IP:18.227.228.95
  • 學位論文

利用結構性支撐向量機的具音樂表現能力之半自動電腦演奏系統

A Semi-automatic Computer Expressive Music Performance System Using Structural Support Vector Machine

指導教授 : 鄭士康
若您是本文的作者,可授權文章由華藝線上圖書館中協助推廣。

摘要


電腦合成的音樂一向被認為是僵硬、機械化而且沒有音樂表現能力。因此能夠產生具有表現能力的電腦自動演奏系統將會對音樂產業、個人化娛樂以及表驗藝術領域有重大的影響。在這篇論文中,我們藉由隱藏式馬可夫模型結構的結構性支撐向量機 (SVM-HMM) 來設計一個可以產生具有表現能力音樂的電腦自動演奏系統。我們邀請六位研究生錄製了克萊門蒂(Muzio Clementi)的小奏鳴曲集 Op.36。我們手動將這些錄音分割成樂句,並且利用程式從中抽取出音樂特徵。這些 音樂特徵藉由 SVM-HMM 訓練成數學模型後,可以利用這個數學模型來演奏訓練過程中沒有見過的樂譜(需要手動標注樂句)。此系統目前只能支援單音旋律。問卷調查的結果顯示,本系統產生的音樂尚不能達到真人的演奏水準。但是根據量化的相似度分析,本系統產生的音樂確實比無表現性的 MIDI 音樂更接近真人演奏。

並列摘要


Computer generated music is known to be robotic and inexpressive. A computer system that can generate expressive performance potentially has significant impact on music production industry, personalized entertainment or even art. In this paper, we have designed and implemented a system that can generate expressive performance using structural support vector machine with hidden Markov model output (SVM-HMM). We recorded six sets of Muzio Clementi's Sonatina Op.36 performed by six graduate students. The recordings and scores are manually split into phrases and had their musical features automatically extracted. Using the SVM-HMM algorithm, a mathematical model of expressive performance knowledge is learned from these features. The trained model can generate expressive performances for previously unseen scores (with user-assigned phrasings). The system currently supports monophonic music only. Subjective test shows that the computer generated performances still cannot achieve the same level of expressiveness of human performers, but quantitative similarity measures show that the computer generated performances are much similar to human performances than inexpressive MIDIs.

參考文獻


[7] A. Friberg, R. Bresin, and J. Sundberg, “Overview of the KTH rule system for musical performance,” Advances in Cognitive Psychology, vol. 2, pp. 145--161, Jan. 2006.
[9] S. R. Livingstone, R. Mühlberger, A. R. Brown, and A. Loch, “Controlling musical
emotionality: an affective computational architecture for influencing musical emotions,” Digital Creativity, vol. 18, pp. 43--53, Mar. 2007.
[10] N. P. M. Todd, “A computational model of rubato,” Contemporary Music Review, vol. 3, pp. 69--88, Jan. 1989.
[11] N. P. McAngus Todd, “The dynamics of dynamics: A model of musical expression,”The Journal of the Acoustical Society of America, vol. 91, p. 3540, June 1992.

延伸閱讀