用户可以建立自己的数据网络使用的脚本语言运行时间。Marsyas提供了一个框架,用于建设应用而不是Marsyas文件的一套应用操作或者在个人声音档或收藏简单文本文件包含soundfiles名单。收藏文件应该包含在soundfiles具有相同的采样率为Marsyas没有执行自动取样44100赫兹之间转换(除了22050赫兹)。特征提取过程的结果都存储在Marsyas作为文本文件可以用来在Weka机器学习的环境。并联Marsyas集一些基本的机器学习的组件。
MIRtoolbox也提供了用一个简单和自适应语法来构建一个接一个复杂计算的可能性。与其相反,MIRtoolbox Marsyas不提供实时的能力。另一方面,其object-based架构使得显著简化语法。MIRtoolbox也可以分析音频文件的文件夹,并能处理不同的采样率文件夹,不需任何转换。MIRtoolbox数据数值可以进一步处理,能够在在Matlab环境下直接的帮助,或者其他工具箱可以出口到文本文件。
5. MIRTOOLBOX的有效性
紧跟着第一个Matlab工具箱的是MIDItoolbox,它致力于分析因音乐的符号表征,而MIRtoolbox是免费提供给研究团体的。它可以从下面的网页下载: http://www.cc.jyu.fi/~lartillo/mirtoolbox
6. 感谢
这份工作得到欧盟委员会的支持(雀巢杯“用音乐来调整大脑”代码028570)。工具箱的发展也从与其他合作伙伴生产合作项目,特别是Eerola Tuomas,何塞。Fornari,马可Fabiani,以及我们部门的学生的合作中得得到了益处。
第 26 页 共 28 页
7. 参考文献
[1] G. Tzanetakis and P. Cook, ―Marsyas: A framework for audio analysis,‖ Organized Sound, vol. 4, no. 3, 2000.
[2] M. Slaney, ―Auditory toolbox version 2,‖ Tech. Rep., Interval Research Corporation, 1998-010, 1998.
[3] I. Nabney, Springer Advances In Pattern Recognition Series,chapter NETLAB: Algorithms for pattern recognition, 2002.
[4] J. Vesanto, ―Proceedings of the matlab dsp conference,‖ in Self-Organizing Map in Matlab: the SOM Toolbox, 1999, pp. 35–40.
[5] G. Tzanetakis and P. Cook, ―Multifeature audio segmentation for browsing and annotation,‖ in Proceedings of the 1999 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 1999.
[6] A. Rauber E. Pampalk and D. Merkl, ―Content-based organization and visualization of music archives,‖ in Proceedings of the 10th ACM International Conference on Multimedia, 2002, pp. 570–579.
[7] E. Terhardt, ―On the perception of periodic sound fluctuations (roughness),‖ Acustica, vol. 30, no. 4, pp. 201–213,1974.
[8] P. Boersma, ―Accurate short-term analysis of the fundamental frequency and the harmonics-to-noise ratio of a sampled sound,‖ IFA Proceedings, vol. 17, pp. 97–110, 1993. [9] T. Tolonen and M. Karjalainen, ―A computationally efficient multipitch analysis model,‖ IEEE Transactions on Speech and Audio Processing, vol. 8, no. 6, pp. 708–716, 2000. [10] G. Peeters, ―Music pitch representation by periodicity measures based on combined temporal and spectral representations,‖in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, 2006.
[11] L. Rabiner and B. H. Juangl, Fundamentals of Speech Recognition, Prentice-Hall, 1993. [12] E. Gomez, ―Tonal description of polyphonic audio for music content processing,‖ INFORMS Journal on Computing, vol. 18, no. 3, pp. 294–304, 2006.
[13] C. Krumhansl, Cognitive Foundations of Musical Pitch, Oxford University Press, 1990.
第 27 页 共 28 页
[14] C. Krumhansl and E. J. Kessler, ―Tracing the dynamic changes in perceived tonal organization in a spatial representation of musical keys,‖ Psychological Review, vol. 89,pp. 334–368, 1982.
[15] P. Toiviainen and C. Krumhansl, ―Measuring and modeling real-time responses to music: The dynamics of tonality induction,‖Perception, vol. 32, no. 6, pp. 741–766, 2003.
[16] P. Toiviainen and J.S. Snyder, ―Tapping to bach: Resonancebased modeling of pulse,‖ Music Perception, vol. 21, no. 1,pp. 43–80, 2003.
[17] J. Foote and M. Cooper, ―Media segmentation using selfsimilarity decomposition,‖ in Proceedings of SPIE Storage and Retrieval for Multimedia Databases, 2003, number 5021, pp. 167–175.
[18] T. Eerola and P. Toiviainen, ―MIR in Matlab: The Midi Toolbox,‖in Proceedings of 5th International Conference on Music Information Retrieval, 2004, pp. 22–27.
[19] P. N. Juslin, ―Emotional communication in music performance:A functionalist perspective and some data,‖ Music Perception, vol. 14, pp. 383–418, 1997.
[20] K. R. Scherer and J. S. Oshinsky, ―Cue utilization in emotion attribution from auditory stimuli,‖ Motivation and Emotion,vol. 1
第 28 页 共 28 页