Acoustic Scene Classification Using Spectrograms

2017 36th International Conference of the Chilean Computer Science Society (SCCC)(2017)

引用 7|浏览0
暂无评分
摘要
For human beings, it's an easy task to hear an environment sound and identify from where it belongs. But for machines, it's different. Thinking on that, the Acoustic Scene Classification task was raised with the objective of automatically classify the place/environment where an audio sample was originally recorded. Thereat, this work investigates the performance of an automatic classification system, using the database of the DCASE 2016 challenge, that recognizes environments/places where an audio sample was originally recorded, considering 15 different categories. The sample's audio signal was converted to a spectrogram and after, features were extracted using texture descriptors. The complementary between the existing signals from the left and right channels of an digital audio recorded in stereo was investigated. Combinations using acoustic and visual features were tested as well, looking for an improvement in the classification rate. After all, an accuracy of 80.17% was reached, surpassing the original baseline provided by the DCASE 2016 challenge itself, that presented accuracy of 77.2%. Then, we have concluded that the features obtained in the visual domain can support the development of an efficient classification system in this application.
更多
查看译文
关键词
acoustic scene classification, pattern recognition, signal processing, spectrogram, visual features
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要