Skip to main navigation Skip to search Skip to main content

Deep belief networks for automatic music genre classification

  • Xiaohong Yang*
  • , Qingcai Chen
  • , Shusen Zhou
  • , Xiaolong Wang
  • *Corresponding author for this work
  • Harbin Institute of Technology Shenzhen

Research output: Contribution to journalConference articlepeer-review

Abstract

This paper proposes an approach to automatic music genre clas- sification using deep belief networks. Based on the restricted Boltzmann machines, the deep belief networks is constructed and takes the acoustic features extracted through content-based analysis of music signals as input. The model parameters are initially determined after the deep belief network is trained by greedy layer-wise learning algorithm with feature vectors that are comprised of short-term and long-term features. Then the parameters are fine-tuned to local optimum according to back propagation algorithm. Experiments on GTZAN dataset show that the performance of music genre classification using deep belief networks is superior to those of widely used classification methods such as support vector machine, K-nearest neighbor, linear discriminant analysis and neural network.

Original languageEnglish
Pages (from-to)2433-2436
Number of pages4
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
StatePublished - 2011
Externally publishedYes
Event12th Annual Conference of the International Speech Communication Association, INTERSPEECH 2011 - Florence, Italy
Duration: 27 Aug 201131 Aug 2011

Keywords

  • Deep belief networks
  • Modulation spectral analysis
  • Music genre classification
  • Timbral texture feature

Fingerprint

Dive into the research topics of 'Deep belief networks for automatic music genre classification'. Together they form a unique fingerprint.

Cite this