Raw Waveform-based Audio Classification Using Sample-level CNN Architectures

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 693
  • Download : 0
Music, speech, and acoustic scene sound are often handled separately in the audio domain because of their different signal characteristics. However, as the image domain grows rapidly by versatile image classification models, it is necessary to study extensible classification models in the audio domain as well. In this study, we approach this problem using two types of sample-level deep convolutional neural networks that take raw waveforms as input and uses filters with small granularity. One is a basic model that consists of convolution and pooling layers. The other is an improved model that additionally has residual connections, squeeze-and-excitation modules and multi-level concatenation. We show that the sample-level models reach state-of-the-art performance levels for the three different categories of sound. Also, we visualize the filters along layers and compare the characteristics of learned filters.
Publisher
Neural Information Processing Systems (NIPS)
Issue Date
2017-12-08
Language
English
Citation

Machine Learning for Audio Signal Processing Workshop, Neural Information Processing Systems (NIPS)

URI
http://hdl.handle.net/10203/238218
Appears in Collection
GCT-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0