ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Convolutional Recurrent Neural Networks for Urban Sound Classification using Raw Waveforms
Cited 56 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Jonghee Sang, Soomyung Park, Junwoo Lee
Issue Date
2018-09
Citation
European Signal Processing Conference (EUSIPCO) 2018, pp.2444-2448
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.23919/EUSIPCO.2018.8553247
Abstract
Recent studies have demonstrated deep learning approaches directly from raw data have been successfully used in image and text. This approach has been applied to audio signals as well but not fully explored yet. In this works, we propose a convolutional recurrent neural network that directly uses time-domain waveforms as input in the domain of urban sound classification. Convolutional recurrent neural network is combined model of convolutional neural networks for extracting sound features and recurrent neural networks for temporal aggregation of the extracted features. The method was evaluated using the UrbanSound8k dataset, the largest public dataset of urban environmental sound sources available for research. The results show how convolutional recurrent neural network with raw waveforms improve the accuracy in urban sound classification and provide effectiveness of its structure with respect to the number of parameters.
KSP Keywords
Audio signal, Convolution neural network(CNN), Environmental sound, Learning approach, Public Datasets, Recurrent Neural Network(RNN), Sound feature, Sound source, Temporal aggregation, Urban sound classification, combined model