ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper A Dual-Staged Context Aggregation Method towards Efficient End-to-End Speech Enhancement
Cited 4 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Kai Zhen, Mi Suk Lee, Minje Kim
Issue Date
2020-05
Citation
International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020, pp.366-370
Publisher
IEEE
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICASSP40776.2020.9054499
Abstract
In speech enhancement, an end-to-end deep neural network converts a noisy speech signal to a clean speech directly in the time domain without time-frequency transformation or mask estimation. However, aggregating contextual information from a high-resolution time domain signal with an affordable model complexity still remains challenging. In this paper, we propose a densely connected convolutional and recurrent network (DCCRN), a hybrid architecture, to enable dual-staged temporal context aggregation. With the dense connectivity and cross-component identical shortcut, DCCRN consistently outperforms competing convolutional baselines with an average STOI improvement of 0.23 and PESQ of 1.38 at three SNR levels. The proposed method is computationally efficient with only 1.38 million parameters. The generalizability performance on the unseen noise types is still decent considering its low complexity, although it is relatively weaker comparing to Wave-U-Net with 7.25 times more parameters.
KSP Keywords
Clean speech, Computationally Efficient, Contextual information, Deep neural network(DNN), End to End(E2E), High-resolution, Hybrid architecture, Noisy speech signal, Recurrent network, Resolution time, Time-Frequency Transformation