Please use this identifier to cite or link to this item:
http://103.99.128.19:8080/xmlui/handle/123456789/368
Title: | A Self-Supervised Convolutional Neural Network Approach for Speech Enhancement |
Authors: | Mamun, Nursadul Majumder, Sharmin Akter, Khadija |
Keywords: | Electrical and Eloctronic Engineering speech enhancement, deep learning, convolutional neural network, self-supervised learning, babble noise, machinery noise |
Issue Date: | 23-Feb-2022 |
Publisher: | Military Institute of Science and Technology (MIST) |
Abstract: | Enhancement of speech means modification to the speech which is degraded by noise. Speech enhancement leads to improvement in the intelligibility of speech to human listeners. Deep learning techniques have drawn tremendous attention for speech enhancement in recent years which require clean speech along with noisy speech for training purpose. However, availability of clean speech signal in naturalistic scenarios is challenging. To ameliorate it, this study proposes a deep neural network- based on speech enhancement approach without the requirement of clean speech to train the model called self-supervised learning. In the proposed framework, two CNN-based speech enhancement models have been deployed for two noisy conditions (babble noise and machinery noise). This work has been accomplished on two different datasets: IEEE speech corpus distorted with real-time noise and recorded speech signals in naturalistic environment. Experimental result demonstrates that the proposed framework achieved significant improvement in both subjective and objective measures. |
Description: | An article published by MIST |
URI: | http://103.99.128.19:8080/xmlui/handle/123456789/368 |
ISBN: | 978-1-6654-9522-6 |
Appears in Collections: | Journals in EEE |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
A_Self-Supervised_Convolutional_Neural_Network_Approach_for_Speech_Enhancement.pdf | 553.03 kB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.