Deep learning-based method for multiple sound source localization with high resolution and accuracy

Cited 34 time in webofscience Cited 0 time in scopus
  • Hit : 66
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorLee, Soo Youngko
dc.contributor.authorChang, Jihoko
dc.contributor.authorLee, Seungchulko
dc.date.accessioned2023-09-13T03:00:45Z-
dc.date.available2023-09-13T03:00:45Z-
dc.date.created2023-09-13-
dc.date.created2023-09-13-
dc.date.issued2021-12-
dc.identifier.citationMECHANICAL SYSTEMS AND SIGNAL PROCESSING, v.161-
dc.identifier.issn0888-3270-
dc.identifier.urihttp://hdl.handle.net/10203/312544-
dc.description.abstractDeep learning-based methods are attracting interest in sound source localization, showing promising results compared to conventional model-based approaches. While these deep learning-based methods have been mainly developed into two approaches, i.e., grid-based and grid-free methods, they inherently involve several limitations that the sound sources should be assumed on the grid points or the number of sound sources should be predefined when constructing a deep neural network's architecture. Breaking away from the existing methods' limitations, we propose a deep learning approach to fulfill multiple sound source localization with high resolution and accuracy, for whether the sound sources are located on the grid points or not. We first suggest a target function to obtain spatial source distribution maps, that can represent multiple sources' positional and strength information, even when the sources are placed off the grid points. While the multiple sound source localization is expanded by the proposed source map into image-to-image pixel-level prediction task, we then propose a fully convolutional neural network (FCN) with an encoder-decoder structure to estimate the multiple sources' positions and strength precisely. Based on the dataset acquired by one to three monopole sources on a square plane of 2.68 x 2.68 m, with a spiral array of 60 microphones at 1, 2, and 10 kHz, we assess both quantitative and qualitative results of the proposed model and demonstrate that our proposed model can achieve highly precise localization results regardless of frequency and the number of sound sources. Besides, we validate that high-resolution source distribution maps can be obtained by the proposed model, from which the positions and the strengths of sound sources are accurately predicted. Lastly, we compare the proposed model with several deconvolution methods, and the results show that the proposed deep learning model significantly outperforms the model-based methods.-
dc.languageEnglish-
dc.publisherACADEMIC PRESS LTD- ELSEVIER SCIENCE LTD-
dc.titleDeep learning-based method for multiple sound source localization with high resolution and accuracy-
dc.typeArticle-
dc.identifier.wosid000670074400008-
dc.identifier.scopusid2-s2.0-85105691197-
dc.type.rimsART-
dc.citation.volume161-
dc.citation.publicationnameMECHANICAL SYSTEMS AND SIGNAL PROCESSING-
dc.identifier.doi10.1016/j.ymssp.2021.107959-
dc.contributor.localauthorLee, Seungchul-
dc.contributor.nonIdAuthorLee, Soo Young-
dc.contributor.nonIdAuthorChang, Jiho-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorSound source localization-
dc.subject.keywordAuthorHigh resolution source map-
dc.subject.keywordAuthorDeep learning-
dc.subject.keywordAuthorConvolutional neural network-
dc.subject.keywordAuthorFully convolutional neural network-
Appears in Collection
ME-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 34 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0