{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T08:03:56Z","timestamp":1777881836862,"version":"3.51.4"},"reference-count":61,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100004543","name":"China Scholarship Council","doi-asserted-by":"publisher","award":["202306250146"],"award-info":[{"award-number":["202306250146"]}],"id":[{"id":"10.13039\/501100004543","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U23B2053"],"award-info":[{"award-number":["U23B2053"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176182"],"award-info":[{"award-number":["62176182"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Expert Systems with Applications"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1016\/j.eswa.2024.125432","type":"journal-article","created":{"date-parts":[[2024,10,4]],"date-time":"2024-10-04T00:36:51Z","timestamp":1728002211000},"page":"125432","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":9,"special_numbering":"C","title":["Dual-stream Noise and Speech Information Perception based Speech Enhancement"],"prefix":"10.1016","volume":"261","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-0397-5569","authenticated-orcid":false,"given":"Nan","family":"Li","sequence":"first","affiliation":[]},{"given":"Longbiao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Qiquan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jianwu","family":"Dang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"3","key":"10.1016\/j.eswa.2024.125432_b1","doi-asserted-by":"crossref","first-page":"247","DOI":"10.1016\/0167-6393(93)90095-3","article-title":"Assessment for automatic speech recognition: Ii. NOISEX-92: A database and an experiment to study the effect of additive noise on speech recognition systems","volume":"12","author":"Andrew","year":"1993","journal-title":"Speech Communication"},{"key":"10.1016\/j.eswa.2024.125432_b2","series-title":"Proc. interspeech 2022","first-page":"936","article-title":"CMGAN: Conformer-based Metric GAN for Speech Enhancement","author":"Cao","year":"2022"},{"key":"10.1016\/j.eswa.2024.125432_b3","unstructured":"Choi, H.-S., Kim, J.-H., Huh, J., Kim, A., Ha, J.-W., & Lee, K. (2018). Phase-aware speech enhancement with deep complex u-net. In International conference on learning representations."},{"key":"10.1016\/j.eswa.2024.125432_b4","doi-asserted-by":"crossref","unstructured":"Defossez, A., Synnaeve, G., & Adi, Y. (2020). Real time speech enhancement in the waveform domain. In Proceedings of interspeech (pp. 3291\u20133295).","DOI":"10.21437\/Interspeech.2020-2409"},{"issue":"2","key":"10.1016\/j.eswa.2024.125432_b5","doi-asserted-by":"crossref","first-page":"443","DOI":"10.1109\/TASSP.1985.1164550","article-title":"Speech enhancement using a minimum mean-square error log-spectral amplitude estimator","volume":"33","author":"Ephraim","year":"1985","journal-title":"IEEE Transactions on Acoustics, Speech, and Signal Processing"},{"key":"10.1016\/j.eswa.2024.125432_b6","doi-asserted-by":"crossref","first-page":"508","DOI":"10.1016\/j.neunet.2023.09.041","article-title":"CompNet: Complementary network for single-channel speech enhancement","volume":"168","author":"Fan","year":"2023","journal-title":"Neural Networks"},{"key":"10.1016\/j.eswa.2024.125432_b7","unstructured":"Fu, S.-W., Liao, C.-F., Tsao, Y., & Lin, S.-D. (2019). MetricGAN: Generative Adversarial Networks based Black-box Metric Scores Optimization for Speech Enhancement. In International conference on machine learning."},{"key":"10.1016\/j.eswa.2024.125432_b8","doi-asserted-by":"crossref","unstructured":"Ge, X., Han, J., Long, Y., & Guan, H. (2022). PercepNet+: A Phase and SNR Aware PercepNet for Real-Time Speech Enhancement. In Proc. interspeech 2022 (pp. 916\u2013920).","DOI":"10.21437\/Interspeech.2022-43"},{"key":"10.1016\/j.eswa.2024.125432_b9","series-title":"Philadelphia : Linguistic data consortium","author":"Graff","year":"2003"},{"key":"10.1016\/j.eswa.2024.125432_b10","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2015). Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In 2015 IEEE international conference on computer vision (pp. 1026\u20131034).","DOI":"10.1109\/ICCV.2015.123"},{"key":"10.1016\/j.eswa.2024.125432_b11","series-title":"2010 IEEE international conference on acoustics, speech and signal processing","first-page":"4266","article-title":"MMSE based noise PSD tracking with low complexity","author":"Hendriks","year":"2010"},{"key":"10.1016\/j.eswa.2024.125432_b12","series-title":"Proc. interspeech 2020","first-page":"2472","article-title":"DCCRN: Deep Complex Convolution Recurrent Network for Phase-Aware Speech Enhancement","author":"Hu","year":"2020"},{"issue":"1","key":"10.1016\/j.eswa.2024.125432_b13","doi-asserted-by":"crossref","first-page":"229","DOI":"10.1109\/TASL.2007.911054","article-title":"Evaluation of objective quality measures for speech enhancement","volume":"16","author":"Hu","year":"2008","journal-title":"IEEE Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b14","doi-asserted-by":"crossref","unstructured":"Hu, Y., et al. (2020). DCCRN: Deep complex convolution recurrent network for phase-aware speech enhancement. In Proceedings of interspeech (pp. 2472\u20132476).","DOI":"10.21437\/Interspeech.2020-2537"},{"key":"10.1016\/j.eswa.2024.125432_b15","series-title":"Blind source separation","first-page":"349","article-title":"On the ideal ratio mask as the goal of computational auditory scene analysis","author":"Hummersone","year":"2014"},{"key":"10.1016\/j.eswa.2024.125432_b16","series-title":"2022 44th annual international conference of the IEEE engineering in medicine & biology society","first-page":"2581","article-title":"A novel speech intelligibility enhancement model based on canonical correlation and deep learning","author":"Hussain","year":"2022"},{"key":"10.1016\/j.eswa.2024.125432_b17","series-title":"Scaling speech enhancement in unseen environments with noise embeddings","author":"Keren","year":"2018"},{"key":"10.1016\/j.eswa.2024.125432_b18","series-title":"Multi-domain processing via hybrid denoising networks for speech enhancement","author":"Kim","year":"2018"},{"issue":"1","key":"10.1016\/j.eswa.2024.125432_b19","doi-asserted-by":"crossref","first-page":"1360","DOI":"10.1515\/jisys-2017-0509","article-title":"Noise reduction using modified wiener filter in digital hearing aid for speech signal enhancement","volume":"29","author":"Kumar","year":"2020","journal-title":"Journal of Intelligent Systems"},{"key":"10.1016\/j.eswa.2024.125432_b20","series-title":"2020 Asia-Pacific signal and information processing association annual summit and conference","first-page":"739","article-title":"Dynamic noise embedding: Noise aware training and adaptation for speech enhancement","author":"Lee","year":"2020"},{"issue":"9","key":"10.1016\/j.eswa.2024.125432_b21","doi-asserted-by":"crossref","first-page":"3230","DOI":"10.3390\/app10093230","article-title":"Multi-task learning u-net for single-channel speech enhancement and mask-based voice activity detection","volume":"10","author":"Lee","year":"2020","journal-title":"Applied Sciences"},{"key":"10.1016\/j.eswa.2024.125432_b22","series-title":"2022 international joint conference on neural networks","first-page":"1","article-title":"Dual-stream speech dereverberation network using long-term and short-term cues","author":"Li","year":"2022"},{"key":"10.1016\/j.eswa.2024.125432_b23","doi-asserted-by":"crossref","first-page":"1829","DOI":"10.1109\/TASLP.2021.3079813","article-title":"Two heads are better than one: A two-stage complex spectral mapping approach for monaural speech enhancement","volume":"29","author":"Li","year":"2021","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b24","doi-asserted-by":"crossref","DOI":"10.1016\/j.specom.2023.103024","article-title":"Robust voice activity detection using an auditory-inspired masked modulation encoder based convolutional attention network","volume":"157","author":"Li","year":"2024","journal-title":"Speech Communication"},{"key":"10.1016\/j.eswa.2024.125432_b25","doi-asserted-by":"crossref","unstructured":"Li, N., Wang, L., Unoki, M., Li, S., Wang, R., Ge, M., et al. (2021). Robust Voice Activity Detection Using a Masked Auditory Encoder Based Convolutional Neural Network. In ICASSP 2021 - 2021 IEEE international conference on acoustics, speech and signal processing (pp. 6828\u20136832).","DOI":"10.1109\/ICASSP39728.2021.9415045"},{"key":"10.1016\/j.eswa.2024.125432_b26","series-title":"Noise tokens: Learning neural noise templates for environment-aware speech enhancement","author":"Li","year":"2020"},{"key":"10.1016\/j.eswa.2024.125432_b27","doi-asserted-by":"crossref","first-page":"3629","DOI":"10.1109\/TASLP.2023.3313442","article-title":"A general unfolding speech enhancement method motivated by taylor\u2019s theorem","volume":"31","author":"Li","year":"2023","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b28","doi-asserted-by":"crossref","first-page":"2156","DOI":"10.1109\/TASLP.2022.3184889","article-title":"Filtering and refining: A collaborative-style framework for single-channel speech enhancement","volume":"30","author":"Li","year":"2022","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"issue":"3","key":"10.1016\/j.eswa.2024.125432_b29","doi-asserted-by":"crossref","first-page":"197","DOI":"10.1109\/TASSP.1978.1163086","article-title":"All-pole modeling of degraded speech","volume":"26","author":"Lim","year":"1978","journal-title":"IEEE Transactions on Acoustics, Speech and Signal Processing"},{"key":"10.1016\/j.eswa.2024.125432_b30","series-title":"Speech enhancement: Theory and practice","author":"Loizou","year":"2007"},{"key":"10.1016\/j.eswa.2024.125432_b31","doi-asserted-by":"crossref","unstructured":"Lu, Y.-X., Ai, Y., & Ling, Z.-H. (2023). MP-SENet: A speech enhancement model with parallel denoising of magnitude and phase spectra. In Proc. iNTERSPEECH 2023 (pp. 3834\u20133838).","DOI":"10.21437\/Interspeech.2023-1441"},{"issue":"8","key":"10.1016\/j.eswa.2024.125432_b32","doi-asserted-by":"crossref","first-page":"1256","DOI":"10.1109\/TASLP.2019.2915167","article-title":"Conv-TasNet: Surpassing ideal time\u2013frequency magnitude masking for speech separation","volume":"27","author":"Luo","year":"2019","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b33","doi-asserted-by":"crossref","first-page":"71","DOI":"10.1016\/j.specom.2022.10.003","article-title":"Analysis of trade-offs between magnitude and phase estimation in loss functions for speech denoising and dereverberation","volume":"145","author":"Luo","year":"2022","journal-title":"Speech Communication"},{"issue":"8","key":"10.1016\/j.eswa.2024.125432_b34","doi-asserted-by":"crossref","first-page":"1256","DOI":"10.1109\/TASLP.2019.2915167","article-title":"Conv-TasNet: Surpassing ideal time\u2013frequency magnitude masking for speech separation","volume":"27","author":"Luo","year":"2019","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b35","doi-asserted-by":"crossref","unstructured":"Pandey, A., & Wang, D. (2020). Densely connected neural network with dilated convolutions for real-time speech enhancement in the time domain. In Proceedings of IEEE international conference on acoustics, speech, and signal processing (pp. 6629\u20136633).","DOI":"10.1109\/ICASSP40776.2020.9054536"},{"key":"10.1016\/j.eswa.2024.125432_b36","doi-asserted-by":"crossref","unstructured":"Pascual, S., Bonafonte, A., & Serr\u00e0, J. (2017). SEGAN: Speech Enhancement Generative Adversarial Network. In Proc. Interspeech 2017 (pp. 3642\u20133646).","DOI":"10.21437\/Interspeech.2017-1428"},{"issue":"10","key":"10.1016\/j.eswa.2024.125432_b37","doi-asserted-by":"crossref","DOI":"10.1121\/10.0014875","article-title":"Phase-aware deep speech enhancement: It\u2019s all about the frame length","volume":"2","author":"Peer","year":"2022","journal-title":"JASA Express Letters"},{"key":"10.1016\/j.eswa.2024.125432_b38","series-title":"2018 IEEE international conference on acoustics, speech and signal processing","first-page":"5069","article-title":"A wavenet for speech denoising","author":"Rethage","year":"2018"},{"key":"10.1016\/j.eswa.2024.125432_b39","first-page":"749","article-title":"Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs","volume":"vol. 2","author":"Rix","year":"2001"},{"key":"10.1016\/j.eswa.2024.125432_b40","doi-asserted-by":"crossref","unstructured":"Rouard, S., Massa, F., & D\u00e9fossez, A. (2023). Hybrid Transformers for Music Source Separation. In ICASSP 23.","DOI":"10.1109\/ICASSP49357.2023.10096956"},{"key":"10.1016\/j.eswa.2024.125432_b41","doi-asserted-by":"crossref","unstructured":"Schroter, H., Escalante-B, A. N., Rosenkranz, T., & Maier, A. (2022). DeepFilterNet: A low complexity speech enhancement framework for full-band audio based on deep filtering. In Proceedings of IEEE international conference on acoustics, speech, signal processing (pp. 7407\u20137411).","DOI":"10.1109\/ICASSP43922.2022.9747055"},{"key":"10.1016\/j.eswa.2024.125432_b42","series-title":"2018 IEEE international conference on acoustics, speech and signal processing","first-page":"5039","article-title":"Time-frequency masking-based speech enhancement using generative adversarial network","author":"Soni","year":"2018"},{"key":"10.1016\/j.eswa.2024.125432_b43","doi-asserted-by":"crossref","unstructured":"Tan, K., Chen, J., & Wang, D. (2018). Gated residual networks with dilated convolutions for supervised speech separation. In Proceedings of IEEE international conference on acoustics, speech, and signal processing (pp. 21\u201325).","DOI":"10.1109\/ICASSP.2018.8461819"},{"key":"10.1016\/j.eswa.2024.125432_b44","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.csl.2019.06.005","article-title":"rVAD: An unsupervised segment-based robust voice activity detection method","volume":"59","author":"Tan","year":"2020","journal-title":"Computer Speech & Language"},{"key":"10.1016\/j.eswa.2024.125432_b45","series-title":"ICASSP 2019-2019 IEEE international conference on acoustics, speech and signal processing","first-page":"6865","article-title":"Complex spectral mapping with a convolutional recurrent network for monaural speech enhancement","author":"Tan","year":"2019"},{"key":"10.1016\/j.eswa.2024.125432_b46","doi-asserted-by":"crossref","first-page":"380","DOI":"10.1109\/TASLP.2019.2955276","article-title":"Learning complex spectral mapping with gated convolutional recurrent networks for monaural speech enhancement","volume":"28","author":"Tan","year":"2020","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b47","article-title":"The diverse environments multi-channel acoustic noise database (demand): A database of multichannel environmental noise recordings","volume":"vol. 19","author":"Thiemann","year":"2013"},{"key":"10.1016\/j.eswa.2024.125432_b48","doi-asserted-by":"crossref","unstructured":"Valentini-Botinhao, C., Wang, X., Takaki, S., & Yamagishi, J. (2016). Investigating RNN-based speech enhancement methods for noise-robust Text-to-Speech. In Speech synthesis workshop (pp. 146\u2013152).","DOI":"10.21437\/SSW.2016-24"},{"key":"10.1016\/j.eswa.2024.125432_b49","doi-asserted-by":"crossref","unstructured":"Valin, J.-M. (2017). A Hybrid DSP\/Deep Learning Approach to Real-Time Full-Band Speech Enhancement. In 2018 IEEE 20th International Workshop on Multimedia Signal Processing (pp. 1\u20135).","DOI":"10.1109\/MMSP.2018.8547084"},{"key":"10.1016\/j.eswa.2024.125432_b50","doi-asserted-by":"crossref","unstructured":"Valin, J.-M., Isik, U., Phansalkar, N., Giri, R., Helwani, K., & Krishnaswamy, A. (2020). A Perceptually-Motivated Approach for Low-Complexity, Real-Time Enhancement of Fullband Speech. In Interspeech.","DOI":"10.21437\/Interspeech.2020-2730"},{"key":"10.1016\/j.eswa.2024.125432_b51","series-title":"2013 international conference oriental COCOSDA held jointly with 2013 conference on Asian spoken language research and evaluation","first-page":"1","article-title":"The voice bank corpus: Design, collection and data analysis of a large regional accent speech database","author":"Veaux","year":"2013"},{"issue":"10","key":"10.1016\/j.eswa.2024.125432_b52","doi-asserted-by":"crossref","first-page":"1702","DOI":"10.1109\/TASLP.2018.2842159","article-title":"Supervised speech separation based on deep learning: An overview","volume":"26","author":"Wang","year":"2018","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"issue":"12","key":"10.1016\/j.eswa.2024.125432_b53","doi-asserted-by":"crossref","first-page":"1849","DOI":"10.1109\/TASLP.2014.2352935","article-title":"On training targets for supervised speech separation","volume":"22","author":"Wang","year":"2014","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b54","doi-asserted-by":"crossref","first-page":"1778","DOI":"10.1109\/TASLP.2020.2998279","article-title":"Complex spectral mapping for single-and multi-channel speech enhancement and robust ASR","volume":"28","author":"Wang","year":"2020","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b55","doi-asserted-by":"crossref","first-page":"2018","DOI":"10.1109\/LSP.2021.3116502","article-title":"On the compensation between magnitude and phase in speech separation","volume":"28","author":"Wang","year":"2021","journal-title":"IEEE Signal Processing Letters"},{"key":"10.1016\/j.eswa.2024.125432_b56","series-title":"2023 Asia Pacific signal and information processing association annual summit and conference","first-page":"2416","article-title":"Nadiffuse: Noise-aware diffusion-based model for speech enhancement","author":"Wang","year":"2023"},{"issue":"3","key":"10.1016\/j.eswa.2024.125432_b57","doi-asserted-by":"crossref","first-page":"483","DOI":"10.1109\/TASLP.2015.2512042","article-title":"Complex ratio masking for monaural speech separation","volume":"24","author":"Williamson","year":"2016","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b58","unstructured":"Xu, Y., Du, J., Dai, L.-R., & Lee, C.-H. A regression approach to speech enhancement based on deep neural networks. IEEE\/ACM Transactions on Audio, Speech."},{"key":"10.1016\/j.eswa.2024.125432_b59","doi-asserted-by":"crossref","first-page":"1404","DOI":"10.1109\/TASLP.2020.2987441","article-title":"Deepmmse: A deep learning approach to MMSE-based noise power spectral density estimation","volume":"28","author":"Zhang","year":"2020","journal-title":"IEEE\/ACM Transactions on Audio, Speech, and Language Processing"},{"key":"10.1016\/j.eswa.2024.125432_b60","first-page":"14549","article-title":"Interactive speech and noise modeling for speech enhancement","volume":"vol. 35","author":"Zheng","year":"2021"},{"key":"10.1016\/j.eswa.2024.125432_b61","doi-asserted-by":"crossref","DOI":"10.1016\/j.dsp.2023.104151","article-title":"A robust and lightweight voice activity detection algorithm for speech enhancement at low signal-to-noise ratio","volume":"141","author":"Zhu","year":"2023","journal-title":"Digital Signal Processing"}],"container-title":["Expert Systems with Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0957417424022991?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0957417424022991?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:54:42Z","timestamp":1777596882000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0957417424022991"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2]]},"references-count":61,"alternative-id":["S0957417424022991"],"URL":"https:\/\/doi.org\/10.1016\/j.eswa.2024.125432","relation":{},"ISSN":["0957-4174"],"issn-type":[{"value":"0957-4174","type":"print"}],"subject":[],"published":{"date-parts":[[2025,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Dual-stream Noise and Speech Information Perception based Speech Enhancement","name":"articletitle","label":"Article Title"},{"value":"Expert Systems with Applications","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.eswa.2024.125432","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"125432"}}