{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,3]],"date-time":"2026-05-03T10:00:51Z","timestamp":1777802451237,"version":"3.51.4"},"reference-count":59,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","award":["2025M781597"],"award-info":[{"award-number":["2025M781597"]}],"id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"publisher","award":["ZR2024QD050"],"award-info":[{"award-number":["ZR2024QD050"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Signal Processing"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.sigpro.2025.110441","type":"journal-article","created":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T02:26:03Z","timestamp":1765592763000},"page":"110441","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Infrared and visible image fusion via spatial-frequency edge-aware network"],"prefix":"10.1016","volume":"242","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-0858-5835","authenticated-orcid":false,"given":"Shuohui","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9675-9016","authenticated-orcid":false,"given":"Qilei","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7273-7499","authenticated-orcid":false,"given":"Mingliang","family":"Gao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9333-5699","authenticated-orcid":false,"given":"Lucia","family":"Cascone","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-2272-0083","authenticated-orcid":false,"given":"Dan","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.sigpro.2025.110441_bib0001","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","article-title":"Pixel-level image fusion: a survey of the state of the art","volume":"33","author":"Li","year":"2017","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0002","doi-asserted-by":"crossref","first-page":"185","DOI":"10.1016\/j.inffus.2022.09.019","article-title":"Current advances and future perspectives of image fusion: a comprehensive review","volume":"90","author":"Karim","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0003","doi-asserted-by":"crossref","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","article-title":"Infrared and visible image fusion methods and applications: a survey","volume":"45","author":"Ma","year":"2019","journal-title":"Inf. fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0004","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2023.109165","article-title":"A comparative review on multi-modal sensors fusion based on deep learning","volume":"213","author":"Tang","year":"2023","journal-title":"Signal Process."},{"key":"10.1016\/j.sigpro.2025.110441_bib0005","article-title":"Cross-modality interaction network for medical image fusion","author":"Song","year":"2024","journal-title":"IEEE Trans. Consum. Electron."},{"issue":"1","key":"10.1016\/j.sigpro.2025.110441_bib0006","doi-asserted-by":"crossref","first-page":"13","DOI":"10.1007\/s13721-024-00449-2","article-title":"CT And MRI image fusion via multimodal feature interaction network","volume":"13","author":"Song","year":"2024","journal-title":"Network Model. Anal. Health Inf. Bioinf."},{"key":"10.1016\/j.sigpro.2025.110441_bib0007","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2024.109749","article-title":"PV-LaP: multi-sensor fusion for 3D scene understanding in intelligent transportation systems","volume":"227","author":"Zhu","year":"2025","journal-title":"Signal Process."},{"key":"10.1016\/j.sigpro.2025.110441_bib0008","doi-asserted-by":"crossref","first-page":"75","DOI":"10.1016\/j.inffus.2016.03.003","article-title":"A review of remote sensing image fusion methods","volume":"32","author":"Ghassemian","year":"2016","journal-title":"Inf. Fusion"},{"issue":"5","key":"10.1016\/j.sigpro.2025.110441_bib0009","first-page":"36","article-title":"Multi-sensor fusion and cooperative perception for autonomous driving: a review","volume":"15","author":"Xiang","year":"2023","journal-title":"IEEE Intell. Transp. Syst. Mag."},{"key":"10.1016\/j.sigpro.2025.110441_bib0010","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"7077","article-title":"Multi-modal fusion transformer for end-to-end autonomous driving","author":"Prakash","year":"2021"},{"key":"10.1016\/j.sigpro.2025.110441_bib0011","doi-asserted-by":"crossref","first-page":"323","DOI":"10.1016\/j.inffus.2021.06.008","article-title":"Image fusion meets deep learning: a survey and perspective","volume":"76","author":"Zhang","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0012","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2025.109955","article-title":"Multi-focus image fusion based on visual depth and fractional-order differentiation operators embedding convolution norm","volume":"233","author":"Xian","year":"2025","journal-title":"Signal Process."},{"key":"10.1016\/j.sigpro.2025.110441_bib0013","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2020.106182","article-title":"Fast infrared and visible image fusion with structural decomposition","volume":"204","author":"Li","year":"2020","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.sigpro.2025.110441_bib0014","doi-asserted-by":"crossref","first-page":"15","DOI":"10.1016\/j.inffus.2015.11.003","article-title":"Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters","volume":"30","author":"Zhou","year":"2016","journal-title":"Inf. fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0015","doi-asserted-by":"crossref","first-page":"109","DOI":"10.1016\/j.inffus.2021.02.008","article-title":"An infrared and visible image fusion method based on multi-scale transformation and norm optimization","volume":"71","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0016","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2021.107087","article-title":"Joint image fusion and denoising via three-layer decomposition and sparse representation","volume":"224","author":"Li","year":"2021","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.sigpro.2025.110441_bib0017","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2021.108254","article-title":"A multi-focus image fusion framework based on multi-scale sparse representation in gradient domain","volume":"189","author":"Wang","year":"2021","journal-title":"Signal Process."},{"issue":"6","key":"10.1016\/j.sigpro.2025.110441_bib0018","doi-asserted-by":"crossref","DOI":"10.14429\/dsj.73.18949","article-title":"Bounded PCA-based multi-sensor image fusion employing curvelet transform coefficients","volume":"73","author":"Singh","year":"2023","journal-title":"Def. Sci. J."},{"key":"10.1016\/j.sigpro.2025.110441_bib0019","doi-asserted-by":"crossref","first-page":"129","DOI":"10.1016\/j.infrared.2018.01.003","article-title":"Infrared and visible image fusion based on robust principal component analysis and compressed sensing","volume":"89","author":"Li","year":"2018","journal-title":"Infrared Phys. Technol."},{"key":"10.1016\/j.sigpro.2025.110441_bib0020","doi-asserted-by":"crossref","first-page":"158","DOI":"10.1016\/j.inffus.2017.10.007","article-title":"Deep learning for pixel-level image fusion: recent advances and future prospects","volume":"42","author":"Liu","year":"2018","journal-title":"Inf. fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0021","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2025.109889","article-title":"LSCANet: Differential features guided long\u2013short cross attention network for infrared and visible image fusion","volume":"231","author":"Guo","year":"2025","journal-title":"Signal Process."},{"key":"10.1016\/j.sigpro.2025.110441_bib0022","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2023.108996","article-title":"MCnet: Multiscale visible image and infrared image fusion network","volume":"208","author":"Sun","year":"2023","journal-title":"Signal Process."},{"key":"10.1016\/j.sigpro.2025.110441_bib0023","doi-asserted-by":"crossref","DOI":"10.1016\/j.sigpro.2025.110073","article-title":"Towards text-refereed multi-modal image fusion by cross-modality interaction","volume":"237","author":"Li","year":"2025","journal-title":"Signal Process."},{"issue":"13","key":"10.1016\/j.sigpro.2025.110441_bib0024","doi-asserted-by":"crossref","first-page":"7296","DOI":"10.1007\/s10489-024-05561-4","article-title":"GAN-GA: Infrared and visible image fusion generative adversarial network based on global awareness","volume":"54","author":"Wu","year":"2024","journal-title":"Appl. Intell."},{"key":"10.1016\/j.sigpro.2025.110441_bib0025","series-title":"2024 4Th International Conference on Computer Science and Blockchain (CCSB)","first-page":"466","article-title":"A lightweight GAN-based image fusion algorithm for visible and infrared images","author":"Wu","year":"2024"},{"issue":"4","key":"10.1016\/j.sigpro.2025.110441_bib0026","doi-asserted-by":"crossref","first-page":"457","DOI":"10.1007\/s00542-022-05315-7","article-title":"IR And visible image fusion using DWT and bilateral filter","volume":"29","author":"Singh","year":"2023","journal-title":"Microsyst. Technol."},{"key":"10.1016\/j.sigpro.2025.110441_bib0027","doi-asserted-by":"crossref","DOI":"10.1016\/j.infrared.2023.104701","article-title":"Infrared and visible image fusion based on domain transform filtering and sparse representation","volume":"131","author":"Li","year":"2023","journal-title":"Infrared Phys. Technol."},{"issue":"6","key":"10.1016\/j.sigpro.2025.110441_bib0028","doi-asserted-by":"crossref","first-page":"1205","DOI":"10.3390\/sym15061205","article-title":"Real-time anomaly detection of network traffic based on CNN","volume":"15","author":"Liu","year":"2023","journal-title":"Symmetry"},{"issue":"5","key":"10.1016\/j.sigpro.2025.110441_bib0029","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","article-title":"DenseFuse: a fusion approach to infrared and visible images","volume":"28","author":"Li","year":"2018","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.sigpro.2025.110441_bib0030","doi-asserted-by":"crossref","DOI":"10.3389\/fnbot.2024.1371385","article-title":"3D Human pose data augmentation using generative adversarial networks for robotic-assisted movement quality assessment","volume":"18","author":"Wang","year":"2024","journal-title":"Front. Neurorobot."},{"issue":"4","key":"10.1016\/j.sigpro.2025.110441_bib0031","doi-asserted-by":"crossref","DOI":"10.1016\/j.ipm.2024.103687","article-title":"FATFusion: A functional\u2013anatomical transformer for medical image fusion","volume":"61","author":"Tang","year":"2024","journal-title":"Inf. Process. Manag."},{"key":"10.1016\/j.sigpro.2025.110441_bib0032","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109295","article-title":"TCCFusion: An infrared and visible image fusion method based on transformer and cross correlation","volume":"137","author":"Tang","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.sigpro.2025.110441_bib0033","doi-asserted-by":"crossref","first-page":"3744","DOI":"10.1109\/TMM.2025.3535390","article-title":"EAT: Multi-Exposure image fusion with adversarial learning and focal transformer","author":"Tang","year":"2025","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.sigpro.2025.110441_bib0034","article-title":"SFCFusion: Spatial-frequency collaborative infrared and visible image fusion","author":"Chen","year":"2024","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.sigpro.2025.110441_bib0035","first-page":"5281","article-title":"A general spatial-frequency learning framework for multimodal image fusion","author":"Zhou","year":"2024","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.sigpro.2025.110441_bib0036","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2020.107404","article-title":"U2-Net: Going deeper with nested U-structure for salient object detection","volume":"106","author":"Qin","year":"2020","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.sigpro.2025.110441_bib0037","doi-asserted-by":"crossref","first-page":"79","DOI":"10.1016\/j.inffus.2022.03.007","article-title":"PIAFusion: A progressive infrared and visible image fusion network based on illumination aware","volume":"83","author":"Tang","year":"2022","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0038","doi-asserted-by":"crossref","first-page":"249","DOI":"10.1016\/j.dib.2017.09.038","article-title":"The TNO multiband image data collection","volume":"15","author":"Toet","year":"2017","journal-title":"Data Brief"},{"key":"10.1016\/j.sigpro.2025.110441_bib0039","series-title":"Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence","article-title":"FusionDN: a unified densely connected network for image fusion","author":"Xu","year":"2020"},{"key":"10.1016\/j.sigpro.2025.110441_bib0040","doi-asserted-by":"crossref","unstructured":"K. Hu, Q. Zhang, M. Yuan, Y. Zhang, SFDFusion: an efficient spatial-frequency domain fusion network for infrared and visible image fusion, (2024). arXiv preprint arXiv:2410.22837.","DOI":"10.3233\/FAIA240524"},{"issue":"4","key":"10.1016\/j.sigpro.2025.110441_bib0041","article-title":"A novel image fusion quality metric based on edge information","volume":"24","author":"Jiang","year":"2015","journal-title":"J. Electron. Imaging"},{"key":"10.1016\/j.sigpro.2025.110441_bib0042","unstructured":"S.D. Metric, Standard Deviation used in Image Quality Evaluation, 2015, Commonly used in image fusion and quality evaluation, no specific paper reference."},{"issue":"3","key":"10.1016\/j.sigpro.2025.110441_bib0043","first-page":"257","article-title":"Image fusion based on mutual information","volume":"38","author":"Goshtasby","year":"2005","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.sigpro.2025.110441_bib0044","first-page":"309","article-title":"A quality evaluation metric for image fusion based on the multi-scale fusion method","author":"Zhang","year":"2009","journal-title":"Proc. 2009 Int. Conf. Image Process."},{"issue":"2","key":"10.1016\/j.sigpro.2025.110441_bib0045","doi-asserted-by":"crossref","first-page":"430","DOI":"10.1109\/TIP.2005.859378","article-title":"Image information and visual quality","volume":"15","author":"Sheikh","year":"2006","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"10.1016\/j.sigpro.2025.110441_bib0046","doi-asserted-by":"crossref","first-page":"156","DOI":"10.1016\/j.inffus.2006.09.001","article-title":"A novel similarity based quality metric for image fusion","volume":"9","author":"Yang","year":"2008","journal-title":"Inf. Fusion"},{"issue":"10","key":"10.1016\/j.sigpro.2025.110441_bib0047","doi-asserted-by":"crossref","first-page":"1421","DOI":"10.1016\/j.imavis.2007.12.002","article-title":"A new automated quality assessment algorithm for image fusion","volume":"27","author":"Chen","year":"2009","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.sigpro.2025.110441_bib0048","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.102147","article-title":"Crossfuse: a novel cross attention mechanism based infrared and visible image fusion approach","volume":"103","author":"Li","year":"2024","journal-title":"Information Fusion"},{"issue":"7","key":"10.1016\/j.sigpro.2025.110441_bib0049","doi-asserted-by":"crossref","first-page":"3159","DOI":"10.1109\/TCSVT.2023.3234340","article-title":"DATFuse: Infrared and visible image fusion via dual attention transformer","volume":"33","author":"Tang","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.sigpro.2025.110441_bib0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109913","article-title":"ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection","volume":"145","author":"Shen","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.sigpro.2025.110441_bib0051","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.101828","article-title":"An interactively reinforced paradigm for joint infrared-visible image fusion and saliency object detection","volume":"98","author":"Wang","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0052","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","article-title":"RFN-Nest: An end-to-end residual fusion network for infrared and visible images","volume":"73","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.sigpro.2025.110441_bib0053","doi-asserted-by":"crossref","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","article-title":"Swinfusion: cross-domain long-range learning for general image fusion via swin transformer","volume":"9","author":"Ma","year":"2022","journal-title":"IEEE\/CAA J. Autom. Sin."},{"issue":"12","key":"10.1016\/j.sigpro.2025.110441_bib0054","doi-asserted-by":"crossref","first-page":"2121","DOI":"10.1109\/JAS.2022.106082","article-title":"Superfusion: a versatile image registration and fusion network with semantic awareness","volume":"9","author":"Tang","year":"2022","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"10.1016\/j.sigpro.2025.110441_bib0055","series-title":"International Joint Conference on Artificial Intelligence (IJCAI)","first-page":"3508","article-title":"Unsupervised misaligned infrared and visible image fusion via cross-Modality image generation and registration","author":"Di","year":"2022"},{"key":"10.1016\/j.sigpro.2025.110441_bib0056","doi-asserted-by":"crossref","first-page":"5413","DOI":"10.1109\/TMM.2022.3192661","article-title":"YDTR: Infrared and visible image fusion via Y-shape dynamic transformer","volume":"25","author":"Tang","year":"2022","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.sigpro.2025.110441_bib0057","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.110822","article-title":"ITFuse: An interactive transformer for infrared and visible image fusion","volume":"156","author":"Tang","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.sigpro.2025.110441_bib0058","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.125472","article-title":"SFINet: A semantic feature interactive learning network for full-time infrared and visible image fusion","volume":"261","author":"Song","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.sigpro.2025.110441_bib0059","first-page":"12077","article-title":"Segformer: simple and efficient design for semantic segmentation with transformers","volume":"34","author":"Xie","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."}],"container-title":["Signal Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0165168425005572?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0165168425005572?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,29]],"date-time":"2026-04-29T20:32:51Z","timestamp":1777494771000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0165168425005572"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":59,"alternative-id":["S0165168425005572"],"URL":"https:\/\/doi.org\/10.1016\/j.sigpro.2025.110441","relation":{},"ISSN":["0165-1684"],"issn-type":[{"value":"0165-1684","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Infrared and visible image fusion via spatial-frequency edge-aware network","name":"articletitle","label":"Article Title"},{"value":"Signal Processing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.sigpro.2025.110441","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"110441"}}