{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T23:25:53Z","timestamp":1771025153732,"version":"3.50.1"},"reference-count":50,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,8,1]],"date-time":"2022-08-01T00:00:00Z","timestamp":1659312000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2021ZD0113002"],"award-info":[{"award-number":["2021ZD0113002"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61572005"],"award-info":[{"award-number":["61572005"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61771058"],"award-info":[{"award-number":["61771058"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62072292"],"award-info":[{"award-number":["62072292"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Signal Processing: Image Communication"],"published-print":{"date-parts":[[2022,8]]},"DOI":"10.1016\/j.image.2022.116722","type":"journal-article","created":{"date-parts":[[2022,5,7]],"date-time":"2022-05-07T11:16:50Z","timestamp":1651922210000},"page":"116722","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":14,"special_numbering":"C","title":["StyleFuse: An unsupervised network based on style loss function for infrared and visible image fusion"],"prefix":"10.1016","volume":"106","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0075-7682","authenticated-orcid":false,"given":"Chen","family":"Cheng","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3807-3569","authenticated-orcid":false,"given":"Cheng","family":"Sun","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2155-0293","authenticated-orcid":false,"given":"Yongqi","family":"Sun","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5757-3753","authenticated-orcid":false,"given":"Jiahui","family":"Zhu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.image.2022.116722_b1","doi-asserted-by":"crossref","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","article-title":"Infrared and visible image fusion methods and applications: A survey","volume":"45","author":"Ma","year":"2019","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.image.2022.116722_b2","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.05.004","article-title":"Pixel-level image fusion: A survey of the state of the art","volume":"33","author":"Li","year":"2017","journal-title":"Inf. Fusion"},{"issue":"3","key":"10.1016\/j.image.2022.116722_b3","doi-asserted-by":"crossref","first-page":"235","DOI":"10.1006\/gmip.1995.1022","article-title":"Multisensor image fusion using the wavelet transform","volume":"57","author":"Li","year":"1995","journal-title":"Graph. Models Image Process."},{"issue":"7","key":"10.1016\/j.image.2022.116722_b4","doi-asserted-by":"crossref","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","article-title":"Image fusion with guided filtering","volume":"22","author":"Li","year":"2013","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.image.2022.116722_b5","series-title":"Infrared and visible image fusion using latent low-rank representation","author":"Li","year":"2019"},{"key":"10.1016\/j.image.2022.116722_b6","article-title":"MDLatLRR: A novel decomposition method for infrared and visible image fusion","author":"Li","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.image.2022.116722_b7","doi-asserted-by":"crossref","first-page":"147","DOI":"10.1016\/j.inffus.2014.09.004","article-title":"A general framework for image fusion based on multi-scale transform and sparse representation","volume":"24","author":"Liu","year":"2015","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.image.2022.116722_b8","doi-asserted-by":"crossref","first-page":"15","DOI":"10.1016\/j.inffus.2015.11.003","article-title":"Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters","volume":"30","author":"Zhou","year":"2016","journal-title":"Inf. Fusion"},{"issue":"5","key":"10.1016\/j.image.2022.116722_b9","doi-asserted-by":"crossref","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","article-title":"DenseFuse: A fusion approach to infrared and visible images","volume":"28","author":"Li","year":"2019","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"10.1016\/j.image.2022.116722_b10","doi-asserted-by":"crossref","first-page":"9645","DOI":"10.1109\/TIM.2020.3005230","article-title":"NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial\/channel attention models","volume":"69","author":"Li","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.image.2022.116722_b11","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1016\/j.inffus.2021.02.023","article-title":"RFN-Nest: An end-to-end residual fusion network for infrared and visible images","volume":"73","author":"Li","year":"2021","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.image.2022.116722_b12","first-page":"1","article-title":"U2Fusion: A unified unsupervised image fusion network","author":"Xu","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.image.2022.116722_b13","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","article-title":"FusionGAN: A generative adversarial network for infrared and visible image fusion","volume":"48","author":"Ma","year":"2019","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.image.2022.116722_b14","doi-asserted-by":"crossref","first-page":"3845","DOI":"10.1109\/TIP.2020.2966075","article-title":"Unsupervised deep image fusion with structure tensor representations","volume":"29","author":"Jung","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.image.2022.116722_b15","article-title":"GANMcC: A generative adversarial network with multi-classification constraints for infrared and visible image fusion","volume":"70","author":"Ma","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.image.2022.116722_b16","first-page":"1","article-title":"Multigrained attention network for infrared and visible image fusion","volume":"70","author":"Li","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.image.2022.116722_b17","doi-asserted-by":"crossref","first-page":"128","DOI":"10.1016\/j.inffus.2020.11.009","article-title":"RXDNFuse: A aggregated residual dense network for infrared and visible image fusion","volume":"69","author":"Long","year":"2021","journal-title":"Inf. Fusion"},{"issue":"7","key":"10.1016\/j.image.2022.116722_b18","doi-asserted-by":"crossref","first-page":"3330","DOI":"10.1109\/TCYB.2019.2894498","article-title":"Enhancing sketch-based image retrieval by CNN semantic re-ranking","volume":"50","author":"Wang","year":"2020","journal-title":"IEEE Trans. Cybern."},{"key":"10.1016\/j.image.2022.116722_b19","series-title":"2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"2414","article-title":"Image style transfer using convolutional neural networks","author":"Gatys","year":"2016"},{"key":"10.1016\/j.image.2022.116722_b20","series-title":"ECCV (2)","first-page":"694","article-title":"Perceptual losses for real-time style transfer and super-resolution","volume":"vol. 9906","author":"Johnson","year":"2016"},{"key":"10.1016\/j.image.2022.116722_b21","doi-asserted-by":"crossref","DOI":"10.1016\/j.image.2021.116452","article-title":"Unsupervised multi-domain image translation with domain representation learning","volume":"99","author":"Liu","year":"2021","journal-title":"Signal Process., Image Commun."},{"issue":"5","key":"10.1016\/j.image.2022.116722_b22","doi-asserted-by":"crossref","first-page":"1193","DOI":"10.1007\/s11760-013-0556-9","article-title":"Image fusion based on pixel significance using cross bilateral filter","volume":"9","author":"Shreyamsha\u00a0Kumar","year":"2015","journal-title":"Signal Image Video Process."},{"key":"10.1016\/j.image.2022.116722_b23","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1016\/j.infrared.2016.01.009","article-title":"Two-scale image fusion of visible and infrared images using saliency detection","volume":"76","author":"Bavirisetti","year":"2016","journal-title":"Infrared Phys. Technol."},{"issue":"1","key":"10.1016\/j.image.2022.116722_b24","doi-asserted-by":"crossref","first-page":"48","DOI":"10.1007\/s12596-013-0148-7","article-title":"Hybrid DDCT-PCA based multi sensor image fusion","volume":"43","author":"Naidu","year":"2014","journal-title":"J. Opt."},{"key":"10.1016\/j.image.2022.116722_b25","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","article-title":"IFCNN: A general image fusion framework based on convolutional neural network","volume":"54","author":"Zhang","year":"2020","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.image.2022.116722_b26","doi-asserted-by":"crossref","DOI":"10.1016\/j.infrared.2019.103144","article-title":"LBP-BEGAN: A generative adversarial network architecture for infrared and visible image fusion","volume":"104","author":"Xu","year":"2020","journal-title":"Infrared Phys. Technol."},{"key":"10.1016\/j.image.2022.116722_b27","series-title":"ECCV (8)","first-page":"715","article-title":"A style-aware content loss for real-time HD style transfer","volume":"vol. 11212","author":"Sanakoyeu","year":"2018"},{"key":"10.1016\/j.image.2022.116722_b28","series-title":"2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"1800","article-title":"Xception: Deep learning with depthwise separable convolutions","author":"Chollet","year":"2017"},{"key":"10.1016\/j.image.2022.116722_b29","series-title":"Proceedings of the 36th International Conference on Machine Learning","first-page":"7324","article-title":"Making convolutional networks shift-invariant again","volume":"vol. 97","author":"Zhang","year":"2019"},{"key":"10.1016\/j.image.2022.116722_b30","series-title":"2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"2261","article-title":"Densely connected convolutional networks","author":"Huang","year":"2017"},{"key":"10.1016\/j.image.2022.116722_b31","series-title":"MICCAI (3)","first-page":"234","article-title":"U-Net: Convolutional networks for biomedical image segmentation","volume":"vol. 9351","author":"Ronneberger","year":"2015"},{"key":"10.1016\/j.image.2022.116722_b32","series-title":"ECCV (25)","first-page":"492","article-title":"Learning enriched features for real image restoration and enhancement","volume":"vol. 12370","author":"Zamir","year":"2020"},{"key":"10.1016\/j.image.2022.116722_b33","series-title":"WACV","first-page":"3559","article-title":"Attentional feature fusion","author":"Dai","year":"2021"},{"key":"10.1016\/j.image.2022.116722_b34","series-title":"ICLR","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2015"},{"key":"10.1016\/j.image.2022.116722_b35","series-title":"Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI-17","first-page":"2230","article-title":"Demystifying neural style transfer","author":"Li","year":"2017"},{"issue":"4","key":"10.1016\/j.image.2022.116722_b36","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.image.2022.116722_b37","series-title":"The Thrity-Seventh Asilomar Conference on Signals, Systems Computers, 2003, Vol. 2","first-page":"1398","article-title":"Multiscale structural similarity for image quality assessment","author":"Wang","year":"2003"},{"key":"10.1016\/j.image.2022.116722_b38","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.inffus.2016.02.001","article-title":"Infrared and visible image fusion via gradient transfer and total variation minimization","volume":"31","author":"Ma","year":"2016","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.image.2022.116722_b39","doi-asserted-by":"crossref","first-page":"4980","DOI":"10.1109\/TIP.2020.2977573","article-title":"DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion","volume":"29","author":"Ma","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.image.2022.116722_b40","first-page":"1","article-title":"Assessment of image fusion procedures using entropy, image quality, and multispectral classification","volume":"2","author":"Roberts","year":"2008","journal-title":"J. Appl. Remote Sens."},{"key":"10.1016\/j.image.2022.116722_b41","doi-asserted-by":"crossref","first-page":"313","DOI":"10.1049\/el:20020212","article-title":"Information measure for performance of image fusion","volume":"38","author":"Qu","year":"2002","journal-title":"Electron. Lett."},{"issue":"4","key":"10.1016\/j.image.2022.116722_b42","doi-asserted-by":"crossref","first-page":"355","DOI":"10.1088\/0957-0233\/8\/4\/002","article-title":"In-fibre Bragg grating sensors","volume":"8","author":"Rao","year":"1997","journal-title":"Meas. Sci. Technol."},{"issue":"4","key":"10.1016\/j.image.2022.116722_b43","doi-asserted-by":"crossref","first-page":"308","DOI":"10.1049\/el:20000267","article-title":"Objective image fusion performance measure","volume":"36","author":"Xydeas","year":"2000","journal-title":"Electron. Lett."},{"issue":"2","key":"10.1016\/j.image.2022.116722_b44","doi-asserted-by":"crossref","first-page":"127","DOI":"10.1016\/j.inffus.2011.08.002","article-title":"A new image fusion performance metric based on visual information fidelity","volume":"14","author":"Han","year":"2013","journal-title":"Inf. Fusion"},{"issue":"11","key":"10.1016\/j.image.2022.116722_b45","doi-asserted-by":"crossref","first-page":"3345","DOI":"10.1109\/TIP.2015.2442920","article-title":"Perceptual quality assessment for multi-exposure image fusion","volume":"24","author":"Ma","year":"2015","journal-title":"IEEE Trans. Image Process."},{"issue":"12","key":"10.1016\/j.image.2022.116722_b46","doi-asserted-by":"crossref","first-page":"1890","DOI":"10.1016\/j.aeue.2015.09.004","article-title":"A new image quality metric for image fusion: The sum of the correlations of differences","volume":"69","author":"Aslantas","year":"2015","journal-title":"AEU - Int. J. Electron. Commun."},{"issue":"6","key":"10.1016\/j.image.2022.116722_b47","doi-asserted-by":"crossref","first-page":"1125","DOI":"10.1007\/s11760-012-0361-x","article-title":"Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform","volume":"7","author":"Shreyamsha\u00a0Kumar","year":"2013","journal-title":"Signal Image Video Process."},{"key":"10.1016\/j.image.2022.116722_b48","series-title":"ICCV","first-page":"618","article-title":"Grad-CAM: Visual explanations from deep networks via gradient-based localization","author":"Selvaraju","year":"2017"},{"issue":"4","key":"10.1016\/j.image.2022.116722_b49","doi-asserted-by":"crossref","first-page":"2049","DOI":"10.1109\/TIP.2018.2794218","article-title":"Learning a deep single image contrast enhancer from multi-exposure images","volume":"27","author":"Cai","year":"2018","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.image.2022.116722_b50","doi-asserted-by":"crossref","first-page":"7203","DOI":"10.1109\/TIP.2020.2999855","article-title":"MEF-GAN: Multi-exposure image fusion via generative adversarial networks","volume":"29","author":"Xu","year":"2020","journal-title":"IEEE Trans. Image Process."}],"container-title":["Signal Processing: Image Communication"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0923596522000571?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0923596522000571?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T05:34:37Z","timestamp":1760333677000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0923596522000571"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8]]},"references-count":50,"alternative-id":["S0923596522000571"],"URL":"https:\/\/doi.org\/10.1016\/j.image.2022.116722","relation":{},"ISSN":["0923-5965"],"issn-type":[{"value":"0923-5965","type":"print"}],"subject":[],"published":{"date-parts":[[2022,8]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"StyleFuse: An unsupervised network based on style loss function for infrared and visible image fusion","name":"articletitle","label":"Article Title"},{"value":"Signal Processing: Image Communication","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.image.2022.116722","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2022 Elsevier B.V. All rights reserved.","name":"copyright","label":"Copyright"}],"article-number":"116722"}}