{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T05:32:07Z","timestamp":1776835927478,"version":"3.51.2"},"reference-count":82,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T00:00:00Z","timestamp":1769904000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100013318","name":"Shanxi Provincial Department of Science and Technology","doi-asserted-by":"publisher","award":["YDZJSX2022A011"],"award-info":[{"award-number":["YDZJSX2022A011"]}],"id":[{"id":"10.13039\/501100013318","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100022976","name":"National Institute of Natural Hazards","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100022976","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62076177"],"award-info":[{"award-number":["62076177"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62027819"],"award-info":[{"award-number":["62027819"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["clinicalkey.com","clinicalkey.com.au","clinicalkey.es","clinicalkey.fr","clinicalkey.jp","elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Computerized Medical Imaging and Graphics"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1016\/j.compmedimag.2026.102719","type":"journal-article","created":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T01:00:05Z","timestamp":1770253205000},"page":"102719","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"C","title":["HF-VLP: A multimodal vision-language pre-trained model for diagnosing heart failure"],"prefix":"10.1016","volume":"128","author":[{"given":"Huiting","family":"Ma","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8310-7684","authenticated-orcid":false,"given":"Dengao","family":"Li","sequence":"additional","affiliation":[]},{"given":"Guiji","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Li","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Xiaole","family":"Fan","sequence":"additional","affiliation":[]},{"given":"Zhe","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yuchen","family":"Liang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"8","key":"10.1016\/j.compmedimag.2026.102719_b1","first-page":"1153","article-title":"Artificial intelligence to diagnose heart failure based on chest X-Rays and potential clinical implications","volume":"37","author":"Adams","year":"2021","journal-title":"Can. J. Cardiol."},{"key":"10.1016\/j.compmedimag.2026.102719_b2","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2023.107448","article-title":"Cov-Fed: Federated learning-based framework for COVID-19 diagnosis using chest X-ray scans","volume":"128","author":"Adjei-Mensah","year":"2024","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.compmedimag.2026.102719_b3","doi-asserted-by":"crossref","unstructured":"Anand, Sidharth, Devulapally, Naresh Kumar, Bhattacharjee, Sreyasee Das, Yuan, Junsong, 2023. Multi-label Emotion Analysis in Conversation via Multimodal Knowledge Distillation. In: Proceedings of the 31st ACM International Conference on Multimedia. pp. 6090\u20136100.","DOI":"10.1145\/3581783.3612517"},{"key":"10.1016\/j.compmedimag.2026.102719_b4","doi-asserted-by":"crossref","unstructured":"Bannur, Shruthi, Hyland, Stephanie, Liu, Qianchu, Perez-Garcia, Fernando, Ilse, Maximilian, Castro, Daniel C, Boecking, Benedikt, Sharma, Harshita, Bouzid, Kenza, Thieme, Anja, et al., 2023. Learning to exploit temporal structure for biomedical vision-language processing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 15016\u201315027.","DOI":"10.1109\/CVPR52729.2023.01442"},{"key":"10.1016\/j.compmedimag.2026.102719_b5","series-title":"European Conference on Computer Vision","first-page":"1","article-title":"Making the most of text semantics to improve biomedical vision\u2013language processing","author":"Boecking","year":"2022"},{"key":"10.1016\/j.compmedimag.2026.102719_b6","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.compmedimag.2026.102719_b7","series-title":"International Conference on Medical Image Computing and Computer-Assisted Intervention","first-page":"679","article-title":"Multi-modal masked autoencoders for medical vision-and-language pre-training","author":"Chen","year":"2022"},{"issue":"8","key":"10.1016\/j.compmedimag.2026.102719_b8","doi-asserted-by":"crossref","first-page":"12","DOI":"10.1109\/TMI.2024.3381123","article-title":"UniChest: Conquer-and-divide Pre-Training for Multi-Source chest X-Ray classification","volume":"43","author":"Dai","year":"2024","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"2","key":"10.1016\/j.compmedimag.2026.102719_b9","doi-asserted-by":"crossref","first-page":"304","DOI":"10.1093\/jamia\/ocv080","article-title":"Preparing a collection of radiology examinations for distribution and retrieval","volume":"23","author":"Demner-Fushman","year":"2016","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"10.1016\/j.compmedimag.2026.102719_b10","series-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018"},{"key":"10.1016\/j.compmedimag.2026.102719_b11","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102536","article-title":"Deep learning based multimodal biomedical data fusion: An overview and comparative review","author":"Duan","year":"2024","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.compmedimag.2026.102719_b12","doi-asserted-by":"crossref","first-page":"12","DOI":"10.1016\/j.inffus.2016.02.006","article-title":"The basic principles of uncertain information fusion. An organised review of merging rules in different representation frameworks","volume":"32","author":"Dubois","year":"2016","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.compmedimag.2026.102719_b13","doi-asserted-by":"crossref","DOI":"10.1016\/j.ebiom.2025.105657","article-title":"Identification of heart failure subtypes using transformer-based deep learning modelling: a population-based study of 379,108 individuals","volume":"114","author":"Fan","year":"2025","journal-title":"EBioMedicine"},{"key":"10.1016\/j.compmedimag.2026.102719_b14","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.101955","article-title":"Explainability meets uncertainty quantification: Insights from feature-based model fusion on multimodal time series","volume":"100","author":"Folgado","year":"2023","journal-title":"Inf. Fusion"},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b15","first-page":"1","article-title":"Domain-specific language model pretraining for biomedical natural language processing","volume":"3","author":"Gu","year":"2021","journal-title":"ACM Trans. Comput. Health (Health)"},{"key":"10.1016\/j.compmedimag.2026.102719_b16","doi-asserted-by":"crossref","DOI":"10.1016\/j.compmedimag.2025.102617","article-title":"Beyond unimodal analysis: Multimodal ensemble learning for enhanced assessment of atherosclerotic disease progression","author":"Guarrasi","year":"2025","journal-title":"Comput. Med. Imaging Graph."},{"key":"10.1016\/j.compmedimag.2026.102719_b17","series-title":"Medical Imaging 2023: Computer-Aided Diagnosis","first-page":"722","article-title":"Cross modal global local representation learning from radiology reports and x-ray chest images","volume":"Vol. 12465","author":"Hadjiyski","year":"2023"},{"key":"10.1016\/j.compmedimag.2026.102719_b18","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102666","article-title":"MMIF-INet: Multimodal medical image fusion by invertible network","volume":"114","author":"He","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.compmedimag.2026.102719_b19","doi-asserted-by":"crossref","unstructured":"He, Kaiming, Zhang, Xiangyu, Ren, Shaoqing, Sun, Jian, 2016. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 770\u2013778.","DOI":"10.1109\/CVPR.2016.90"},{"key":"10.1016\/j.compmedimag.2026.102719_b20","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2024.103279","article-title":"Interpretable medical image visual question answering via multi-modal relationship graph learning","volume":"97","author":"Hu","year":"2024","journal-title":"Med. Image Anal."},{"key":"10.1016\/j.compmedimag.2026.102719_b21","series-title":"Lora: Low-rank adaptation of large language models","author":"Hu","year":"2021"},{"key":"10.1016\/j.compmedimag.2026.102719_b22","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102648","article-title":"Deep evidential fusion with uncertainty quantification and reliability learning for multimodal medical image segmentation","volume":"113","author":"Huang","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.compmedimag.2026.102719_b23","doi-asserted-by":"crossref","unstructured":"Huang, Shih-Cheng, Shen, Liyue, Lungren, Matthew P, Yeung, Serena, 2021. Gloria: A multimodal global-local representation learning framework for label-efficient medical image recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 3942\u20133951.","DOI":"10.1109\/ICCV48922.2021.00391"},{"key":"10.1016\/j.compmedimag.2026.102719_b24","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2023.119900","article-title":"Lightweight deep CNN-based models for early detection of COVID-19 patients from chest X-ray images","volume":"223","author":"Hussein","year":"2023","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.compmedimag.2026.102719_b25","doi-asserted-by":"crossref","unstructured":"Irvin, Jeremy, Rajpurkar, Pranav, Ko, Michael, Yu, Yifan, Ciurea-Ilcus, Silviana, Chute, Chris, Marklund, Henrik, Haghgoo, Behzad, Ball, Robyn, Shpanskaya, Katie, et al., 2019. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In: Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 33, pp. 590\u2013597.","DOI":"10.1609\/aaai.v33i01.3301590"},{"issue":"10159","key":"10.1016\/j.compmedimag.2026.102719_b26","doi-asserted-by":"crossref","first-page":"1789","DOI":"10.1016\/S0140-6736(18)32279-7","article-title":"Global, regional, and national incidence, prevalence, and years lived with disability for 354 diseases and injuries for 195 countries and territories, 1990\u20132017: a systematic analysis for the Global Burden of Disease Study 2017","volume":"392","author":"James","year":"2018","journal-title":"Lancet"},{"key":"10.1016\/j.compmedimag.2026.102719_b27","series-title":"MIMIC-CXR-JPG, a large publicly available database of labeled chest radiographs","author":"Johnson","year":"2019"},{"key":"10.1016\/j.compmedimag.2026.102719_b28","series-title":"A rank stabilization scaling factor for fine-tuning with lora","author":"Kalajdzievski","year":"2023"},{"key":"10.1016\/j.compmedimag.2026.102719_b29","doi-asserted-by":"crossref","unstructured":"Kong, Shu, Fowlkes, Charless, 2017. Low-rank bilinear pooling for fine-grained classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 365\u2013374.","DOI":"10.1109\/CVPR.2017.743"},{"issue":"11","key":"10.1016\/j.compmedimag.2026.102719_b30","first-page":"908","article-title":"Biomarkers in the clinical management of patients with atrial fibrillation and heart failure","volume":"18","author":"Koniari","year":"2021","journal-title":"J. Geriatr. Cardiol.: JGC"},{"key":"10.1016\/j.compmedimag.2026.102719_b31","series-title":"International Conference on Machine Learning","first-page":"19730","article-title":"Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models","author":"Li","year":"2023"},{"key":"10.1016\/j.compmedimag.2026.102719_b32","series-title":"International Conference on Machine Learning","first-page":"12888","article-title":"BLIP: Bootstrapping Language-Image Pre-training for unified Vision-Language understanding and generation","author":"Li","year":"2022"},{"key":"10.1016\/j.compmedimag.2026.102719_b33","doi-asserted-by":"crossref","DOI":"10.1016\/j.compmedimag.2025.102694","article-title":"TGIAlign: Text-guided dual-branch bidirectional framework for cross-modal semantic alignment in medical vision-language","author":"Li","year":"2026","journal-title":"Comput. Med. Imaging Graph."},{"key":"10.1016\/j.compmedimag.2026.102719_b34","article-title":"Llava-med: Training a large language-and-vision assistant for biomedicine in one day","volume":"36","author":"Li","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.compmedimag.2026.102719_b35","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2022.104538","article-title":"Diagnosis of heart failure from imbalance datasets using multi-level classification","volume":"81","author":"Li","year":"2023","journal-title":"Biomed. Signal Process. Control."},{"key":"10.1016\/j.compmedimag.2026.102719_b36","series-title":"Dora: Weight-decomposed low-rank adaptation","author":"Liu","year":"2024"},{"key":"10.1016\/j.compmedimag.2026.102719_b37","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2023.107037","article-title":"Scaling-up medical vision-and-language representation learning with federated learning","volume":"126","author":"Lu","year":"2023","journal-title":"Eng. Appl. Artif. Intell."},{"issue":"4","key":"10.1016\/j.compmedimag.2026.102719_b38","doi-asserted-by":"crossref","first-page":"781","DOI":"10.1536\/ihj.19-714","article-title":"Diagnosing heart failure from chest X-ray images using deep learning","volume":"61","author":"Matsumoto","year":"2020","journal-title":"Int. Heart J."},{"key":"10.1016\/j.compmedimag.2026.102719_b39","series-title":"International Conference on Medical Image Computing and Computer-Assisted Intervention","first-page":"725","article-title":"Berthop: An effective vision-and-language model for chest x-ray disease diagnosis","author":"Monajatipoor","year":"2022"},{"issue":"12","key":"10.1016\/j.compmedimag.2026.102719_b40","doi-asserted-by":"crossref","first-page":"6070","DOI":"10.1109\/JBHI.2022.3207502","article-title":"Multi-modal understanding and generation for medical images and text via vision-language pre-training","volume":"26","author":"Moon","year":"2022","journal-title":"IEEE J. Biomed. Health Inform."},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b41","doi-asserted-by":"crossref","first-page":"429","DOI":"10.1038\/s41597-022-01498-w","article-title":"VinDr-CXR: An open dataset of chest X-rays with radiologist\u2019s annotations","volume":"9","author":"Nguyen","year":"2022","journal-title":"Sci. Data"},{"key":"10.1016\/j.compmedimag.2026.102719_b42","doi-asserted-by":"crossref","DOI":"10.1016\/j.compmedimag.2023.102277","article-title":"Deep reinforcement learning framework for thoracic diseases classification via prior knowledge guidance","volume":"108","author":"Nie","year":"2023","journal-title":"Comput. Med. Imaging Graph."},{"key":"10.1016\/j.compmedimag.2026.102719_b43","doi-asserted-by":"crossref","DOI":"10.1016\/j.compmedimag.2023.102320","article-title":"Deep learning for report generation on chest X-ray images","volume":"111","author":"Ouis","year":"2024","journal-title":"Comput. Med. Imaging Graph."},{"issue":"10","key":"10.1016\/j.compmedimag.2026.102719_b44","doi-asserted-by":"crossref","first-page":"2698","DOI":"10.1109\/TMI.2020.3042773","article-title":"Learning hierarchical attention for weakly-supervised chest X-ray abnormality localization and diagnosis","volume":"40","author":"Ouyang","year":"2020","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b45","doi-asserted-by":"crossref","first-page":"3848","DOI":"10.1038\/s41467-022-31514-x","article-title":"Self-evolving vision transformer for chest X-ray diagnosis through knowledge distillation","volume":"13","author":"Park","year":"2022","journal-title":"Nat. Commun."},{"key":"10.1016\/j.compmedimag.2026.102719_b46","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2023.106126","article-title":"Vision Transformers in medical computer vision\u2014A contemplative retrospection","volume":"122","author":"Parvaiz","year":"2023","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.compmedimag.2026.102719_b47","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2020.101911","article-title":"Discriminative ensemble learning for few-shot chest x-ray diagnosis","volume":"68","author":"Paul","year":"2021","journal-title":"Med. Image Anal."},{"key":"10.1016\/j.compmedimag.2026.102719_b48","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2024.103225","article-title":"MMGPL: Multimodal medical data analysis with graph prompt learning","author":"Peng","year":"2024","journal-title":"Med. Image Anal."},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b49","first-page":"19","article-title":"COVID-19 diagnosis from chest X-ray images using transfer learning: Enhanced performance by debiasing dataloader","volume":"29","author":"Polat","year":"2021","journal-title":"J. X-Ray Sci. Technol."},{"key":"10.1016\/j.compmedimag.2026.102719_b50","doi-asserted-by":"crossref","DOI":"10.1016\/j.compmedimag.2023.102232","article-title":"Semantically preserving adversarial unsupervised domain adaptation network for improving disease recognition from chest x-rays","volume":"107","author":"Qin","year":"2023","journal-title":"Comput. Med. Imaging Graph."},{"key":"10.1016\/j.compmedimag.2026.102719_b51","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s12938-018-0544-y","article-title":"Computer-aided detection in chest radiography based on artificial intelligence: a survey","volume":"17","author":"Qin","year":"2018","journal-title":"Biomed. Eng. Online"},{"issue":"8","key":"10.1016\/j.compmedimag.2026.102719_b52","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"issue":"6","key":"10.1016\/j.compmedimag.2026.102719_b53","doi-asserted-by":"crossref","first-page":"4756","DOI":"10.1007\/s10489-024-05420-2","article-title":"A novel lightweight CNN for chest X-ray-based lung disease identification on heterogeneous embedded system","volume":"54","author":"Sanida","year":"2024","journal-title":"Appl. Intell."},{"issue":"17","key":"10.1016\/j.compmedimag.2026.102719_b54","doi-asserted-by":"crossref","first-page":"3272","DOI":"10.1093\/cvr\/cvac013","article-title":"Global burden of heart failure: a comprehensive and updated review of epidemiology","volume":"118","author":"Savarese","year":"2022","journal-title":"Cardiovasc. Res."},{"key":"10.1016\/j.compmedimag.2026.102719_b55","series-title":"Outrageously large neural networks: The sparsely-gated mixture-of-experts layer","author":"Shazeer","year":"2017"},{"key":"10.1016\/j.compmedimag.2026.102719_b56","article-title":"Semi-supervised medical image classification via feature-level multi-scale consistency and adversarial training","author":"Shiyan","year":"2025","journal-title":"Comput. Med. Imaging Graph."},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b57","doi-asserted-by":"crossref","first-page":"9025","DOI":"10.1038\/s41598-023-35922-x","article-title":"Pneumonia detection with QCSA network on chest X-ray","volume":"13","author":"Singh","year":"2023","journal-title":"Sci. Rep."},{"key":"10.1016\/j.compmedimag.2026.102719_b58","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2023.127063","article-title":"Roformer: Enhanced transformer with rotary position embedding","volume":"568","author":"Su","year":"2024","journal-title":"Neurocomputing"},{"key":"10.1016\/j.compmedimag.2026.102719_b59","series-title":"International Conference on Machine Learning","first-page":"6105","article-title":"Efficientnet: Rethinking model scaling for convolutional neural networks","author":"Tan","year":"2019"},{"issue":"3","key":"10.1016\/j.compmedimag.2026.102719_b60","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1038\/s41569-018-0108-7","article-title":"Dietary metabolism, the gut microbiome, and heart failure","volume":"16","author":"Tang","year":"2019","journal-title":"Nat. Rev. Cardiol."},{"issue":"11","key":"10.1016\/j.compmedimag.2026.102719_b61","first-page":"35","article-title":"Clinical prompt learning with frozen language models","author":"Taylor","year":"2024","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"12","key":"10.1016\/j.compmedimag.2026.102719_b62","doi-asserted-by":"crossref","first-page":"1399","DOI":"10.1038\/s41551-022-00936-9","article-title":"Expert-level detection of pathologies from unannotated chest X-ray images via self-supervised learning","volume":"6","author":"Tiu","year":"2022","journal-title":"Nat. Biomed. Eng."},{"issue":"8","key":"10.1016\/j.compmedimag.2026.102719_b63","doi-asserted-by":"crossref","first-page":"e93","DOI":"10.1161\/CIR.0000000000001123","article-title":"Heart disease and stroke statistics\u20142023 update: a report from the American Heart Association","volume":"147","author":"Tsao","year":"2023","journal-title":"Circulation"},{"key":"10.1016\/j.compmedimag.2026.102719_b64","doi-asserted-by":"crossref","DOI":"10.1017\/ATSIP.2020.27","article-title":"Subspace learning for facial expression recognition: an overview and a new perspective","volume":"10","author":"Turan","year":"2021","journal-title":"APSIPA Trans. Signal Inf. Process."},{"key":"10.1016\/j.compmedimag.2026.102719_b65","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"4","key":"10.1016\/j.compmedimag.2026.102719_b66","doi-asserted-by":"crossref","first-page":"352","DOI":"10.1016\/j.jaapos.2007.11.022","article-title":"Agreement among pediatric ophthalmologists in diagnosing plus and pre-plus disease in retinopathy of prematurity","volume":"12","author":"Wallace","year":"2008","journal-title":"J. Am. Assoc. Pediatr. Ophthalmol. Strabismus"},{"key":"10.1016\/j.compmedimag.2026.102719_b67","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1016\/j.neucom.2022.01.059","article-title":"Multimodal medical image fusion based on multichannel coupled neural P systems and max-cloud models in spectral total variation domain","volume":"480","author":"Wang","year":"2022","journal-title":"Neurocomputing"},{"key":"10.1016\/j.compmedimag.2026.102719_b68","doi-asserted-by":"crossref","unstructured":"Wang, Xiaosong, Peng, Yifan, Lu, Le, Lu, Zhiyong, Bagheri, Mohammadhadi, Summers, Ronald M, 2017. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 2097\u20132106.","DOI":"10.1109\/CVPR.2017.369"},{"key":"10.1016\/j.compmedimag.2026.102719_b69","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.102016","article-title":"A multi-objective segmentation method for chest X-rays based on collaborative learning from multiple partially annotated datasets","volume":"102","author":"Wang","year":"2024","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.compmedimag.2026.102719_b70","first-page":"ocae108","article-title":"Deep learning with noisy labels in medical prediction problems: a scoping review","author":"Wei","year":"2024","journal-title":"J. Am. Med. Inform. Assoc."},{"key":"10.1016\/j.compmedimag.2026.102719_b71","doi-asserted-by":"crossref","unstructured":"Wu, Chaoyi, Zhang, Xiaoman, Zhang, Ya, Wang, Yanfeng, Xie, Weidi, 2023. Medklip: Medical knowledge enhanced language-image pre-training for x-ray diagnosis. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp. 21372\u201321383.","DOI":"10.1109\/ICCV51070.2023.01954"},{"key":"10.1016\/j.compmedimag.2026.102719_b72","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.111395","article-title":"Knowledge-based dynamic prompt learning for multi-label disease diagnosis","volume":"286","author":"Xie","year":"2024","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.compmedimag.2026.102719_b73","series-title":"Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment","author":"Xu","year":"2023"},{"key":"10.1016\/j.compmedimag.2026.102719_b74","article-title":"Interpretability research of deep learning: A literature survey","author":"Xua","year":"2024","journal-title":"Inf. Fusion"},{"issue":"6","key":"10.1016\/j.compmedimag.2026.102719_b75","doi-asserted-by":"crossref","first-page":"776","DOI":"10.1016\/j.jacc.2017.04.025","volume":"70","author":"Yancy","year":"2017","journal-title":"J. Am. Coll. Cardiol."},{"issue":"11","key":"10.1016\/j.compmedimag.2026.102719_b76","doi-asserted-by":"crossref","first-page":"3348","DOI":"10.1109\/TMI.2023.3283517","article-title":"F-DARTS: Foveated differentiable architecture search based multimodal medical image fusion","volume":"42","author":"Ye","year":"2023","journal-title":"IEEE Trans. Med. Imaging"},{"key":"10.1016\/j.compmedimag.2026.102719_b77","series-title":"Adalora: Adaptive budget allocation for parameter-efficient fine-tuning","author":"Zhang","year":"2023"},{"key":"10.1016\/j.compmedimag.2026.102719_b78","series-title":"Machine Learning for Healthcare Conference","first-page":"2","article-title":"Contrastive learning of medical visual representations from paired images and text","author":"Zhang","year":"2022"},{"key":"10.1016\/j.compmedimag.2026.102719_b79","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2021.107673","article-title":"Category supervised cross-modal hashing retrieval for chest x-ray and radiology reports","volume":"98","author":"Zhang","year":"2022","journal-title":"Comput. Electr. Eng."},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b80","doi-asserted-by":"crossref","first-page":"4542","DOI":"10.1038\/s41467-023-40260-7","article-title":"Knowledge-enhanced visual-language pre-training on chest radiology images","volume":"14","author":"Zhang","year":"2023","journal-title":"Nat. Commun."},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b81","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1038\/s42256-021-00425-9","article-title":"Generalized radiograph representation learning via cross-supervision between images and free-text radiology reports","volume":"4","author":"Zhou","year":"2022","journal-title":"Nat. Mach. Intell."},{"issue":"1","key":"10.1016\/j.compmedimag.2026.102719_b82","doi-asserted-by":"crossref","first-page":"5649","DOI":"10.1038\/s41467-024-50043-3","article-title":"Pre-trained multimodal large language model enhances dermatological diagnosis using SkinGPT-4","volume":"15","author":"Zhou","year":"2024","journal-title":"Nat. Commun."}],"container-title":["Computerized Medical Imaging and Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0895611126000224?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0895611126000224?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T11:53:51Z","timestamp":1771502031000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0895611126000224"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2]]},"references-count":82,"alternative-id":["S0895611126000224"],"URL":"https:\/\/doi.org\/10.1016\/j.compmedimag.2026.102719","relation":{},"ISSN":["0895-6111"],"issn-type":[{"value":"0895-6111","type":"print"}],"subject":[],"published":{"date-parts":[[2026,2]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"HF-VLP: A multimodal vision-language pre-trained model for diagnosing heart failure","name":"articletitle","label":"Article Title"},{"value":"Computerized Medical Imaging and Graphics","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.compmedimag.2026.102719","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"102719"}}