{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T00:10:41Z","timestamp":1760141441055,"version":"build-2065373602"},"reference-count":42,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100002663","name":"Northwestern Polytechnical University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100002663","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neurocomputing"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1016\/j.neucom.2025.131382","type":"journal-article","created":{"date-parts":[[2025,9,13]],"date-time":"2025-09-13T16:27:17Z","timestamp":1757780837000},"page":"131382","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Breaking the alignment barrier: A spatiotemporal alignment-free RGBT tracking approach"],"prefix":"10.1016","volume":"656","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-2711-5708","authenticated-orcid":false,"given":"Xiaodong","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9781-5462","authenticated-orcid":false,"given":"Meibo","family":"Lv","sequence":"additional","affiliation":[]},{"given":"Daming","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Lingyu","family":"Si","sequence":"additional","affiliation":[]},{"given":"Ruiheng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Hui","family":"Xu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neucom.2025.131382_bib0005","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TII.2019.2947293","article-title":"Robust multi-modality anchor graph-based label prediction for RGB-infrared tracking","author":"Lan","year":"2024","journal-title":"IEEE Trans. Ind. Inf."},{"key":"10.1016\/j.neucom.2025.131382_bib0010","doi-asserted-by":"crossref","DOI":"10.1007\/s11042-024-20546-8","article-title":"Residual channel-attention (rca) network for remote sensing image scene classification","author":"Gomaa","year":"2025","journal-title":"Multimed. Tools Appl."},{"key":"10.1016\/j.neucom.2025.131382_bib0015","series-title":"2024 6th Novel Intelligent and Leading Emerging Sciences Conference (NILES)","first-page":"211","article-title":"Advanced domain adaptation technique for object detection leveraging semi-automated dataset construction and enhanced YOLOv8","author":"Gomaa","year":"2024"},{"issue":"35","key":"10.1016\/j.neucom.2025.131382_bib0020","doi-asserted-by":"crossref","first-page":"26023","DOI":"10.1007\/s11042-020-09242-5","article-title":"Efficient vehicle detection and tracking strategy in aerial videos by employing morphological operations and feature points motion analysis","volume":"79","author":"Gomaa","year":"2020","journal-title":"Multimed. Tools Appl."},{"key":"10.1016\/j.neucom.2025.131382_bib0025","series-title":"IGARSS 2023 - 2023 IEEE International Geoscience and Remote Sensing Symposium","first-page":"2350","article-title":"Detection of earthquake-induced building damages using remote sensing data and deep learning: a case study of Mashiki town, Japan","author":"Salem","year":"2023"},{"issue":"6","key":"10.1016\/j.neucom.2025.131382_bib0030","article-title":"Novel deep learning domain adaptation approach for object detection using semi-self building dataset and modified YOLOv4","volume":"15","author":"Gomaa","year":"2024","journal-title":"World Electr. Veh. J."},{"key":"10.1016\/j.neucom.2025.131382_bib0035","series-title":"2018 IEEE 61st International Midwest Symposium on Circuits and Systems (MWSCAS)","first-page":"222","article-title":"Real-time algorithm for simultaneous vehicle detection and tracking in aerial view videos","author":"Gomaa","year":"2018"},{"key":"10.1016\/j.neucom.2025.131382_bib0040","doi-asserted-by":"crossref","DOI":"10.1088\/1361-6501\/ad633d","article-title":"Enhanced feature extraction YOLO industrial small object detection algorithm based on receptive-field attention and multi-scale features","volume":"35","author":"Tao","year":"2024","journal-title":"Meas. Sci. Technol."},{"key":"10.1016\/j.neucom.2025.131382_bib0045","doi-asserted-by":"crossref","DOI":"10.1016\/j.aei.2024.102986","article-title":"Pseudo-label guided dual classifier domain adversarial network for unsupervised cross-domain fault diagnosis with small samples","volume":"64","author":"Sun","year":"2025","journal-title":"Adv. Eng. Inform."},{"key":"10.1016\/j.neucom.2025.131382_bib0050","doi-asserted-by":"crossref","first-page":"10990","DOI":"10.1002\/rnc.7553","article-title":"Interval observer design for unobservable switched nonlinear partial differential equation systems and its application","volume":"34","author":"Song","year":"2024","journal-title":"Int. J. Robust Nonlinear Control"},{"key":"10.1016\/j.neucom.2025.131382_bib0055","doi-asserted-by":"crossref","first-page":"3473","DOI":"10.1002\/acs.3885","article-title":"Spatiotemporal fault estimation for switched nonlinear reaction\u2013diffusion systems via adaptive iterative learning","volume":"38","author":"Peng","year":"2024","journal-title":"Int. J. Adapt. Control Signal Process."},{"key":"10.1016\/j.neucom.2025.131382_bib0060","series-title":"European Conference on Computer Vision","article-title":"Distractor-aware siamese networks for visual object tracking","author":"Zhu","year":"2018"},{"key":"10.1016\/j.neucom.2025.131382_bib0065","first-page":"1","article-title":"Detail-aware network for infrared image enhancement","volume":"63","author":"Zhang","year":"2025","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"10.1016\/j.neucom.2025.131382_bib0070","series-title":"2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW)","article-title":"Multi-adapter rgbt tracking","author":"Li","year":"2019"},{"key":"10.1016\/j.neucom.2025.131382_bib0075","series-title":"2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW)","article-title":"Deep adaptive fusion network for high performance rgbt tracking","author":"Gao","year":"2019"},{"issue":"1","key":"10.1016\/j.neucom.2025.131382_bib0080","doi-asserted-by":"crossref","first-page":"121","DOI":"10.1109\/TIV.2020.2980735","article-title":"Quality-aware feature aggregation network for robust rgbt tracking","volume":"6","author":"Zhu","year":"2021","journal-title":"IEEE Trans. Intell. Veh."},{"key":"10.1016\/j.neucom.2025.131382_bib0085","series-title":"2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","article-title":"Bridging search region interaction with template for RGB-T tracking","author":"Hui","year":"2023"},{"key":"10.1016\/j.neucom.2025.131382_bib0090","series-title":"2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"4293","article-title":"Learning multi-domain convolutional neural networks for visual tracking","author":"Nam","year":"2016"},{"issue":"2","key":"10.1016\/j.neucom.2025.131382_bib0095","doi-asserted-by":"crossref","first-page":"1900","DOI":"10.1109\/TITS.2024.3512551","article-title":"Siamtfa: Siamese triple-stream feature aggregation network for efficient rgbt tracking","volume":"26","author":"Zhang","year":"2025","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.neucom.2025.131382_bib0100","first-page":"1","article-title":"Duality-gated mutual condition network for rgbt tracking","author":"Lu","year":"2024","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.neucom.2025.131382_bib0105","series-title":"2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","article-title":"Visible-thermal UAV tracking: a large-scale benchmark and new baseline","author":"Zhang","year":"2022"},{"key":"10.1016\/j.neucom.2025.131382_bib0110","doi-asserted-by":"crossref","first-page":"1247","DOI":"10.1007\/s40815-023-01664-1","article-title":"Improved event-triggered-based output tracking for a class of delayed networked t-s fuzzy systems","volume":"26","author":"Aslam","year":"2024","journal-title":"Int. J. Fuzzy Syst."},{"key":"10.1016\/j.neucom.2025.131382_bib0115","doi-asserted-by":"crossref","first-page":"2247","DOI":"10.1080\/00207721.2014.981236","article-title":"Stochastic sampled data robust stabilisation of t-s fuzzy neutral systems with randomly occurring uncertainties and time-varying delays","volume":"47","author":"Rakkiyappan","year":"2016","journal-title":"Int. J. Syst. Sci."},{"issue":"9","key":"10.1016\/j.neucom.2025.131382_bib0120","doi-asserted-by":"crossref","first-page":"2714","DOI":"10.1007\/s11263-021-01495-3","article-title":"Learning adaptive attribute-driven representation for real-time RGB-T tracking","volume":"129","author":"Zhang","year":"2021","journal-title":"Int. J. Comput. Vis."},{"issue":"3","key":"10.1016\/j.neucom.2025.131382_bib0125","first-page":"2831","article-title":"Attribute-based progressive fusion network for rgbt tracking","volume":"36","author":"Xiao","year":"2022","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"10.1016\/j.neucom.2025.131382_bib0130","series-title":"2019 IEEE\/CVF International Conference on Computer Vision (ICCV)","first-page":"6181","article-title":"Learning discriminative model prediction for tracking","author":"Bhat","year":"2019"},{"key":"10.1016\/j.neucom.2025.131382_bib0135","series-title":"2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW)","article-title":"Multi-modal fusion for end-to-end RGB-T tracking","author":"Zhang","year":"2019"},{"key":"10.1016\/j.neucom.2025.131382_bib0140","first-page":"4009","article-title":"Learning the model update for siamese trackers","author":"Zhang","year":"2019","journal-title":"IEEE\/CVF Int. Conf. Comput. Vis."},{"key":"10.1016\/j.neucom.2025.131382_bib0145","first-page":"1","article-title":"Putrack: improved underwater object tracking via progressive prompting","author":"Zhang","year":"2025","journal-title":"IEEE Trans. Ind. Inf."},{"key":"10.1016\/j.neucom.2025.131382_bib0150","doi-asserted-by":"crossref","first-page":"15156","DOI":"10.1109\/TNNLS.2023.3282905","article-title":"Learning cross-attention discriminators via alternating time\u2013space transformers for visual tracking","volume":"35","author":"Wang","year":"2023","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"12","key":"10.1016\/j.neucom.2025.131382_bib0155","doi-asserted-by":"crossref","first-page":"5743","DOI":"10.1109\/TIP.2016.2614135","article-title":"Learning collaborative sparse representation for grayscale-thermal tracking","volume":"25","author":"Li","year":"2016","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.neucom.2025.131382_bib0160","series-title":"Proceedings of the 25th ACM International Conference on Multimedia","article-title":"Weighted sparse representation regularized graph learning for RGB-T object tracking","author":"Li","year":"2017"},{"key":"10.1016\/j.neucom.2025.131382_bib0165","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2019.106977","article-title":"RGB-T object tracking: benchmark and baseline","volume":"96","author":"Li","year":"2019","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.neucom.2025.131382_bib0170","doi-asserted-by":"crossref","first-page":"392","DOI":"10.1109\/TIP.2021.3130533","article-title":"Lasher: a large-scale high-diversity benchmark for rgbt tracking","volume":"31","author":"Li","year":"2022","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"10.1016\/j.neucom.2025.131382_bib0175","doi-asserted-by":"crossref","first-page":"393","DOI":"10.3390\/s20020393","article-title":"Object tracking in RGB-T videos using modal-aware attention network and competitive learning","volume":"20","author":"Zhang","year":"2020","journal-title":"Sensors"},{"key":"10.1016\/j.neucom.2025.131382_bib0180","series-title":"Challenge-Aware RGBT Tracking","first-page":"222","author":"Li","year":"2020"},{"key":"10.1016\/j.neucom.2025.131382_bib0185","doi-asserted-by":"crossref","first-page":"5613","DOI":"10.1109\/TIP.2021.3087341","article-title":"rgbt tracking via multi-adapter network with hierarchical divergence loss","volume":"30","author":"Lu","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.neucom.2025.131382_bib0190","doi-asserted-by":"crossref","first-page":"3335","DOI":"10.1109\/TIP.2021.3060862","article-title":"Jointly modeling motion and appearance cues for robust RGB-T tracking","volume":"30","author":"Zhang","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.neucom.2025.131382_bib0195","series-title":"2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"9516","article-title":"Visual prompt multi-modal tracking","author":"Zhu","year":"2023"},{"key":"10.1016\/j.neucom.2025.131382_bib0200","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.101881","article-title":"Exploring fusion strategies for accurate rgbt visual object tracking","volume":"99","author":"Tang","year":"2023","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.neucom.2025.131382_bib0205","series-title":"Proceedings of the ACM International Conference on Multimedia","first-page":"3129","article-title":"Quality-aware rgbt tracking via supervised reliability learning and weighted residual guidance","author":"Liu","year":"2023"},{"key":"10.1016\/j.neucom.2025.131382_bib0210","series-title":"AAAI Conference on Artificial Intelligence","article-title":"Bi-directional adapter for multimodal tracking","author":"Zhu","year":"2024"}],"container-title":["Neurocomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231225020545?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231225020545?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T01:15:19Z","timestamp":1760058919000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0925231225020545"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":42,"alternative-id":["S0925231225020545"],"URL":"https:\/\/doi.org\/10.1016\/j.neucom.2025.131382","relation":{},"ISSN":["0925-2312"],"issn-type":[{"type":"print","value":"0925-2312"}],"subject":[],"published":{"date-parts":[[2025,12]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Breaking the alignment barrier: A spatiotemporal alignment-free RGBT tracking approach","name":"articletitle","label":"Article Title"},{"value":"Neurocomputing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neucom.2025.131382","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"131382"}}