{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T02:07:33Z","timestamp":1773022053288,"version":"3.50.1"},"reference-count":66,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans.Inform.Forensic Secur."],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/tifs.2022.3198857","type":"journal-article","created":{"date-parts":[[2022,8,15]],"date-time":"2022-08-15T19:47:06Z","timestamp":1660592826000},"page":"3816-3830","source":"Crossref","is-referenced-by-count":53,"title":["TnT Attacks! Universal Naturalistic Adversarial Patches Against Deep Neural Network Systems"],"prefix":"10.1109","volume":"17","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6941-2765","authenticated-orcid":false,"given":"Bao Gia","family":"Doan","sequence":"first","affiliation":[{"name":"School of Computer Science, The University of Adelaide, Adelaide, SA, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9172-4252","authenticated-orcid":false,"given":"Minhui","family":"Xue","sequence":"additional","affiliation":[{"name":"School of Computer Science, The University of Adelaide, Adelaide, SA, Australia"}]},{"given":"Shiqing","family":"Ma","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Rutgers University, New Brunswick, NJ, USA"}]},{"given":"Ehsan","family":"Abbasnejad","sequence":"additional","affiliation":[{"name":"School of Computer Science, The University of Adelaide, Adelaide, SA, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2008-9255","authenticated-orcid":false,"given":"Damith","family":"C. Ranasinghe","sequence":"additional","affiliation":[{"name":"School of Computer Science, The University of Adelaide, Adelaide, SA, Australia"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/s10916-018-1088-1"},{"key":"ref2","first-page":"284","article-title":"Synthesizing robust adversarial examples","volume-title":"Proc. ICML","author":"Athalye"},{"key":"ref3","first-page":"1505","article-title":"Blind backdoors in deep learning models","volume-title":"Proc. USENIX Secur.","author":"Bagdasaryan"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11672"},{"key":"ref5","article-title":"Unrestricted adversarial examples via semantic manipulation","volume-title":"Proc. ICLR","author":"Bhattad"},{"key":"ref6","article-title":"Adversarial patch","volume-title":"Proc. NIPS Workshop","author":"Tom Brown"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW50498.2020.00337"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-56877-1_7"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140444"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.49"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.312"},{"key":"ref12","article-title":"Certified defenses for adversarial patches","volume-title":"Proc. ICLR","author":"Chiang"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/SPW50608.2020.00025"},{"key":"ref14","first-page":"1310","article-title":"Certified adversarial robustness via randomized smoothing","volume-title":"Proc. ICML","author":"Cohen"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/543"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2012.02.016"},{"key":"ref17","first-page":"3578","article-title":"Differentiable abstract interpretation for provably robust neural networks","volume-title":"Proc. ICML","author":"Mirman"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2019.00143"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref20","article-title":"BadNets: Identifying vulnerabilities in the machine learning model supply chain","volume-title":"arXiv:1708.06733","author":"Gu","year":"2017"},{"key":"ref21","article-title":"Approximating CNNs with bag-of-local-features models works surprisingly well on ImageNet","volume-title":"Proc. ICLR","author":"Brendel"},{"key":"ref22","article-title":"Targeted backdoor attacks on deep learning systems using data poisoning","volume-title":"arXiv:1712.05526","author":"Chen","year":"2017"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2018.23291"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00175"},{"key":"ref25","article-title":"Explaining and harnessing adversarial examples","volume-title":"Proc. ICLR","author":"Goodfellow"},{"key":"ref26","first-page":"2672","article-title":"Generative adversarial nets","volume-title":"Proc. NIPS","author":"Goodfellow"},{"key":"ref27","first-page":"5767","article-title":"Improved training of Wasserstein GANs","volume-title":"Proc. NIPS","author":"Gulrajani"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00210"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.123"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01501"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00775"},{"key":"ref32","volume-title":"Google Images","year":"2020"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00257"},{"key":"ref34","first-page":"2507","article-title":"LaVAN: Localized and visible adversarial noise","volume-title":"Proc. ICML","author":"Karmon"},{"key":"ref35","volume-title":"CIFAR-10 (Canadian Institute for Advanced Research)","author":"Krizhevsky"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2021\/694"},{"key":"ref37","first-page":"6465","article-title":"(De)randomized smoothing for certifiable defense against patch attacks","volume-title":"Proc. NeurIPS","author":"Levine"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33011028"},{"key":"ref39","article-title":"No need to worry about adversarial examples in object detection in autonomous vehicles","volume-title":"Proc. CVPR Workshops","author":"Lu"},{"key":"ref40","article-title":"Towards deep learning models resistant to adversarial attacks","volume-title":"Proc. ICLR","author":"Madry"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.17"},{"key":"ref42","volume-title":"Image Thresholding","year":"2020"},{"key":"ref43","article-title":"TorchGAN: A flexible framework for GAN training and Evaluation","volume-title":"arXiv:1909.03410","author":"Pal","year":"2019"},{"key":"ref44","article-title":"Transferability in machine learning: From phenomena to black-box attacks using adversarial samples","volume-title":"arXiv:1605.07277","author":"Papernot","year":"2016"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1145\/3052973.3053009"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.5244\/C.29.41"},{"key":"ref47","first-page":"8026","article-title":"PyTorch: An imperative style, high-performance deep learning library","volume-title":"Proc. NeurIPS","author":"Paszke"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2011.5981788"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-68238-5_32"},{"key":"ref50","first-page":"8178","article-title":"Reverse-engineering deep ReLU networks","volume-title":"Proc. ICML","author":"Rolnick"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978392"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1145\/3317611"},{"key":"ref53","article-title":"Very deep convolutional networks for large-scale image recognition","volume-title":"Proc. ICLR","author":"Simonyan"},{"key":"ref54","article-title":"Intriguing properties of neural networks","volume-title":"arXiv:1312.6199","author":"Szegedy","year":"2013"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.220"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475653"},{"key":"ref57","first-page":"601","article-title":"Stealing machine learning models via prediction APIs","volume-title":"Proc. USENIX Secur.","author":"Tram\u00e8r"},{"key":"ref58","article-title":"On adaptive attacks to adversarial example defenses","volume-title":"arXiv:2002.08347","author":"Tramer","year":"2020"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2019.00031"},{"key":"ref60","article-title":"Defending against physically realizable attacks on image classification","volume-title":"Proc. ICLR","author":"Wu"},{"key":"ref61","first-page":"3177","article-title":"PatchGuard: A provably robust defense against adversarial patches via small receptive fields and masking","volume-title":"Proc. USENIX Secur.","author":"Xiang"},{"key":"ref62","article-title":"Spatially transformed adversarial examples","volume-title":"Proc. ICLR","author":"Xiao"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.5244\/C.30.87"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_40"},{"key":"ref65","first-page":"1659","article-title":"Interpretable deep learning under fire","volume-title":"Proc. USENIX Secur.","author":"Zhang"},{"key":"ref66","article-title":"Generating natural adversarial examples","volume-title":"Proc. ICLR","author":"Zhao"}],"container-title":["IEEE Transactions on Information Forensics and Security"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10206\/9652463\/09856683.pdf?arnumber=9856683","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,24]],"date-time":"2024-01-24T02:58:33Z","timestamp":1706065113000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9856683\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":66,"URL":"https:\/\/doi.org\/10.1109\/tifs.2022.3198857","relation":{},"ISSN":["1556-6013","1556-6021"],"issn-type":[{"value":"1556-6013","type":"print"},{"value":"1556-6021","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]}}}