{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T18:58:47Z","timestamp":1778180327408,"version":"3.51.4"},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,6,1]],"date-time":"2020-06-01T00:00:00Z","timestamp":1590969600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,6,1]],"date-time":"2020-06-01T00:00:00Z","timestamp":1590969600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,6,1]],"date-time":"2020-06-01T00:00:00Z","timestamp":1590969600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,6]]},"DOI":"10.1109\/cvpr42600.2020.00126","type":"proceedings-article","created":{"date-parts":[[2020,8,5]],"date-time":"2020-08-05T21:20:05Z","timestamp":1596662405000},"page":"1178-1187","source":"Crossref","is-referenced-by-count":85,"title":["Efficient Adversarial Training With Transferable Adversarial Examples"],"prefix":"10.1109","author":[{"given":"Haizhong","family":"Zheng","sequence":"first","affiliation":[]},{"given":"Ziqi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Juncheng","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Honglak","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Atul","family":"Prakash","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","article-title":"The limitations of adversarial training and the blind-spot attack","author":"zhang","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref32","article-title":"You only propagate once: Accelerating adversarial training via maximal principle","author":"zhang","year":"0","journal-title":"Neural Information Processing Systems (NeurIPS)"},{"key":"ref31","author":"zagoruyko","year":"2016","journal-title":"Wide residual networks"},{"key":"ref30","article-title":"Provable defenses against adversarial examples via the convex outer adversarial polytope","author":"wong","year":"0","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref34","article-title":"Theoretically principled trade-off between robustness and accuracy","author":"zhang","year":"0","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref10","article-title":"Using pre-training can improve model robustness and uncertainty","author":"hendrycks","year":"0","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00283"},{"key":"ref12","article-title":"Learning multiple layers of features from tiny images","author":"krizhevsky","year":"2009","journal-title":"Technical Report Citeseer"},{"key":"ref13","article-title":"Adversarial machine learning at scale","author":"kurakin","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref15","author":"lee","year":"2017","journal-title":"Generative adversarial trainer Defense to adversarial perturbations with GAN"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6810"},{"key":"ref17","article-title":"Delving into transferable adversarial examples and black-box attacks","author":"liu","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref18","article-title":"Towards deep learning models resistant to adversarial attacks","author":"madry","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref19","author":"papernot","year":"2016","journal-title":"Transferability in machine learning from phenomena to black-box attacks using adversarial samples"},{"key":"ref28","article-title":"Ensemble adversarial training: Attacks and defenses","author":"tram\u00e8r","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref4","article-title":"Unlabeled data improves adversarial robustness","author":"carmon","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref27","article-title":"Intriguing properties of neural networks","author":"szegedy","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140444"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00673"},{"key":"ref5","article-title":"Certified adversarial robustness via randomized smoothing","author":"cohen","year":"0","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref8","article-title":"Explaining and harnessing adversarial examples","author":"goodfellow","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref7","article-title":"Robust physical-world attacks on deep learning models","author":"eykholt","year":"0","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2018\/520"},{"key":"ref9","first-page":"2266","article-title":"Formal guarantees on the robustness of a classifier against adversarial manipulation","author":"hein","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref1","first-page":"2687","article-title":"Learning to attack: Adversarial transformation networks","author":"baluja","year":"0","journal-title":"AAAI"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3052973.3053009"},{"key":"ref22","first-page":"5014","article-title":"Adversarially robust generalization requires more data","author":"schmidt","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref21","article-title":"Certified defenses against adversarial examples","author":"raghunathan","year":"0","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref24","author":"shafahi","year":"2018","journal-title":"Universal adversarial training"},{"key":"ref23","article-title":"Adversarial training for free!","author":"shafahi","year":"0","journal-title":"Neural Information Processing Systems (NeurIPS)"},{"key":"ref26","article-title":"Physical adversarial examples for object detectors","author":"song","year":"0","journal-title":"12th USENIX Workshop on Offensive Technologies (WOOT 18)"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978392"}],"event":{"name":"2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","location":"Seattle, WA, USA","start":{"date-parts":[[2020,6,13]]},"end":{"date-parts":[[2020,6,19]]}},"container-title":["2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9142308\/9156271\/09157681.pdf?arnumber=9157681","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,27]],"date-time":"2022-06-27T15:44:53Z","timestamp":1656344693000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9157681\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,6]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/cvpr42600.2020.00126","relation":{},"subject":[],"published":{"date-parts":[[2020,6]]}}}