{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,7]],"date-time":"2025-08-07T09:15:56Z","timestamp":1754558156496},"publisher-location":"California","reference-count":0,"publisher":"International Joint Conferences on Artificial Intelligence Organization","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,7]]},"abstract":"<jats:p>Despite the success of vision Transformers for the image deraining task, they are limited by computation-heavy and slow runtime. In this work, we investigate Transformer decoder is not necessary and has huge computational costs. Therefore, we revisit the standard vision Transformer as well as its successful variants and propose a novel Decoder-Free Transformer-Like (DFTL) architecture for fast and accurate single image deraining. Specifically, we adopt a cheap linear projection to represent visual information with lower\n\ncomputational costs than previous linear projections. Then we replace standard Transformer decoder block with designed Progressive Patch Merging (PPM), which attains comparable performance and efficiency. DFTL could significantly alleviate the computation and GPU memory requirements through proposed modules. Extensive experiments\n\ndemonstrate the superiority of DFTL compared with competitive Transformer architectures, e.g., ViT, DETR, IPT, Uformer, and Restormer. The code is available at https:\/\/github.com\/XiaoXiao-Woo\/derain.<\/jats:p>","DOI":"10.24963\/ijcai.2022\/205","type":"proceedings-article","created":{"date-parts":[[2022,7,16]],"date-time":"2022-07-16T02:55:56Z","timestamp":1657940156000},"page":"1474-1480","source":"Crossref","is-referenced-by-count":6,"title":["A Decoder-free Transformer-like Architecture for High-efficiency Single Image Deraining"],"prefix":"10.24963","author":[{"given":"Xiao","family":"Wu","sequence":"first","affiliation":[{"name":"University of Electronic Science and Technology of China"}]},{"given":"Ting-Zhu","family":"Huang","sequence":"additional","affiliation":[{"name":"University of Electronic Science and Technology of China"}]},{"given":"Liang-Jian","family":"Deng","sequence":"additional","affiliation":[{"name":"University of Electronic Science and Technology of China"}]},{"given":"Tian-Jing","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Electronic Science and Technology of China"}]}],"member":"10584","event":{"number":"31","sponsor":["International Joint Conferences on Artificial Intelligence Organization (IJCAI)"],"acronym":"IJCAI-2022","name":"Thirty-First International Joint Conference on Artificial Intelligence {IJCAI-22}","start":{"date-parts":[[2022,7,23]]},"theme":"Artificial Intelligence","location":"Vienna, Austria","end":{"date-parts":[[2022,7,29]]}},"container-title":["Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence"],"original-title":[],"deposited":{"date-parts":[[2022,7,18]],"date-time":"2022-07-18T11:08:20Z","timestamp":1658142500000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.ijcai.org\/proceedings\/2022\/205"}},"subtitle":[],"proceedings-subject":"Artificial Intelligence Research Articles","short-title":[],"issued":{"date-parts":[[2022,7]]},"references-count":0,"URL":"https:\/\/doi.org\/10.24963\/ijcai.2022\/205","relation":{},"subject":[],"published":{"date-parts":[[2022,7]]}}}