{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T17:23:20Z","timestamp":1777569800452,"version":"3.51.4"},"publisher-location":"Cham","reference-count":70,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031250651","type":"print"},{"value":"9783031250668","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25066-8_7","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:18:05Z","timestamp":1676621885000},"page":"153-173","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":17,"title":["Realistic Bokeh Effect Rendering on\u00a0Mobile GPUs, Mobile AI &amp; AIM 2022 Challenge: Report"],"prefix":"10.1007","author":[{"given":"Andrey","family":"Ignatov","sequence":"first","affiliation":[]},{"given":"Radu","family":"Timofte","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Feng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Gaocheng","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Zhe","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Hongbin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Minsu","family":"Kwon","sequence":"additional","affiliation":[]},{"given":"Haotian","family":"Qian","sequence":"additional","affiliation":[]},{"given":"Wentao","family":"Tong","sequence":"additional","affiliation":[]},{"given":"Pan","family":"Mu","sequence":"additional","affiliation":[]},{"given":"Ziping","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Guangjing","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Brian","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Fei","sequence":"additional","affiliation":[]},{"given":"Huaijin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Hyebin","family":"Cho","sequence":"additional","affiliation":[]},{"given":"Byeongjun","family":"Kwon","sequence":"additional","affiliation":[]},{"given":"Munchurl","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Mingyang","family":"Qian","sequence":"additional","affiliation":[]},{"given":"Huixin","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Yanan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xiaotao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Lei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"key":"7_CR1","doi-asserted-by":"crossref","unstructured":"Afifi, M., Brubaker, M.A., Brown, M.S.: Histogan: controlling colors of GAN-generated and real images via color histograms. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7941\u20137950 (2021)","DOI":"10.1109\/CVPR46437.2021.00785"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Adams, A., Shih, Y., Hern\u00e1ndez, C.: Fast bilateral-space stereo for synthetic defocus. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4466\u20134474 (2015)","DOI":"10.1109\/CVPR.2015.7299076"},{"key":"7_CR3","unstructured":"Benavides, F.T., Ignatov, A., Timofte, R.: Phonedepth: a dataset for monocular depth estimation on mobile devices. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 0\u20130 (2022)"},{"key":"7_CR4","unstructured":"Cai, J., Gu, S., Timofte, R., Zhang, L.: Ntire 2019 challenge on real image super-resolution: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2019)"},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Cai, Y., Yao, Z., Dong, Z., Gholami, A., Mahoney, M.W., Keutzer, K.: Zeroq: a novel zero shot quantization framework. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13169\u201313178 (2020)","DOI":"10.1109\/CVPR42600.2020.01318"},{"key":"7_CR6","doi-asserted-by":"crossref","unstructured":"Chen, L., Chu, X., Zhang, X., Sun, J.: Simple baselines for image restoration. arXiv preprint arXiv:2204.04676 (2022)","DOI":"10.1007\/978-3-031-20071-7_2"},{"key":"7_CR7","doi-asserted-by":"crossref","unstructured":"Chiang, C.M., et al.: Deploying image deblurring across mobile devices: a perspective of quality and latency. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 502\u2013503 (2020)","DOI":"10.1109\/CVPRW50498.2020.00259"},{"key":"7_CR8","unstructured":"Conde, M.V., Timofte, R., et al.: Reversed Image signal Processing and RAW Reconstruction. AIM 2022 Challenge Report. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)"},{"key":"7_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2021.103089","volume":"77","author":"S Dutta","year":"2021","unstructured":"Dutta, S.: Depth-aware blending of smoothed images for bokeh effect generation. J. Vis. Commun. Image Represent. 77, 103089 (2021)","journal-title":"J. Vis. Commun. Image Represent."},{"key":"7_CR10","doi-asserted-by":"crossref","unstructured":"Dutta, S., Das, S.D., Shah, N.A., Tiwari, A.K.: Stacked deep multi-scale hierarchical network for fast bokeh effect rendering from a single image. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2398\u20132407 (2021)","DOI":"10.1109\/CVPRW53098.2021.00272"},{"key":"7_CR11","unstructured":"in the new Google Camera app, L.B.: https:\/\/ai.googleblog.com\/2014\/04\/lens-blur-in-new-google-camera-app.html"},{"key":"7_CR12","doi-asserted-by":"crossref","unstructured":"Ha, H., Im, S., Park, J., Jeon, H.G., So Kweon, I.: High-quality depth from uncalibrated small motion clip. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5413\u20135421 (2016)","DOI":"10.1109\/CVPR.2016.584"},{"key":"7_CR13","doi-asserted-by":"crossref","unstructured":"Howard, A., et al.: Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"7_CR14","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Byeoung-su, K., Timofte, R.: Fast camera image denoising on mobile GPUs with deep learning, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00285"},{"key":"7_CR15","unstructured":"Ignatov, A., Chiang, J., Kuo, H.K., Sycheva, A., Timofte, R.: Learned smartphone ISP on mobile NPUs with deep learning, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)"},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Kobyshev, N., Timofte, R., Vanhoey, K., Van Gool, L.: Dslr-quality photos on mobile devices with deep convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3277\u20133285 (2017)","DOI":"10.1109\/ICCV.2017.355"},{"key":"7_CR17","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Malivenko, G., Plowman, D., Shukla, S., Timofte, R.: Fast and accurate single-image depth estimation on mobile devices, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00288"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Malivenko, G., Timofte, R.: Fast and accurate quantized camera scene detection on smartphones, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00289"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Pynet-v2 mobile: efficient on-device photo processing with neural networks. In: 2021 26th International Conference on Pattern Recognition (ICPR). IEEE (2022)","DOI":"10.1109\/ICPR56361.2022.9956598"},{"key":"7_CR20","unstructured":"Ignatov, A., Malivenko, G., Timofte, R., et al.: Efficient single-image depth estimation on mobile devices, mobile AI & AIM 2022 challenge: Report. In: European Conference on Computer Vision (2022)"},{"key":"7_CR21","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Patel, J., Timofte, R.: Rendering natural camera bokeh effect with deep learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 418\u2013419 (2020)","DOI":"10.1109\/CVPRW50498.2020.00217"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Aim 2019 challenge on bokeh effect synthesis: methods and results. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3591\u20133598. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00444"},{"key":"7_CR23","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Microisp: Processing 32mp photos on mobile devices with deep learning. In: European Conference on Computer Vision (2022)","DOI":"10.1007\/978-3-031-25063-7_46"},{"key":"7_CR24","unstructured":"Ignatov, A., Timofte, R.: Ntire 2019 challenge on image enhancement: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Power efficient video super-resolution on mobile NPUs with deep learning, mobile AI & aim 2022 challenge: Report. In: European Conference on Computer Vision (2022)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"7_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/978-3-030-11021-5_19","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Ignatov","year":"2019","unstructured":"Ignatov, A., et al.: AI benchmark: running deep neural networks on android smartphones. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 288\u2013314. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_19"},{"key":"7_CR27","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Timofte, R., Denna, M., Younes, A.: Real-time quantized image super-resolution on mobile NPUs, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"7_CR28","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Timofte, R., Denna, M., Younes, A., et al.: Efficient and accurate quantized image super-resolution on mobile NPUs, mobile AI & AIM 2022 challenge: Report. In: European Conference on Computer Vision (2022)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Aim 2019 challenge on raw to RGB mapping: methods and results. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3584\u20133590. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00443"},{"key":"7_CR30","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Ai benchmark: All about deep learning on smartphones in 2019. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3617\u20133635. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00447"},{"key":"7_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-67070-2_13","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"A Ignatov","year":"2020","unstructured":"Ignatov, A., et al.: AIM 2020 challenge on rendering realistic bokeh. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12537, pp. 213\u2013228. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-67070-2_13"},{"key":"7_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"315","DOI":"10.1007\/978-3-030-11021-5_20","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Ignatov","year":"2019","unstructured":"Ignatov, A., et al.: PIRM challenge on perceptual image enhancement on smartphones: report. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 315\u2013333. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_20"},{"key":"7_CR33","unstructured":"Ignatov, A., et al.: Aim 2020 challenge on learned image signal processing pipeline. arXiv preprint arXiv:2011.04994 (2020)"},{"key":"7_CR34","unstructured":"Ignatov, A., Timofte, R., et al.: Learned smartphone ISP on mobile GPUs with deep learning, mobile AI & aim 2022 challenge: Report. In: European Conference on Computer Vision (2022)"},{"key":"7_CR35","doi-asserted-by":"publisher","first-page":"276","DOI":"10.1016\/j.patrec.2020.07.033","volume":"138","author":"D Ignatov","year":"2020","unstructured":"Ignatov, D., Ignatov, A.: Controlling information capacity of binary neural network. Pattern Recogn. Lett. 138, 276\u2013281 (2020)","journal-title":"Pattern Recogn. Lett."},{"key":"7_CR36","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"7_CR37","doi-asserted-by":"crossref","unstructured":"Jacob, B., et al.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"7_CR38","unstructured":"Jain, S.R., Gural, A., Wu, M., Dick, C.H.: trained quantization thresholds for accurate and efficient fixed-point inference of deep neural networks. arXiv preprint arXiv:1903.08066 (2019)"},{"key":"7_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"7_CR40","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"7_CR41","doi-asserted-by":"crossref","unstructured":"K\u0131nl\u0131, F.O., Mente\u015f, S., \u00d6zcan, B., Kirac, F., Timofte, R., et al.: Aim 2022 challenge on instagram filter removal: Methods and results. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)","DOI":"10.1007\/978-3-031-25066-8_2"},{"key":"7_CR42","doi-asserted-by":"crossref","unstructured":"Lee, B.J., Lei, F., Chen, H., Baudron, A.: Bokeh-loss GAN: multi-stage adversarial training for realistic edge-aware bokeh. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)","DOI":"10.1007\/978-3-031-25063-7_39"},{"key":"7_CR43","unstructured":"Lee, J., et al.: On-device neural net inference with mobile GPUs. arXiv preprint arXiv:1907.01989 (2019)"},{"key":"7_CR44","doi-asserted-by":"crossref","unstructured":"Li, Y., Gu, S., Gool, L.V., Timofte, R.: Learning filter basis for convolutional neural network compression. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5623\u20135632 (2019)","DOI":"10.1109\/ICCV.2019.00572"},{"key":"7_CR45","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Metapruning: Meta learning for automatic neural network channel pruning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3296\u20133305 (2019)","DOI":"10.1109\/ICCV.2019.00339"},{"key":"7_CR46","doi-asserted-by":"crossref","unstructured":"Liu, Z., Wu, B., Luo, W., Yang, X., Liu, W., Cheng, K.T.: Bi-real net: Enhancing the performance of 1-bit cnns with improved representational capability and advanced training algorithm. In: Proceedings of the European conference on computer vision (ECCV), pp. 722\u2013737 (2018)","DOI":"10.1007\/978-3-030-01267-0_44"},{"key":"7_CR47","unstructured":"Lugmayr, A., Danelljan, M., Timofte, R.: Ntire 2020 challenge on real-world image super-resolution: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 494\u2013495 (2020)"},{"key":"7_CR48","unstructured":"Mehta, S., Rastegari, M.: Separable self-attention for mobile vision transformers. arXiv preprint arXiv:2206.02680 (2022)"},{"key":"7_CR49","unstructured":"Obukhov, A., Rakhuba, M., Georgoulis, S., Kanakis, M., Dai, D., Van Gool, L.: T-basis: a compact representation for neural networks. In: International Conference on Machine Learning, pp. 7392\u20137404. PMLR (2020)"},{"key":"7_CR50","doi-asserted-by":"crossref","unstructured":"Peng, J., Cao, Z., Luo, X., Lu, H., Xian, K., Zhang, J.: Bokehme: when neural rendering meets classical rendering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16283\u201316292 (2022)","DOI":"10.1109\/CVPR52688.2022.01580"},{"key":"7_CR51","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1007\/978-3-030-67070-2_14","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"M Qian","year":"2020","unstructured":"Qian, M., et al.: BGGAN: bokeh-glass generative adversarial network for rendering realistic bokeh. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12537, pp. 229\u2013244. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-67070-2_14"},{"key":"7_CR52","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Bochkovskiy, A., Koltun, V.: Vision transformers for dense prediction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12179\u201312188 (2021)","DOI":"10.1109\/ICCV48922.2021.01196"},{"key":"7_CR53","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. 44, 1623\u20131637 (2020)","DOI":"10.1109\/TPAMI.2020.3019967"},{"key":"7_CR54","doi-asserted-by":"crossref","unstructured":"Romero, A., Ignatov, A., Kim, H., Timofte, R.: Real-time video super-resolution on smartphones with deep learning, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00287"},{"key":"7_CR55","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: Mobilenetv 2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"7_CR56","doi-asserted-by":"crossref","unstructured":"Shen, X., et al.: Automatic portrait segmentation for image stylization. In: Computer Graphics Forum, vol. 35, pp. 93\u2013102. Wiley Online Library (2016)","DOI":"10.1111\/cgf.12814"},{"key":"7_CR57","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"92","DOI":"10.1007\/978-3-319-46448-0_6","volume-title":"Computer Vision \u2013 ECCV 2016","author":"X Shen","year":"2016","unstructured":"Shen, X., Tao, X., Gao, H., Zhou, C., Jia, J.: Deep automatic portrait matting. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 92\u2013107. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_6"},{"key":"7_CR58","doi-asserted-by":"crossref","unstructured":"Tan, M., et al.: MnasNet: platform-aware neural architecture search for mobile. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2820\u20132828 (2019)","DOI":"10.1109\/CVPR.2019.00293"},{"key":"7_CR59","unstructured":"TensorFlow-Lite: https:\/\/www.tensorflow.org\/lite"},{"key":"7_CR60","unstructured":"Timofte, R., Gu, S., Wu, J., Van Gool, L.: Ntire 2018 challenge on single image super-resolution: methods and results. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 852\u2013863 (2018)"},{"key":"7_CR61","unstructured":"Uhlich, S., et al.: Mixed precision DNNs: all you need is a good parametrization. arXiv preprint arXiv:1905.11452 (2019)"},{"key":"7_CR62","doi-asserted-by":"crossref","unstructured":"Wan, A., et al.: Fbnetv2: differentiable neural architecture search for spatial and channel dimensions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12965\u201312974 (2020)","DOI":"10.1109\/CVPR42600.2020.01298"},{"issue":"5","key":"7_CR63","doi-asserted-by":"publisher","first-page":"286","DOI":"10.3390\/machines10050286","volume":"10","author":"F Wang","year":"2022","unstructured":"Wang, F., Zhang, Y., Ai, Y., Zhang, W.: Rendering natural bokeh effects based on depth estimation to improve the aesthetic ability of machine vision. Machines 10(5), 286 (2022)","journal-title":"Machines"},{"key":"7_CR64","doi-asserted-by":"crossref","unstructured":"Wu, B., et al.: Fbnet: hardware-aware efficient convnet design via differentiable neural architecture search. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10734\u201310742 (2019)","DOI":"10.1109\/CVPR.2019.01099"},{"key":"7_CR65","doi-asserted-by":"crossref","unstructured":"Yang, J., et al.: Quantization networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7308\u20137316 (2019)","DOI":"10.1109\/CVPR.2019.00748"},{"key":"7_CR66","unstructured":"Yang, R., Timofte, R., et al.: Aim 2022 challenge on super-resolution of compressed image and video: Dataset, methods and results. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)"},{"key":"7_CR67","doi-asserted-by":"crossref","unstructured":"Yu, F., Gallup, D.: 3D reconstruction from accidental motion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3986\u20133993 (2014)","DOI":"10.1109\/CVPR.2014.509"},{"key":"7_CR68","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"7_CR69","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1109\/TCI.2022.3171417","volume":"8","author":"B Zheng","year":"2022","unstructured":"Zheng, B., et al.: Constrained predictive filters for single image bokeh rendering. IEEE Trans. Comput. Imaging 8, 346\u2013357 (2022)","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"7_CR70","doi-asserted-by":"crossref","unstructured":"Zhu, B., Chen, Y., Wang, J., Liu, S., Zhang, B., Tang, M.: Fast deep matting for portrait animation on mobile phone. In: Proceedings of the 25th ACM International Conference on Multimedia, pp. 297\u2013305 (2017)","DOI":"10.1145\/3123266.3123286"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25066-8_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T13:13:34Z","timestamp":1728911614000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25066-8_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250651","9783031250668"],"references-count":70,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25066-8_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}