{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T09:20:53Z","timestamp":1778145653299,"version":"3.51.4"},"publisher-location":"Cham","reference-count":61,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197895","type":"print"},{"value":"9783031197901","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19790-1_40","type":"book-chapter","created":{"date-parts":[[2022,10,23]],"date-time":"2022-10-23T11:02:44Z","timestamp":1666522964000},"page":"668-685","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":279,"title":["FlowFormer: A Transformer Architecture for\u00a0Optical Flow"],"prefix":"10.1007","author":[{"given":"Zhaoyang","family":"Huang","sequence":"first","affiliation":[]},{"given":"Xiaoyu","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Chao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ka Chun","family":"Cheung","sequence":"additional","affiliation":[]},{"given":"Hongwei","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Jifeng","family":"Dai","sequence":"additional","affiliation":[]},{"given":"Hongsheng","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,24]]},"reference":[{"key":"40_CR1","doi-asserted-by":"crossref","unstructured":"Black, M.J., Anandan, P.: A framework for the robust estimation of optical flow. In: 1993 (4th) International Conference on Computer Vision, pp. 231\u2013236. IEEE (1993)","DOI":"10.1109\/ICCV.1993.378214"},{"issue":"3","key":"40_CR2","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1023\/B:VISI.0000045324.43199.43","volume":"61","author":"A Bruhn","year":"2005","unstructured":"Bruhn, A., Weickert, J., Schn\u00f6rr, C.: Lucas\/Kanade meets Horn\/Schunck: combining local and global optic flow methods. Int. J. Comput. Vis. 61(3), 211\u2013231 (2005)","journal-title":"Int. J. Comput. Vis."},{"key":"40_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1007\/978-3-642-33783-3_44","volume-title":"Computer Vision \u2013 ECCV 2012","author":"DJ Butler","year":"2012","unstructured":"Butler, D.J., Wulff, J., Stanley, G.B., Black, M.J.: A naturalistic open source movie for optical flow evaluation. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7577, pp. 611\u2013625. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33783-3_44"},{"key":"40_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision \u2013 ECCV 2020","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"40_CR5","doi-asserted-by":"crossref","unstructured":"Chan, K.C., Wang, X., Yu, K., Dong, C., Loy, C.C.: BasicVSR: the search for essential components in video super-resolution and beyond. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4947\u20134956 (2021)","DOI":"10.1109\/CVPR46437.2021.00491"},{"key":"40_CR6","doi-asserted-by":"crossref","unstructured":"Chen, H., et al.: Pre-trained image processing transformer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12299\u201312310 (2021)","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"40_CR7","unstructured":"Cho, S., Hong, S., Jeon, S., Lee, Y., Sohn, K., Kim, S.: CATs: cost aggregation transformers for visual correspondence. In: Advances in Neural Information Processing Systems, vol. 34 (2021)"},{"key":"40_CR8","unstructured":"Chu, X., et al.: Twins: revisiting spatial attention design in vision transformers. arXiv preprint arXiv:2104.13840 (2021)"},{"key":"40_CR9","doi-asserted-by":"crossref","unstructured":"Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.V., Salakhutdinov, R.: Transformer-XL: attentive language models beyond a fixed-length context. arXiv preprint arXiv:1901.02860 (2019)","DOI":"10.18653\/v1\/P19-1285"},{"key":"40_CR10","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"40_CR11","unstructured":"Dosovitskiy, A., et al.: An image is worth $$16 \\times 16$$ words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"40_CR12","doi-asserted-by":"crossref","unstructured":"Dosovitskiy, A., et al.: FlowNet: learning optical flow with convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2758\u20132766 (2015)","DOI":"10.1109\/ICCV.2015.316"},{"key":"40_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"713","DOI":"10.1007\/978-3-030-58610-2_42","volume-title":"Computer Vision \u2013 ECCV 2020","author":"C Gao","year":"2020","unstructured":"Gao, C., Saraf, A., Huang, J.-B., Kopf, J.: Flow-edge guided video completion. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 713\u2013729. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_42"},{"issue":"11","key":"40_CR14","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the KITTI dataset. Int. J. Robot. Res. 32(11), 1231\u20131237 (2013)","journal-title":"Int. J. Robot. Res."},{"issue":"2","key":"40_CR15","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1007\/s41095-021-0229-5","volume":"7","author":"MH Guo","year":"2021","unstructured":"Guo, M.H., Cai, J.X., Liu, Z.N., Mu, T.J., Martin, R.R., Hu, S.M.: PCT: point cloud transformer. Comput. Vis. Media 7(2), 187\u2013199 (2021)","journal-title":"Comput. Vis. Media"},{"key":"40_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"770","DOI":"10.1007\/978-3-030-58604-1_46","volume-title":"Computer Vision \u2013 ECCV 2020","author":"M Hofinger","year":"2020","unstructured":"Hofinger, M., Bul\u00f2, S.R., Porzi, L., Knapitsch, A., Pock, T., Kontschieder, P.: Improving optical flow on a pyramid level. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12373, pp. 770\u2013786. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58604-1_46"},{"issue":"1\u20133","key":"40_CR17","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/0004-3702(81)90024-2","volume":"17","author":"BK Horn","year":"1981","unstructured":"Horn, B.K., Schunck, B.G.: Determining optical flow. Artif. Intell. 17(1\u20133), 185\u2013203 (1981)","journal-title":"Artif. Intell."},{"key":"40_CR18","unstructured":"Huang, Z., et al.: Life: lighting invariant flow estimation. arXiv preprint arXiv:2104.03097 (2021)"},{"key":"40_CR19","doi-asserted-by":"crossref","unstructured":"Huang, Z., et al.: VS-Net: voting with segmentation for visual localization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6101\u20136111 (2021)","DOI":"10.1109\/CVPR46437.2021.00604"},{"key":"40_CR20","unstructured":"Huang, Z., Zhang, T., Heng, W., Shi, B., Zhou, S.: RIFE: real-time intermediate flow estimation for video frame interpolation. arXiv preprint arXiv:2011.06294 (2020)"},{"key":"40_CR21","doi-asserted-by":"crossref","unstructured":"Hui, T.W., Tang, X., Loy, C.C.: LiteFlowNet: a lightweight convolutional neural network for optical flow estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8981\u20138989 (2018)","DOI":"10.1109\/CVPR.2018.00936"},{"issue":"8","key":"40_CR22","doi-asserted-by":"publisher","first-page":"2555","DOI":"10.1109\/TPAMI.2020.2976928","volume":"43","author":"TW Hui","year":"2020","unstructured":"Hui, T.W., Tang, X., Loy, C.C.: A lightweight optical flow CNN-revisiting data fidelity and regularization. IEEE Trans. Pattern Anal. Mach. Intell. 43(8), 2555\u20132569 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"40_CR23","doi-asserted-by":"crossref","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: FlowNet 2.0: evolution of optical flow estimation with deep networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2462\u20132470 (2017)","DOI":"10.1109\/CVPR.2017.179"},{"key":"40_CR24","unstructured":"Jaegle, A., et al.: Perceiver IO: a general architecture for structured inputs & outputs. arXiv preprint arXiv:2107.14795 (2021)"},{"key":"40_CR25","doi-asserted-by":"crossref","unstructured":"Jiang, S., Campbell, D., Lu, Y., Li, H., Hartley, R.: Learning to estimate hidden motions with global motion aggregation. arXiv preprint arXiv:2104.02409 (2021)","DOI":"10.1109\/ICCV48922.2021.00963"},{"key":"40_CR26","doi-asserted-by":"crossref","unstructured":"Jiang, S., Lu, Y., Li, H., Hartley, R.: Learning optical flow from a few matches. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16592\u201316600 (2021)","DOI":"10.1109\/CVPR46437.2021.01632"},{"key":"40_CR27","doi-asserted-by":"crossref","unstructured":"Jiang, W., Trulls, E., Hosang, J., Tagliasacchi, A., Yi, K.M.: COTR: correspondence transformer for matching across images. arXiv preprint arXiv:2103.14167 (2021)","DOI":"10.1109\/ICCV48922.2021.00615"},{"key":"40_CR28","doi-asserted-by":"crossref","unstructured":"Kim, D., Woo, S., Lee, J.Y., Kweon, I.S.: Deep video inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5792\u20135801 (2019)","DOI":"10.1109\/CVPR.2019.00594"},{"key":"40_CR29","doi-asserted-by":"crossref","unstructured":"Kondermann, D., et al.: The HCI benchmark suite: stereo and flow ground truth with uncertainties for urban autonomous driving. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 19\u201328 (2016)","DOI":"10.1109\/CVPRW.2016.10"},{"key":"40_CR30","doi-asserted-by":"crossref","unstructured":"Lai, W.S., Huang, J.B., Ahuja, N., Yang, M.H.: Deep Laplacian pyramid networks for fast and accurate super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 624\u2013632 (2017)","DOI":"10.1109\/CVPR.2017.618"},{"key":"40_CR31","doi-asserted-by":"crossref","unstructured":"Liang, J., Cao, J., Sun, G., Zhang, K., Van Gool, L., Timofte, R.: SwinIR: image restoration using Swin transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1833\u20131844 (2021)","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"40_CR32","doi-asserted-by":"crossref","unstructured":"Liu, R., et al.: FuseFormer: fusing fine-grained information in transformers for video inpainting. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14040\u201314049 (2021)","DOI":"10.1109\/ICCV48922.2021.01378"},{"issue":"12","key":"40_CR33","doi-asserted-by":"publisher","first-page":"2087","DOI":"10.1002\/int.22285","volume":"35","author":"X Liu","year":"2020","unstructured":"Liu, X., Liu, H., Lin, Y.: Video frame interpolation via optical flow estimation with image inpainting. Int. J. Intell. Syst. 35(12), 2087\u20132102 (2020)","journal-title":"Int. J. Intell. Syst."},{"key":"40_CR34","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. arXiv preprint arXiv:2103.14030 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"40_CR35","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4040\u20134048 (2016)","DOI":"10.1109\/CVPR.2016.438"},{"key":"40_CR36","doi-asserted-by":"crossref","unstructured":"Piergiovanni, A., Ryoo, M.S.: Representation flow for action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9945\u20139953 (2019)","DOI":"10.1109\/CVPR.2019.01018"},{"key":"40_CR37","doi-asserted-by":"crossref","unstructured":"Ranjan, A., Black, M.J.: Optical flow estimation using a spatial pyramid network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4161\u20134170 (2017)","DOI":"10.1109\/CVPR.2017.291"},{"key":"40_CR38","doi-asserted-by":"crossref","unstructured":"Sajjadi, M.S., Vemulapalli, R., Brown, M.: Frame-recurrent video super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6626\u20136634 (2018)","DOI":"10.1109\/CVPR.2018.00693"},{"key":"40_CR39","doi-asserted-by":"crossref","unstructured":"Sarlin, P.E., DeTone, D., Malisiewicz, T., Rabinovich, A.: SuperGlue: learning feature matching with graph neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4938\u20134947 (2020)","DOI":"10.1109\/CVPR42600.2020.00499"},{"issue":"2","key":"40_CR40","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1007\/s11263-013-0644-x","volume":"106","author":"D Sun","year":"2014","unstructured":"Sun, D., Roth, S., Black, M.J.: A quantitative analysis of current practices in optical flow estimation and the principles behind them. Int. J. Comput. Vis. 106(2), 115\u2013137 (2014)","journal-title":"Int. J. Comput. Vis."},{"key":"40_CR41","doi-asserted-by":"crossref","unstructured":"Sun, D., et al.: AutoFlow: learning a better training set for optical flow. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10093\u201310102 (2021)","DOI":"10.1109\/CVPR46437.2021.00996"},{"key":"40_CR42","doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M.Y., Kautz, J.: PWC-Net: CNNs for optical flow using pyramid, warping, and cost volume. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8934\u20138943 (2018)","DOI":"10.1109\/CVPR.2018.00931"},{"issue":"6","key":"40_CR43","doi-asserted-by":"publisher","first-page":"1408","DOI":"10.1109\/TPAMI.2019.2894353","volume":"42","author":"D Sun","year":"2019","unstructured":"Sun, D., Yang, X., Liu, M.Y., Kautz, J.: Models matter, so does training: an empirical study of CNNs for optical flow estimation. IEEE Trans. Pattern Anal. Mach. Intell. 42(6), 1408\u20131423 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"40_CR44","doi-asserted-by":"crossref","unstructured":"Sun, J., Shen, Z., Wang, Y., Bao, H., Zhou, X.: LoFTR: detector-free local feature matching with transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8922\u20138931 (2021)","DOI":"10.1109\/CVPR46437.2021.00881"},{"key":"40_CR45","doi-asserted-by":"crossref","unstructured":"Sun, S., Kuang, Z., Sheng, L., Ouyang, W., Zhang, W.: Optical flow guided feature: a fast and robust motion representation for video action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1390\u20131399 (2018)","DOI":"10.1109\/CVPR.2018.00151"},{"key":"40_CR46","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-030-58536-5_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Teed","year":"2020","unstructured":"Teed, Z., Deng, J.: RAFT: recurrent all-pairs field transforms for optical flow. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 402\u2013419. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_24"},{"key":"40_CR47","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"40_CR48","doi-asserted-by":"crossref","unstructured":"Xu, H., Yang, J., Cai, J., Zhang, J., Tong, X.: High-resolution optical flow from 1D attention and correlation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10498\u201310507 (2021)","DOI":"10.1109\/ICCV48922.2021.01033"},{"key":"40_CR49","doi-asserted-by":"crossref","unstructured":"Xu, R., Li, X., Zhou, B., Loy, C.C.: Deep flow-guided video inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3723\u20133732 (2019)","DOI":"10.1109\/CVPR.2019.00384"},{"key":"40_CR50","unstructured":"Xu, X., Siyao, L., Sun, W., Yin, Q., Yang, M.H.: Quadratic video interpolation. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"40_CR51","doi-asserted-by":"crossref","unstructured":"Xu, Y., Lin, K.Y., Zhang, G., Wang, X., Li, H.: RNNPose: recurrent 6-DoF object pose refinement with robust correspondence field estimation and pose optimization (2022)","DOI":"10.1109\/CVPR52688.2022.01446"},{"key":"40_CR52","doi-asserted-by":"crossref","unstructured":"Yan, W., Sharma, A., Tan, R.T.: Optical flow in dense foggy scenes using semi-supervised learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13259\u201313268 (2020)","DOI":"10.1109\/CVPR42600.2020.01327"},{"key":"40_CR53","unstructured":"Yang, G., Ramanan, D.: Volumetric correspondence networks for optical flow. In: Advances in Neural Information Processing Systems, vol. 32, pp. 794\u2013805 (2019)"},{"key":"40_CR54","doi-asserted-by":"crossref","unstructured":"Yang, L., Xu, Y., Yuan, C., Liu, W., Li, B., Hu, W.: Improving visual grounding with visual-linguistic verification and iterative reasoning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9499\u20139508 (2022)","DOI":"10.1109\/CVPR52688.2022.00928"},{"key":"40_CR55","doi-asserted-by":"crossref","unstructured":"Yin, Z., Darrell, T., Yu, F.: Hierarchical discrete distribution decomposition for match density estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6044\u20136053 (2019)","DOI":"10.1109\/CVPR.2019.00620"},{"key":"40_CR56","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"528","DOI":"10.1007\/978-3-030-58517-4_31","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Zeng","year":"2020","unstructured":"Zeng, Y., Fu, J., Chao, H.: Learning joint spatial-temporal transformations for video inpainting. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12361, pp. 528\u2013543. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58517-4_31"},{"key":"40_CR57","doi-asserted-by":"crossref","unstructured":"Zhang, F., Woodford, O.J., Prisacariu, V.A., Torr, P.H.: Separable flow: learning motion cost volumes for optical flow estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10807\u201310817 (2021)","DOI":"10.1109\/ICCV48922.2021.01063"},{"key":"40_CR58","doi-asserted-by":"crossref","unstructured":"Zhao, H., Jiang, L., Jia, J., Torr, P.H., Koltun, V.: Point transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 16259\u201316268 (2021)","DOI":"10.1109\/ICCV48922.2021.01595"},{"key":"40_CR59","doi-asserted-by":"crossref","unstructured":"Zhao, S., Sheng, Y., Dong, Y., Chang, E.I., Xu, Y., et al.: MaskFlowNet: asymmetric feature matching with learnable occlusion mask. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6278\u20136287 (2020)","DOI":"10.1109\/CVPR42600.2020.00631"},{"issue":"1","key":"40_CR60","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s13640-020-00501-x","volume":"2020","author":"Y Zhao","year":"2020","unstructured":"Zhao, Y., Man, K.L., Smith, J., Siddique, K., Guan, S.U.: Improved two-stream model for human action recognition. EURASIP J. Image Video Process. 2020(1), 1\u20139 (2020)","journal-title":"EURASIP J. Image Video Process."},{"key":"40_CR61","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Zhang, M., Lu, F.: Optical flow in the dark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6749\u20136757 (2020)","DOI":"10.1109\/CVPR42600.2020.00678"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19790-1_40","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T10:53:33Z","timestamp":1728212013000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19790-1_40"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197895","9783031197901"],"references-count":61,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19790-1_40","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"24 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}