{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,26]],"date-time":"2026-04-26T12:35:04Z","timestamp":1777206904176,"version":"3.51.4"},"reference-count":22,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,9,1]],"date-time":"2025-09-01T00:00:00Z","timestamp":1756684800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,9,1]],"date-time":"2025-09-01T00:00:00Z","timestamp":1756684800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,8,5]],"date-time":"2025-08-05T00:00:00Z","timestamp":1754352000000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100012325","name":"National Office for Philosophy and Social Sciences","doi-asserted-by":"publisher","award":["22ZD10"],"award-info":[{"award-number":["22ZD10"]}],"id":[{"id":"10.13039\/501100012325","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012456","name":"National Social Science Fund of China","doi-asserted-by":"publisher","award":["20ZD313"],"award-info":[{"award-number":["20ZD313"]}],"id":[{"id":"10.13039\/501100012456","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013071","name":"Major Program of National Fund of Philosophy and Social Science of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100013071","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["CUC25GT05"],"award-info":[{"award-number":["CUC25GT05"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["SoftwareX"],"published-print":{"date-parts":[[2025,9]]},"DOI":"10.1016\/j.softx.2025.102299","type":"journal-article","created":{"date-parts":[[2025,8,18]],"date-time":"2025-08-18T09:18:15Z","timestamp":1755508695000},"page":"102299","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["PyCinemetricsV2: Interactive computational film software based on transformers and PySide6"],"prefix":"10.1016","volume":"31","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3869-125X","authenticated-orcid":false,"given":"Chunfang","family":"Li","sequence":"first","affiliation":[]},{"given":"Yalv","family":"Fan","sequence":"additional","affiliation":[]},{"given":"Yushi","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Kun","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yuhe","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yuchen","family":"Pei","sequence":"additional","affiliation":[]},{"given":"Tongtong","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Zhuoqi","family":"Shi","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.softx.2025.102299_b1","series-title":"Sculpting in time: reflections on the cinema. Translated by Kitty Hunter-Blair","author":"Tarkovsky","year":"1986"},{"key":"10.1016\/j.softx.2025.102299_b2","series-title":"Learning transferable visual models from natural language supervision","author":"Radford","year":"2021"},{"key":"10.1016\/j.softx.2025.102299_b3","series-title":"Robust speech recognition via large-scale weak supervision","author":"Radford","year":"2022"},{"key":"10.1016\/j.softx.2025.102299_b4","series-title":"GIT: A generative image-to-text transformer for vision and language","author":"Wang","year":"2022"},{"key":"10.1016\/j.softx.2025.102299_b5","unstructured":"Cinemetrics, https:\/\/cinemetrics.uchicago.edu\/how-to-use, 2025.01.08."},{"key":"10.1016\/j.softx.2025.102299_b6","doi-asserted-by":"crossref","unstructured":"Springstein M, Stamatakis M, Plank M, et al. TIB AV-Analytics: A web-based platform for scholarly video analysis and film studies. In: ACM international conference on research and development in information retrieval. Taipei, China; 2023, p. 1\u20135. http:\/\/dx.doi.org\/10.1145\/3539618.3591820.","DOI":"10.1145\/3539618.3591820"},{"key":"10.1016\/j.softx.2025.102299_b7","unstructured":"Azure AI Video Indexer, https:\/\/www.videoindexer.ai\/, 2025.01.09."},{"key":"10.1016\/j.softx.2025.102299_b8","doi-asserted-by":"crossref","DOI":"10.1016\/j.softx.2024.101686","article-title":"Pycinemetrics: Computational film studies tool based on deep learning and PySide2","volume":"26","author":"Li","year":"2024","journal-title":"SoftwareX"},{"key":"10.1016\/j.softx.2025.102299_b9","doi-asserted-by":"crossref","unstructured":"Sou\u010dek T, Loko\u010d J. Transnet v2: An effective deep network architecture for fast shot transition detection. In: The 32nd ACM international conference on multimedia. New York, United Statesp; 2024, p. 11218\u201321. http:\/\/dx.doi.org\/10.1145\/3664647.3685517.","DOI":"10.1145\/3664647.3685517"},{"key":"10.1016\/j.softx.2025.102299_b10","series-title":"IEEE conference on computer vision and pattern recognition","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He","year":"2016"},{"key":"10.1016\/j.softx.2025.102299_b11","series-title":"MovieNet: A holistic dataset for movie understanding","author":"Huang","year":"2020"},{"key":"10.1016\/j.softx.2025.102299_b12","series-title":"IEEE\/CVF conference on computer vision and pattern recognition","first-page":"4690","article-title":"ArcFace: Additive angular margin loss for deep face recognition","author":"Deng","year":"2019"},{"key":"10.1016\/j.softx.2025.102299_b13","series-title":"Python in science conference","doi-asserted-by":"crossref","first-page":"18","DOI":"10.25080\/Majora-7b98e3ed-003","article-title":"Librosa: Audio and music signal analysis in Python","author":"McFee","year":"2015"},{"key":"10.1016\/j.softx.2025.102299_b14","first-page":"919","article-title":"Real-time scene text detection with differentiable binarization and adaptive scale fusion","volume":"vol. 45, no. 1","author":"Liao","year":"2023"},{"key":"10.1016\/j.softx.2025.102299_b15","first-page":"172","article-title":"OpenPose: realtime multi-person 2D pose estimation using part affinity fields","volume":"vol. 43, no. 1","author":"Cao","year":"2021"},{"key":"10.1016\/j.softx.2025.102299_b16","series-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2014"},{"key":"10.1016\/j.softx.2025.102299_b17","doi-asserted-by":"crossref","first-page":"713","DOI":"10.1007\/s10579-023-09704-w","article-title":"Democratizing neural machine translation with OPUS-MT","volume":"58","author":"Tiedemann","year":"2023","journal-title":"Lang Resour Eval"},{"key":"10.1016\/j.softx.2025.102299_b18","series-title":"PaddleOCR: A practical ultra-lightweight OCR system","author":"Liu","year":"2020"},{"key":"10.1016\/j.softx.2025.102299_b19","unstructured":"PySide6, https:\/\/wiki.qt.io\/Qt_for_Python, 2025.01.08."},{"key":"10.1016\/j.softx.2025.102299_b20","first-page":"30","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"Adv Neural Inf Process Syst"},{"key":"10.1016\/j.softx.2025.102299_b21","series-title":"Swin transformer: Hierarchical vision transformer using shifted windows","author":"Liu","year":"2021"},{"key":"10.1016\/j.softx.2025.102299_b22","series-title":"Hugging face\u2019s transformers: State-of-the-art natural language processing","author":"Wolf","year":"2020"}],"container-title":["SoftwareX"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2352711025002651?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S2352711025002651?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,26]],"date-time":"2026-04-26T11:37:04Z","timestamp":1777203424000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S2352711025002651"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9]]},"references-count":22,"alternative-id":["S2352711025002651"],"URL":"https:\/\/doi.org\/10.1016\/j.softx.2025.102299","relation":{},"ISSN":["2352-7110"],"issn-type":[{"value":"2352-7110","type":"print"}],"subject":[],"published":{"date-parts":[[2025,9]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"PyCinemetricsV2: Interactive computational film software based on transformers and PySide6","name":"articletitle","label":"Article Title"},{"value":"SoftwareX","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.softx.2025.102299","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"simple-article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 The Authors. Published by Elsevier B.V.","name":"copyright","label":"Copyright"}],"article-number":"102299"}}