{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,25]],"date-time":"2026-04-25T09:38:47Z","timestamp":1777109927169,"version":"3.51.4"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T00:00:00Z","timestamp":1721606400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T00:00:00Z","timestamp":1721606400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62337001"],"award-info":[{"award-number":["62337001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["226-2022-00051"],"award-info":[{"award-number":["226-2022-00051"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"name":"HKUST Special Support for Young Faculty","award":["F0927"],"award-info":[{"award-number":["F0927"]}]},{"name":"HKUST Sports Science and Technology Research Grant","award":["SSTRG24EG04"],"award-info":[{"award-number":["SSTRG24EG04"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,1]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>\n                    Controllable Image Captioning (CIC)\u2014generating natural language descriptions about images under the guidance of given control signals\u2014is one of the most promising directions toward next-generation captioning systems. Till now, various kinds of control signals for CIC have been proposed, ranging from content-related control to structure-related control. However, due to the format and target gaps of different control signals, all existing CIC works (or architectures) only focus on one certain control signal, and overlook the human-like combinatorial ability. By \u201ccombinatorial\", we mean that our humans can easily meet multiple needs (or constraints) simultaneously when generating descriptions. To this end, we propose a novel prompt-based framework for CIC by learning\n                    <jats:bold>Com<\/jats:bold>\n                    binatorial\n                    <jats:bold>Pro<\/jats:bold>\n                    mpts, dubbed as\n                    <jats:bold>ComPro<\/jats:bold>\n                    . Specifically, we directly utilize a pretrained language model GPT-2 Radford et al. (OpenAI blog 1:9, 2019) as our language model, which can help to bridge the gap between different signal-specific CIC architectures. Then, we reformulate the CIC as a prompt-guide sentence generation problem, and propose a new lightweight prompt generation network to generate the combinatorial prompts for different kinds of control signals. For different control signals, we further design a new mask attention mechanism to realize the prompt-based CIC. Due to its simplicity, our ComPro can be further extended to more kinds of combined control signals by concatenating these prompts. Extensive experiments on two prevalent CIC benchmarks have verified the effectiveness and efficiency of our ComPro on both single and combined control signals.\n                  <\/jats:p>","DOI":"10.1007\/s11263-024-02179-4","type":"journal-article","created":{"date-parts":[[2024,7,22]],"date-time":"2024-07-22T05:01:54Z","timestamp":1721624514000},"page":"129-150","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["Learning Combinatorial Prompts for Universal Controllable Image Captioning"],"prefix":"10.1007","volume":"133","author":[{"given":"Zhen","family":"Wang","sequence":"first","affiliation":[]},{"given":"Jun","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Yueting","family":"Zhuang","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Shao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6148-9709","authenticated-orcid":false,"given":"Long","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,22]]},"reference":[{"key":"2179_CR1","doi-asserted-by":"crossref","unstructured":"Anderson, P., Fernando, B., Johnson, M., & Gould, S. (2016). Spice: Semantic propositional image caption evaluation. In ECCV (pp. 382\u2013398)","DOI":"10.1007\/978-3-319-46454-1_24"},{"key":"2179_CR2","doi-asserted-by":"crossref","unstructured":"Anderson, P., He, X., Buehler, C., Teney, D., Johnson, M., Gould, S., & Zhang, L. (2018). Bottom-up and top-down attention for image captioning and visual question answering. In CVPR (pp. 6077\u20136086).","DOI":"10.1109\/CVPR.2018.00636"},{"key":"2179_CR3","unstructured":"Banerjee, S., & Lavie, A. (2005). Meteor: An automatic metric for mt evaluation with improved correlation with human judgments. In ACL workshop (pp. 65\u201372)."},{"key":"2179_CR4","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al. (2020). Language models are few-shot learners. NeurIPS,33, 1877\u20131901.","journal-title":"Language models are few-shot learners. NeurIPS"},{"key":"2179_CR5","doi-asserted-by":"crossref","unstructured":"Chen, J., Guo, H., Yi, K., Li, B., & Elhoseiny, M. (2022). Visualgpt: Data-efficient adaptation of pretrained language models for image captioning. In CVPR (pp. 18030\u201318040).","DOI":"10.1109\/CVPR52688.2022.01750"},{"key":"2179_CR6","doi-asserted-by":"crossref","unstructured":"Chen, L., Zhang, H., Xiao, J., Nie, L, Shao, J., Liu, W., & Chua, T. S. (2017) Sca-cnn: Spatial and channel-wise attention in convolutional networks for image captioning. In CVPR (pp. 5659\u20135667).","DOI":"10.1109\/CVPR.2017.667"},{"key":"2179_CR7","doi-asserted-by":"crossref","unstructured":"Chen, L., Jiang, Z., Xiao, J., & Liu, W. (2021). Human-like controllable image captioning with verb-specific semantic roles. In CVPR (pp. 16846\u201316856)","DOI":"10.1109\/CVPR46437.2021.01657"},{"key":"2179_CR8","doi-asserted-by":"crossref","unstructured":"Chen, S., Jin, Q., Wang, P., & Wu, Q. (2020). Say as you wish: Fine-grained control of image caption generation with abstract scene graphs. In CVPR (pp. 9962\u20139971).","DOI":"10.1109\/CVPR42600.2020.00998"},{"key":"2179_CR9","doi-asserted-by":"crossref","unstructured":"Chen, T., Zhang, Z., You, Q., Fang, C., Wang, Z., Jin, H., & Luo, J. (2018). \u201cfactual\u201dor\u201cemotional\u201d: Stylized image captioning with adaptive learning and attention. In ECCV (pp. 519\u2013535).","DOI":"10.1007\/978-3-030-01249-6_32"},{"key":"2179_CR10","doi-asserted-by":"crossref","unstructured":"Cornia, M., Baraldi, L., & Cucchiara, R. (2019). Show, control and tell: A framework for generating controllable and grounded captions. In CVPR (pp. 8307\u20138316).","DOI":"10.1109\/CVPR.2019.00850"},{"key":"2179_CR11","doi-asserted-by":"crossref","unstructured":"Dai, B., Fidler, S., Urtasun, R., & Lin, D. (2017). Towards diverse and natural image descriptions via a conditional Gan. In: ICCV (pp. 2970\u20132979).","DOI":"10.1109\/ICCV.2017.323"},{"key":"2179_CR12","doi-asserted-by":"crossref","unstructured":"Deng, C., Ding, N., Tan, M., & Wu, Q. (2020). Length-controllable image captioning. In: ECCV (pp. 712\u2013729).","DOI":"10.1007\/978-3-030-58601-0_42"},{"key":"2179_CR13","doi-asserted-by":"crossref","unstructured":"Deshpande, A., Aneja, J., Wang, L., Schwing, A. G., & Forsyth, D. (2019). Fast, diverse and accurate image captioning guided by part-of-speech. In CVPR (pp. 10695\u201310704).","DOI":"10.1109\/CVPR.2019.01095"},{"key":"2179_CR14","unstructured":"Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding."},{"key":"2179_CR15","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., & Uszkoreit, J. (2020) An image is worth 16x16 words: Transformers for image recognition at scale."},{"key":"2179_CR16","doi-asserted-by":"crossref","unstructured":"Gan, C., Gan, Z., He, X., Gao, J., & Deng, L. (2017) Stylenet: Generating attractive visual captions with styles. In CVPR (pp. 3137\u20133146).","DOI":"10.1109\/CVPR.2017.108"},{"key":"2179_CR17","doi-asserted-by":"crossref","unstructured":"Han, X., Zhao, W., Ding, N., Liu, Z., & Sun, M. (2021). Ptr: Prompt tuning with rules for text classification.","DOI":"10.1016\/j.aiopen.2022.11.003"},{"key":"2179_CR18","doi-asserted-by":"crossref","unstructured":"Karpathy, A., & Fei-Fei, L. (2015). Deep visual-semantic alignments for generating image descriptions. In CVPR (pp. 3128\u20133137).","DOI":"10.1109\/CVPR.2015.7298932"},{"key":"2179_CR19","unstructured":"Keskar, N. S., McCann, B., Varshney, L. R., Xiong, C., & Socher, R. (2019). Ctrl: A conditional transformer language model for controllable generation."},{"key":"2179_CR20","doi-asserted-by":"crossref","unstructured":"Kikuchi, Y., Neubig, G., Sasano, R., Takamura, H., & Okumura, M. (2016) Controlling output length in neural encoder-decoders.","DOI":"10.18653\/v1\/D16-1140"},{"key":"2179_CR21","doi-asserted-by":"crossref","unstructured":"Lester, B., Al-Rfou, R., & Constant, N. (2021). The power of scale for parameter-efficient prompt tuning.","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"2179_CR22","unstructured":"Li, J., Li, D., Xiong, C., & Hoi, S. (2022). Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation."},{"key":"2179_CR23","doi-asserted-by":"crossref","unstructured":"Li, X. L., & Liang, P. (2021). Prefix-tuning: Optimizing continuous prompts for generation.","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"2179_CR24","unstructured":"Lin, C. Y. (2004). Rouge: A package for automatic evaluation of summaries. In ACL workshop (pp. 74\u201381)."},{"key":"2179_CR25","doi-asserted-by":"crossref","unstructured":"Lin, T. Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., & Zitnick, C. L. (2014). Microsoft coco: Common objects in context. In ECCV (pp. 740\u2013755).","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2179_CR26","doi-asserted-by":"crossref","unstructured":"Lindh, A., Ross, R. J., & Kelleher, J. D. (2020). Language-driven region pointer advancement for controllable image captioning.","DOI":"10.18653\/v1\/2020.coling-main.174"},{"key":"2179_CR27","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., & Neubig, G. (2021). Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing."},{"key":"2179_CR28","unstructured":"Luo, Z., Xi, Y., Zhang, R., & Ma, J. (2022a). I-tuning: Tuning language models with image for caption generation."},{"key":"2179_CR29","unstructured":"Luo, Z., Xi., Y., Zhang, R., & Ma, J. (2022b). Vc-gpt: Visual conditioned GPT for end-to-end generative vision-and-language pre-training."},{"key":"2179_CR30","doi-asserted-by":"crossref","unstructured":"Mao, Y., Chen, L., Jiang, Z., Zhang, D., Zhang, Z., Shao, J., & Xiao, J. (2022). Rethinking the reference-based distinctive image captioning. In ACM MM (pp. 4374\u20134384)","DOI":"10.1145\/3503161.3548358"},{"key":"2179_CR31","doi-asserted-by":"crossref","unstructured":"Mathews, A., Xie, L., & He, X. (2016). Senticap: Generating image descriptions with sentiments. In AAAI Vol.\u00a030","DOI":"10.1609\/aaai.v30i1.10475"},{"key":"2179_CR32","doi-asserted-by":"crossref","unstructured":"Mathews, A., Xie, L., & He, X. (2018). Semstyle: Learning to generate stylised image captions using unaligned text. In CVPR (pp. 8591\u20138600).","DOI":"10.1109\/CVPR.2018.00896"},{"key":"2179_CR33","doi-asserted-by":"crossref","unstructured":"Meng, Z., Yu, L., Zhang, N., Berg, T. L., Damavandi, B., Singh, V., & Bearman, A. (2021). Connecting what to say with where to look by modeling human attention traces. In CVPR (pp. 12679\u201312688).","DOI":"10.1109\/CVPR46437.2021.01249"},{"key":"2179_CR34","unstructured":"Mokady, R., Hertz, A., & Bermano, A. H. (2021). Clipcap: Clip prefix for image captioning."},{"issue":"3","key":"2179_CR35","doi-asserted-by":"publisher","first-page":"443","DOI":"10.1016\/0022-2836(70)90057-4","volume":"48","author":"SB Needleman","year":"1970","unstructured":"Needleman, S. B., & Wunsch, C. D. (1970). A general method applicable to the search for similarities in the amino acid sequence of two proteins. Journal of Molecular Biology,48(3), 443\u2013453.","journal-title":"Journal of Molecular Biology"},{"key":"2179_CR36","doi-asserted-by":"crossref","unstructured":"Papineni, K., Roukos, S., Ward, T., & Zhu, W.J. (2002). Bleu: A method for automatic evaluation of machine translation. In: ACL, (pp. 311\u2013318).","DOI":"10.3115\/1073083.1073135"},{"key":"2179_CR37","doi-asserted-by":"crossref","unstructured":"Plummer, B. A., Wang, L., Cervantes, C. M., Caicedo, J. C., Hockenmaier, J., Lazebnik, S. (2015). Flickr30k entities: Collecting region-to-phrase correspondences for richer image-to-sentence models. In ICCV (pp. 2641\u20132649).","DOI":"10.1109\/ICCV.2015.303"},{"key":"2179_CR38","doi-asserted-by":"crossref","unstructured":"Pont-Tuset, J., Uijlings, J., Changpinyo, S., Soricut, R., & Ferrari, V. (2020). Connecting vision and language with localized narratives. In ECCV (pp. 647\u2013664). Springer.","DOI":"10.1007\/978-3-030-58558-7_38"},{"key":"2179_CR39","unstructured":"Radford, A., Narasimhan, K., Salimans, T., & Sutskever, I. (2018). Improving language understanding by generative pre-training. OpenAI."},{"issue":"8","key":"2179_CR40","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al. (2019). Language models are unsupervised multitask learners. OpenAI blog,1(8), 9.","journal-title":"OpenAI blog"},{"key":"2179_CR41","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J. & Krueger, G. (2021) Learning transferable visual models from natural language supervision. In ICML (pp. 8748\u20138763)."},{"key":"2179_CR42","unstructured":"Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks. NeurIPS"},{"key":"2179_CR43","doi-asserted-by":"crossref","unstructured":"Schick, T., & Sch\u00fctze, H. (2020). Exploiting cloze questions for few shot text classification and natural language inference.","DOI":"10.18653\/v1\/2021.eacl-main.20"},{"key":"2179_CR44","doi-asserted-by":"crossref","unstructured":"Sharma, P., Ding, N., Goodman, S., Soricut, R. (2018). Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In ACL (pp. 2556\u20132565).","DOI":"10.18653\/v1\/P18-1238"},{"key":"2179_CR45","doi-asserted-by":"crossref","unstructured":"Shuster, K., Humeau, S., Hu, H., Bordes, A., & Weston, J. (2019). Engaging image captioning via personality. In CVPR (pp. 12516\u201312526).","DOI":"10.1109\/CVPR.2019.01280"},{"key":"2179_CR46","unstructured":"Su, Y., Lan, T., Liu, Y., Liu, F., Yogatama, D., Wang, Y., Kong, L., & Collier, N. (2022). Language models can see: Plugging visual controls in text generation."},{"key":"2179_CR47","unstructured":"Sun, Y., Wang, S., Feng, S., Ding, S., Pang, C., Shang, J., Liu, J., Chen, X., Zhao, Y., & Lu, Y., et\u00a0al. (2021) Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation."},{"key":"2179_CR48","unstructured":"Tang, T., Li, J., Zhao, W. X., Wen, J. R. (2022). Context-tuning: Learning contextualized prompts for natural language generation."},{"key":"2179_CR49","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, \u0141., & Polosukhin, I. (2017). Attention is all you need. NeurIPS30."},{"key":"2179_CR50","doi-asserted-by":"crossref","unstructured":"Vedantam, R., Lawrence Zitnick, C., & Parikh, D. (2015). Cider: Consensus-based image description evaluation. In CVPR (pp. 4566\u20134575).","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"2179_CR51","doi-asserted-by":"crossref","unstructured":"Vinyals, O., Toshev, A., Bengio, S., & Erhan, D. (2015). Show and tell: A neural image caption generator. In CVPR (pp. 3156\u20133164).","DOI":"10.1109\/CVPR.2015.7298935"},{"key":"2179_CR52","doi-asserted-by":"crossref","unstructured":"Wang, J., Xu, W., Wang, Q., & Chan, A. B. (2020). Compare and reweight: Distinctive image captioning using similar images sets. In ECCV (pp. 370\u2013386).","DOI":"10.1007\/978-3-030-58452-8_22"},{"key":"2179_CR53","first-page":"2617","volume":"37","author":"N Wang","year":"2023","unstructured":"Wang, N., Xie, J., Wu, J., Jia, M., & Li, L. (2023). Controllable image captioning via prompting. AAAI,37, 2617\u20132625.","journal-title":"Controllable image captioning via prompting. AAAI"},{"key":"2179_CR54","doi-asserted-by":"crossref","unstructured":"Wang, Z., Chen, L., Ma, W., Han, G., Niu, Y., Shao, J., & Xiao, J. (2022). Explicit image caption editing. In ECCV (pp. 113\u2013129). Springer","DOI":"10.1007\/978-3-031-20059-5_7"},{"key":"2179_CR55","doi-asserted-by":"crossref","unstructured":"Wang, Z., Xiao, J., Chen, T., Chen, L. (2023b). Decap: Towards generalized explicit caption editing via diffusion mechanism.","DOI":"10.1007\/978-3-031-72775-7_21"},{"key":"2179_CR56","unstructured":"Xu, K., Ba, J., Kiros, R., Cho, K., Courville, A., Salakhudinov, R., Zemel, R., & Bengio, Y. (2015). Show, attend and tell: Neural image caption generation with visual attention. In ICML, (pp. 2048\u20132057)."},{"key":"2179_CR57","doi-asserted-by":"crossref","unstructured":"Yan, K., Ji, L., Luo, H., Zhou, M., Duan, N., & Ma, S. (2021). Control image captioning spatially and temporally. In ACL-IJCNLP, (pp. 2014\u20132025).","DOI":"10.18653\/v1\/2021.acl-long.157"},{"key":"2179_CR58","doi-asserted-by":"crossref","unstructured":"Yang, K., Liu, D., Lei, W., Yang, B., Xue, M., Chen, B., & Xie, J. (2022). Tailor: A prompt-based approach to attribute-based controlled text generation.","DOI":"10.18653\/v1\/2023.acl-long.25"},{"key":"2179_CR59","doi-asserted-by":"crossref","unstructured":"Young, P., Lai, A., Hodosh, M., & Hockenmaier, J. (2014). From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. TACL,2, 67\u201378.","DOI":"10.1162\/tacl_a_00166"},{"key":"2179_CR60","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Li, Y., & Wang, S. (2019). Intention oriented image captions with guiding objects. In CVPR, (pp. 8395\u20138404).","DOI":"10.1109\/CVPR.2019.00859"},{"key":"2179_CR61","doi-asserted-by":"crossref","unstructured":"Zhong, Y., Wang, L., Chen, J., Yu, D., & Li, Y. (2020). Comprehensive image captioning via scene graph decomposition. In ECCV, (pp. 211\u2013229).","DOI":"10.1007\/978-3-030-58568-6_13"},{"key":"2179_CR62","unstructured":"Zhu, Z., Wang, T., & Qu, H. (2021). Self-annotated training for controllable image captioning"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02179-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-024-02179-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02179-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,25]],"date-time":"2026-04-25T08:41:40Z","timestamp":1777106500000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-024-02179-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,22]]},"references-count":62,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["2179"],"URL":"https:\/\/doi.org\/10.1007\/s11263-024-02179-4","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,22]]},"assertion":[{"value":"27 July 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 July 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 April 2026","order":5,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Update","order":6,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The original online version of this article was revised due to a retrospective Open Access order.","order":7,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}}]}}