{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T02:29:09Z","timestamp":1760149749454,"version":"build-2065373602"},"reference-count":43,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,6,1]],"date-time":"2023-06-01T00:00:00Z","timestamp":1685577600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/100012542","name":"Sichuan Province Science and Technology Support Program","doi-asserted-by":"publisher","award":["2022ZHCG0007"],"award-info":[{"award-number":["2022ZHCG0007"]}],"id":[{"id":"10.13039\/100012542","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004829","name":"Sichuan Province Department of Science and Technology","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100004829","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100018542","name":"Natural Science Foundation of Sichuan Province","doi-asserted-by":"publisher","award":["2022NSFSC0503"],"award-info":[{"award-number":["2022NSFSC0503"]}],"id":[{"id":"10.13039\/501100018542","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Computer Speech &amp; Language"],"published-print":{"date-parts":[[2023,6]]},"DOI":"10.1016\/j.csl.2023.101504","type":"journal-article","created":{"date-parts":[[2023,3,7]],"date-time":"2023-03-07T02:06:13Z","timestamp":1678154773000},"page":"101504","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":2,"special_numbering":"C","title":["Improving named entity correctness of abstractive summarization by generative negative sampling"],"prefix":"10.1016","volume":"81","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4013-3492","authenticated-orcid":false,"given":"Zheng","family":"Chen","sequence":"first","affiliation":[]},{"given":"Hongyu","family":"Lin","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.csl.2023.101504_b1","doi-asserted-by":"crossref","DOI":"10.1016\/j.csl.2021.101276","article-title":"Deep reinforcement and transfer learning for abstractive text summarization: A review","volume":"71","author":"Alomari","year":"2022","journal-title":"Comput. Speech Lang."},{"key":"10.1016\/j.csl.2023.101504_b2","unstructured":"Bao,\u00a0H., Dong,\u00a0L., Wei,\u00a0F., Wang,\u00a0W., Yang,\u00a0N., Liu,\u00a0X., Wang,\u00a0Y., Piao,\u00a0S., Gao,\u00a0J., Zhou,\u00a0M., Hon,\u00a0H.-W., 2020. UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training. In: International Conference on Machine Learning. pp. 642\u2013652."},{"key":"10.1016\/j.csl.2023.101504_b3","series-title":"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing","article-title":"Factual error correction for abstractive summarization models","author":"Cao","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b4","doi-asserted-by":"crossref","unstructured":"Cao,\u00a0Z., Wei,\u00a0F., Li,\u00a0W., Li,\u00a0S., 2018. Faithful to the Original: Fact Aware Neural Abstractive Summarization. In: AAAI Conference on Artificial Intelligence.","DOI":"10.1609\/aaai.v32i1.11912"},{"key":"10.1016\/j.csl.2023.101504_b5","series-title":"Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations","first-page":"42","article-title":"N-LTP: An open-source neural language technology platform for Chinese","author":"Che","year":"2021"},{"key":"10.1016\/j.csl.2023.101504_b6","doi-asserted-by":"crossref","unstructured":"Chen,\u00a0S., Zhang,\u00a0F., Sone,\u00a0K., Roth,\u00a0D., 2021. Improving Faithfulness in Abstractive Summarization with Contrast Candidate Generation and Selection. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. pp. 5935\u20135941.","DOI":"10.18653\/v1\/2021.naacl-main.475"},{"key":"10.1016\/j.csl.2023.101504_b7","series-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2019"},{"key":"10.1016\/j.csl.2023.101504_b8","series-title":"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing","article-title":"Multi-fact correction in abstractive text summarization","author":"Dong","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b9","doi-asserted-by":"crossref","unstructured":"Elkan,\u00a0C., Noto,\u00a0K., 2008. Learning classifiers from only positive and unlabeled data. In: Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. pp. 213\u2013220.","DOI":"10.1145\/1401890.1401920"},{"key":"10.1016\/j.csl.2023.101504_b10","series-title":"Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics","article-title":"Ranking generated summaries by correctness: An interesting but challenging application for natural language inference","author":"Falke","year":"2019"},{"issue":"5","key":"10.1016\/j.csl.2023.101504_b11","doi-asserted-by":"crossref","first-page":"378","DOI":"10.1037\/h0031619","article-title":"Measuring nominal scale agreement among many raters","volume":"76","author":"Fleiss","year":"1971","journal-title":"Psychol. Bull."},{"key":"10.1016\/j.csl.2023.101504_b12","series-title":"Findings of the Association for Computational Linguistics","first-page":"478","article-title":"GO FIGURE: A meta evaluation of factuality in summarization","author":"Gabriel","year":"2021"},{"key":"10.1016\/j.csl.2023.101504_b13","first-page":"166","article-title":"Assessing the factual accuracy of generated text","author":"Goodrich","year":"2019"},{"key":"10.1016\/j.csl.2023.101504_b14","series-title":"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","article-title":"Soft layer-specific multi-task summarization with entailment and question generation","author":"Guo","year":"2018"},{"key":"10.1016\/j.csl.2023.101504_b15","series-title":"Natural Language Processing and Chinese Computing","first-page":"942","article-title":"Overview of the NLPCC 2017 shared task: Single document summarization","author":"Hua","year":"2018"},{"key":"10.1016\/j.csl.2023.101504_b16","first-page":"592","article-title":"An evaluation of two-step techniques for positive-unlabeled learning in text classification","volume":"3","author":"Kaboutari","year":"2014","journal-title":"Int. J. Comput. Appl. Technol. Res."},{"year":"1990","series-title":"Rank Correlation Methods","author":"Kendall","key":"10.1016\/j.csl.2023.101504_b17"},{"key":"10.1016\/j.csl.2023.101504_b18","series-title":"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing","article-title":"Evaluating the factual consistency of abstractive text summarization","author":"Kryscinski","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b19","series-title":"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing","article-title":"Improving abstraction in text summarization","author":"Kry\u015bci\u0144ski","year":"2018"},{"key":"10.1016\/j.csl.2023.101504_b20","doi-asserted-by":"crossref","unstructured":"Lewis,\u00a0M., Liu,\u00a0Y., Goyal,\u00a0N., Ghazvininejad,\u00a0M., Mohamed,\u00a0A., Levy,\u00a0O., Stoyanov,\u00a0V., Zettlemoyer,\u00a0L., 2020. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. pp. 7871\u20137880.","DOI":"10.18653\/v1\/2020.acl-main.703"},{"year":"2021","series-title":"XIANDAI HANYU CHANGYONGCIBLAO","author":"Li","key":"10.1016\/j.csl.2023.101504_b21"},{"key":"10.1016\/j.csl.2023.101504_b22","series-title":"Proceedings of the 27th International Conference on Computational Linguistics","article-title":"Ensure the correctness of the summary: Incorporate entailment knowledge into abstractive sentence summarization","author":"Li","year":"2018"},{"key":"10.1016\/j.csl.2023.101504_b23","series-title":"Text Summarization Branches Out","article-title":"ROUGE: A package for automatic evaluation of summaries","author":"Lin","year":"2004"},{"key":"10.1016\/j.csl.2023.101504_b24","series-title":"Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing","article-title":"Text summarization with pretrained encoders","author":"Liu","year":"2019"},{"key":"10.1016\/j.csl.2023.101504_b25","series-title":"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics","article-title":"On faithfulness and factuality in abstractive summarization","author":"Maynez","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b26","series-title":"Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning","article-title":"Abstractive text summarization using sequence-to-sequence RNNs and beyond","author":"Nallapati","year":"2016"},{"key":"10.1016\/j.csl.2023.101504_b27","series-title":"Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume","article-title":"Entity-level factual consistency of abstractive text summarization","author":"Nan","year":"2021"},{"key":"10.1016\/j.csl.2023.101504_b28","series-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","first-page":"6881","article-title":"Improving factual consistency of abstractive summarization via question answering","author":"Nan","year":"2021"},{"key":"10.1016\/j.csl.2023.101504_b29","series-title":"Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","first-page":"4812","article-title":"Understanding factuality in abstractive summarization with FRANK: A benchmark for factuality metrics","author":"Pagnoni","year":"2021"},{"key":"10.1016\/j.csl.2023.101504_b30","series-title":"Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics","article-title":"Bleu: a method for automatic evaluation of machine translation","author":"Papineni","year":"2002"},{"key":"10.1016\/j.csl.2023.101504_b31","series-title":"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing","article-title":"On extractive and abstractive neural document summarization with transformer language models","author":"Pilault","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b32","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"J. Mach. Learn. Res."},{"key":"10.1016\/j.csl.2023.101504_b33","series-title":"Proceedings of the 36th International Conference on Machine Learning","first-page":"5926","article-title":"MASS: Masked sequence to sequence pre-training for language generation","author":"Song","year":"2019"},{"key":"10.1016\/j.csl.2023.101504_b34","doi-asserted-by":"crossref","unstructured":"Stahlberg,\u00a0F., Byrne,\u00a0B., 2019. On NMT Search Errors and Model Errors: Cat Got Your Tongue?. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. EMNLP-IJCNLP, pp. 3356\u20133362.","DOI":"10.18653\/v1\/D19-1331"},{"key":"10.1016\/j.csl.2023.101504_b35","series-title":"Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","author":"Sutskever","year":"2014"},{"key":"10.1016\/j.csl.2023.101504_b36","series-title":"Advances in Neural Information Processing Systems, Vol. 30","article-title":"Attention is all you need","author":"Vaswani","year":"2017"},{"year":"2018","series-title":"Diverse beam search: Decoding diverse solutions from neural sequence models","author":"Vijayakumar","key":"10.1016\/j.csl.2023.101504_b37"},{"key":"10.1016\/j.csl.2023.101504_b38","series-title":"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics","article-title":"Asking and answering questions to evaluate the factual consistency of summaries","author":"Wang","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b39","doi-asserted-by":"crossref","unstructured":"Wei,\u00a0B., Wang,\u00a0M., Zhou,\u00a0H., Lin,\u00a0J., Sun,\u00a0X., 2019. Imitation Learning for Non-Autoregressive Neural Machine Translation. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. pp. 1304\u20131312.","DOI":"10.18653\/v1\/P19-1125"},{"key":"10.1016\/j.csl.2023.101504_b40","series-title":"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations","article-title":"Transformers: State-of-the-art natural language processing","author":"Wolf","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b41","series-title":"Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","first-page":"483","article-title":"mT5: A massively multilingual pre-trained text-to-text transformer","author":"Xue","year":"2021"},{"key":"10.1016\/j.csl.2023.101504_b42","series-title":"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics","article-title":"Optimizing the factual correctness of a summary: A study of summarizing radiology reports","author":"Zhang","year":"2020"},{"key":"10.1016\/j.csl.2023.101504_b43","series-title":"Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","first-page":"718","article-title":"Enhancing factual consistency of abstractive summarization","author":"Zhu","year":"2021"}],"container-title":["Computer Speech &amp; Language"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0885230823000232?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0885230823000232?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T20:46:23Z","timestamp":1760129183000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0885230823000232"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6]]},"references-count":43,"alternative-id":["S0885230823000232"],"URL":"https:\/\/doi.org\/10.1016\/j.csl.2023.101504","relation":{},"ISSN":["0885-2308"],"issn-type":[{"type":"print","value":"0885-2308"}],"subject":[],"published":{"date-parts":[[2023,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Improving named entity correctness of abstractive summarization by generative negative sampling","name":"articletitle","label":"Article Title"},{"value":"Computer Speech & Language","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.csl.2023.101504","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2023 Elsevier Ltd. All rights reserved.","name":"copyright","label":"Copyright"}],"article-number":"101504"}}