{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,25]],"date-time":"2026-04-25T11:09:21Z","timestamp":1777115361398,"version":"3.51.4"},"reference-count":38,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["2523029"],"award-info":[{"award-number":["2523029"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013804","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100013804","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2021YFC2902703"],"award-info":[{"award-number":["2021YFC2902703"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neurocomputing"],"published-print":{"date-parts":[[2025,11]]},"DOI":"10.1016\/j.neucom.2025.131022","type":"journal-article","created":{"date-parts":[[2025,7,19]],"date-time":"2025-07-19T06:32:29Z","timestamp":1752906749000},"page":"131022","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":2,"special_numbering":"C","title":["Safe coordinated optimization of the thickening-dewatering process via reinforcement learning with real-time human guidance"],"prefix":"10.1016","volume":"652","author":[{"given":"Ranmeng","family":"Lin","sequence":"first","affiliation":[]},{"given":"Runda","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Fengyang","family":"Jiang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-7083-9455","authenticated-orcid":false,"given":"Jun","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Dakuo","family":"He","sequence":"additional","affiliation":[]},{"given":"Kang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Fuli","family":"Wang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neucom.2025.131022_bib0005","doi-asserted-by":"crossref","DOI":"10.1016\/j.mineng.2019.106141","article-title":"Data-driven-based self-healing control of abnormal feeding conditions in thickening-dewatering process","volume":"146","author":"Jia","year":"2020","journal-title":"Min. Eng."},{"issue":"9","key":"10.1016\/j.neucom.2025.131022_bib0010","doi-asserted-by":"crossref","first-page":"3173","DOI":"10.1080\/00207543.2017.1401236","article-title":"Scheduling on a two-machine permutation flow shop under time-of-use electricity tariffs","volume":"56","author":"Wang","year":"2018","journal-title":"Int. J. Prod. Res."},{"issue":"5","key":"10.1016\/j.neucom.2025.131022_bib0015","doi-asserted-by":"crossref","first-page":"3085","DOI":"10.1109\/TII.2019.2912018","article-title":"New methods for optimal operational control of industrial processes using reinforcement learning on two time scales","volume":"16","author":"Xue","year":"2020","journal-title":"IEEE Trans. Ind. Informat."},{"key":"10.1016\/j.neucom.2025.131022_bib0020","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1016\/j.compchemeng.2012.02.016","article-title":"On reliable and unreliable numerical methods for the simulation of secondary settling tanks in wastewater treatment","volume":"41","author":"B\u00fcrger","year":"2012","journal-title":"Comput. Chem. Eng."},{"key":"10.1016\/j.neucom.2025.131022_bib0025","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1016\/j.mineng.2017.01.011","article-title":"A study on model predictive control in paste thickeners with rake torque constraint","volume":"105","author":"Tan","year":"2017","journal-title":"Min. Eng."},{"issue":"3","key":"10.1016\/j.neucom.2025.131022_bib0030","doi-asserted-by":"crossref","first-page":"742","DOI":"10.1002\/cjce.23886","article-title":"A two-step coordinated optimization model for a dewatering process","volume":"99","author":"Zhang","year":"2021","journal-title":"Can. J. Chem. Eng."},{"key":"10.1016\/j.neucom.2025.131022_bib0035","doi-asserted-by":"crossref","first-page":"267","DOI":"10.1016\/j.cherd.2021.07.013","article-title":"Stochastic chance-constrained optimization framework for the thickening-dewatering process with an uncertain feed quantity","volume":"173","author":"Zhang","year":"2021","journal-title":"Chem. Eng. Res. Des."},{"key":"10.1016\/j.neucom.2025.131022_bib0040","first-page":"1","article-title":"Collaborative optimization framework for the industrial thickening\u2013dewatering process based on mixed integer linear programming","volume":"72","author":"Shulei","year":"2023","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"10.1016\/j.neucom.2025.131022_bib0045","doi-asserted-by":"crossref","first-page":"103","DOI":"10.1016\/j.cherd.2023.12.025","article-title":"Robust coordinated optimization of the thickening-dewatering process under uncertain feeding conditions","volume":"202","author":"Jia","year":"2024","journal-title":"Chem. Eng. Res. Des."},{"key":"10.1016\/j.neucom.2025.131022_bib0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.ress.2023.109682","article-title":"A dynamic mission abort policy for transportation systems with stochastic dependence by deep reinforcement learning","volume":"241","author":"Liu","year":"2024","journal-title":"Rel. Eng. Syst. Saf."},{"key":"10.1016\/j.neucom.2025.131022_bib0055","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.124402","article-title":"Sample-efficient reinforcement learning with knowledge-embedded hybrid model for optimal control of mining industry","volume":"254","author":"Zheng","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.neucom.2025.131022_bib0060","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1016\/j.neucom.2021.12.043","article-title":"Supervised assisted deep reinforcement learning for emergency voltage control of power systems","volume":"475","author":"Li","year":"2022","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2025.131022_bib0065","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2023.126986","article-title":"Personalized robotic control via constrained multi-objective reinforcement learning","volume":"565","author":"He","year":"2024","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2025.131022_bib0070","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2024.127904","article-title":"A deep reinforcement learning based distributed multi-uav dynamic area coverage algorithm for complex environment","volume":"595","author":"Xiao","year":"2024","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2025.131022_bib0075","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2025.129343","article-title":"Optimal scheduling of shared autonomous electric vehicles with multi-agent reinforcement learning: a mappo-based approach","volume":"622","author":"Tian","year":"2025","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2025.131022_bib0080","doi-asserted-by":"crossref","DOI":"10.1016\/j.ress.2022.109081","article-title":"Safe multi-agent deep reinforcement learning for joint bidding and maintenance scheduling of generation units","volume":"232","author":"Rokhforoz","year":"2023","journal-title":"Rel. Eng. Syst. Saf."},{"key":"10.1016\/j.neucom.2025.131022_bib0085","series-title":"The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event","article-title":"Efficient learning of safe driving policy via human-ai copilot optimization","author":"Li","year":"2022"},{"issue":"1","key":"10.1016\/j.neucom.2025.131022_bib0090","doi-asserted-by":"crossref","first-page":"855","DOI":"10.1109\/TNNLS.2022.3177685","article-title":"Prioritized experience-based reinforcement learning with human guidance for autonomous driving","volume":"35","author":"Wu","year":"2024","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"3","key":"10.1016\/j.neucom.2025.131022_bib0095","doi-asserted-by":"crossref","first-page":"4129","DOI":"10.1109\/TASE.2023.3292388","article-title":"Safe model-based reinforcement learning with an uncertainty-aware reachability certificate","volume":"21","author":"Yu","year":"2024","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"10.1016\/j.neucom.2025.131022_bib0100","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2025.110791","article-title":"Autonomous control of soft robots using safe reinforcement learning and covariance matrix adaptation","volume":"153","author":"Garg","year":"2025","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.neucom.2025.131022_bib0105","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2024.128677","article-title":"Off-policy safe reinforcement learning for nonlinear discrete-time systems","volume":"611","author":"Jha","year":"2025","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2025.131022_bib0110","doi-asserted-by":"crossref","first-page":"562","DOI":"10.1016\/j.neucom.2022.11.006","article-title":"Safe reinforcement learning for affine nonlinear systems with state constraints and input saturation using control barrier functions","volume":"518","author":"Liu","year":"2023","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2025.131022_bib0115","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.112312","article-title":"A safe reinforcement learning algorithm for supervisory control of power plants","volume":"301","author":"Sun","year":"2024","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.neucom.2025.131022_bib0120","doi-asserted-by":"crossref","DOI":"10.1016\/j.ress.2024.110639","article-title":"A novel sim2real reinforcement learning algorithm for process control","volume":"254","author":"Liang","year":"2025","journal-title":"Rel. Eng. Syst. Saf."},{"key":"10.1016\/j.neucom.2025.131022_bib0125","article-title":"Deep reinforcement learning for maintenance optimization of a scrap-based steel production line","volume":"249","author":"Neto","year":"2024","journal-title":"Rel. Eng. Syst. Saf."},{"key":"10.1016\/j.neucom.2025.131022_bib0130","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2023.107784","article-title":"Balancing therapeutic effect and safety in ventilator parameter recommendation: an offline reinforcement learning approach","volume":"131","author":"Zhang","year":"2024","journal-title":"Eng. Appl. Artif. Intell."},{"key":"10.1016\/j.neucom.2025.131022_bib0135","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2025.129912","article-title":"Hybrid safe reinforcement learning: tackling distribution shift and outliers with the Student-t\u2019s process","volume":"634","author":"Hickman","year":"2025","journal-title":"Neurocomputing"},{"issue":"12","key":"10.1016\/j.neucom.2025.131022_bib0140","doi-asserted-by":"crossref","first-page":"11216","DOI":"10.1109\/TPAMI.2024.3457538","article-title":"A review of safe reinforcement learning: methods, theories, and applications","volume":"46","author":"Gu","year":"2024","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"7","key":"10.1016\/j.neucom.2025.131022_bib0145","doi-asserted-by":"crossref","first-page":"3374","DOI":"10.1109\/TAI.2024.3351797","article-title":"Learn from safe experience: safe reinforcement learning for task automation of surgical robot","volume":"5","author":"Fan","year":"2024","journal-title":"IEEE Trans. Artif. Intell."},{"issue":"1","key":"10.1016\/j.neucom.2025.131022_bib0150","article-title":"Human-in-the-loop transfer learning in collision avoidance of autonomous robots","volume":"5","author":"Oriyama","year":"2025","journal-title":"Biomim. Intell. Robot."},{"issue":"11","key":"10.1016\/j.neucom.2025.131022_bib0155","doi-asserted-by":"crossref","first-page":"16181","DOI":"10.1109\/TITS.2024.3420959","article-title":"Safety-aware human-in-the-loop reinforcement learning with shared control for autonomous driving","volume":"25","author":"Huang","year":"2024","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"10.1016\/j.neucom.2025.131022_bib0160","doi-asserted-by":"crossref","DOI":"10.1016\/j.commtr.2023.100103","article-title":"Envisioning the future of transportation: inspiration of chatgpt and large models","volume":"3","author":"Qu","year":"2023","journal-title":"Commun. Transp. Res."},{"issue":"5","key":"10.1016\/j.neucom.2025.131022_bib0165","first-page":"1095","article-title":"Data driven coordinated optimization control of thickening-filter process","volume":"36","author":"Zhang","year":"2021","journal-title":"Kongzhi Yu Juece\/Control And Decis."},{"key":"10.1016\/j.neucom.2025.131022_bib0170","series-title":"Proceedings of the 35th International Conference on Machine Learning, ICML 2018","first-page":"1856","article-title":"Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor","author":"Haarnoja","year":"2018"},{"key":"10.1016\/j.neucom.2025.131022_bib0175","series-title":"4th International Conference on Learning Representations, ICLR 2016","article-title":"Continuous control with deep reinforcement learning","author":"Lillicrap","year":"2016"},{"key":"10.1016\/j.neucom.2025.131022_bib0180","series-title":"Proceedings of the 35th International Conference on Machine Learning, ICML 2018","first-page":"1582","article-title":"Addressing function approximation error in actor-critic methods","author":"Fujimoto","year":"2018"},{"key":"10.1016\/j.neucom.2025.131022_bib0185","series-title":"Machine Learning, Proceedings of the Nineteenth International Conference","first-page":"267","article-title":"Approximately optimal approximate reinforcement learning","author":"Kakade","year":"2002"},{"key":"10.1016\/j.neucom.2025.131022_bib0190","doi-asserted-by":"crossref","DOI":"10.1016\/j.commtr.2024.100127","article-title":"Human as ai mentor: enhanced human-in-the-loop reinforcement learning for safe and efficient autonomous driving","volume":"4","author":"Huang","year":"2024","journal-title":"Commun. Transp. Res."}],"container-title":["Neurocomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231225016947?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231225016947?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T02:27:44Z","timestamp":1772764064000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0925231225016947"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11]]},"references-count":38,"alternative-id":["S0925231225016947"],"URL":"https:\/\/doi.org\/10.1016\/j.neucom.2025.131022","relation":{},"ISSN":["0925-2312"],"issn-type":[{"value":"0925-2312","type":"print"}],"subject":[],"published":{"date-parts":[[2025,11]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Safe coordinated optimization of the thickening-dewatering process via reinforcement learning with real-time human guidance","name":"articletitle","label":"Article Title"},{"value":"Neurocomputing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neucom.2025.131022","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"131022"}}