{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T12:00:38Z","timestamp":1778155238759,"version":"3.51.4"},"reference-count":371,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,8,13]],"date-time":"2025-08-13T00:00:00Z","timestamp":1755043200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,8,13]],"date-time":"2025-08-13T00:00:00Z","timestamp":1755043200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI Ethics"],"published-print":{"date-parts":[[2025,10]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>Large Language Models (LLMs) have achieved unparalleled success across diverse language modeling tasks in recent years. However, this progress has also intensified ethical concerns, impacting the deployment of LLMs in everyday contexts. This paper provides a comprehensive survey of ethical challenges associated with LLMs, from longstanding issues such as copyright infringement, systematic bias, and data privacy, to emerging problems like truthfulness and social norms. We critically analyze existing research aimed at understanding, examining, and mitigating these ethical risks. Our survey underscores integrating ethical standards and societal values into the development of LLMs, thereby guiding the development of responsible and ethically aligned language models.<\/jats:p>","DOI":"10.1007\/s43681-025-00797-3","type":"journal-article","created":{"date-parts":[[2025,8,13]],"date-time":"2025-08-13T07:56:03Z","timestamp":1755071763000},"page":"4745-4771","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Deconstructing the ethics of large language models from long-standing issues to new-emerging dilemmas: a survey"],"prefix":"10.1007","volume":"5","author":[{"given":"Chengyuan","family":"Deng","sequence":"first","affiliation":[]},{"given":"Yiqun","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Heng","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Yijun","family":"Tian","sequence":"additional","affiliation":[]},{"given":"Han","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yichen","family":"Wang","sequence":"additional","affiliation":[]},{"given":"kuofeng","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Henry Peng","family":"Zou","sequence":"additional","affiliation":[]},{"given":"Yiqiao","family":"jin","sequence":"additional","affiliation":[]},{"given":"Yijia","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Shenghao","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Zongxing","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Weimin","family":"Lyu","sequence":"additional","affiliation":[]},{"given":"Sihong","family":"He","sequence":"additional","affiliation":[]},{"given":"Lu","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Haohan","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7142-2193","authenticated-orcid":false,"given":"Jun","family":"Zhuang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,8,13]]},"reference":[{"key":"797_CR1","doi-asserted-by":"crossref","unstructured":"Abadi, M., Chu, A., Goodfellow., I, McMahan., H\u00a0B, Mironov., I, Talwar., K, Zhang., L: Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, pages 308\u2013318, (2016)","DOI":"10.1145\/2976749.2978318"},{"key":"797_CR2","doi-asserted-by":"crossref","unstructured":"Abascal, J., Wu, S., Oprea, A., Ullman, J.: Tmi! finetuned models leak private information from their pretraining data. arXiv preprint arXiv:2306.01181, (2023)","DOI":"10.56553\/popets-2024-0075"},{"key":"797_CR3","unstructured":"Abbas, A., Tirumala, K., Simig, D., Ganguli, S., Morcos, A\u00a0S.: Semdedup: Data-efficient learning at web-scale through semantic deduplication. arXiv preprint arXiv:2303.09540, (2023)"},{"key":"797_CR4","doi-asserted-by":"crossref","unstructured":"Abid, A., Farooqi, M., Zou, J.: Persistent anti-muslim bias in large language models. In Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society, pages 298\u2013306, (2021)","DOI":"10.1145\/3461702.3462624"},{"key":"797_CR5","unstructured":"Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, D., Almeida, F\u0152., Altenschmidt, J., Altman, S., Anadkat, S., et\u00a0al.: Gpt-4 technical report. arXiv preprint arXiv:2303.08774, (2023)"},{"key":"797_CR6","unstructured":"Adi, Y., Baum, C., Cisse, M., Pinkas, B., Keshet, J.: Turning your weakness into a strength: Watermarking deep neural networks by backdooring. In 27th USENIX Security Symposium (USENIX Security 18), pages 1615\u20131631, (2018)"},{"key":"797_CR7","doi-asserted-by":"crossref","unstructured":"Ahn, J., Oh, A.: Mitigating language-dependent ethnic bias in BERT. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 533\u2013549, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.emnlp-main.42"},{"key":"797_CR8","doi-asserted-by":"crossref","unstructured":"Ahvanooey, M\u00a0T., Li, Q., Shim, H\u0160., Huang, Y.: A comparative analysis of information hiding techniques for copyright protection of text documents. Secur. Commun. Networks, 2018:5325040:1\u20135325040:22, (2018)","DOI":"10.1155\/2018\/5325040"},{"key":"797_CR9","first-page":"191","volume":"25","author":"MT Ahvanooey","year":"2016","unstructured":"Ahvanooey, M.T., Mazraeh, H.D., Tabasi, S.H.: An innovative technique for web text watermarking (aitw). Inf. Secur. J.: Global Perspect. 25, 191\u2013196 (2016)","journal-title":"Inf. Secur. J.: Global Perspect."},{"key":"797_CR10","doi-asserted-by":"crossref","unstructured":"Anil, R., Ghazi, B., Gupta, V., Kumar, R., Manurangsi, P.: Large-scale differentially private bert. arXiv preprint arXiv:2108.01624, (2021)","DOI":"10.18653\/v1\/2022.findings-emnlp.484"},{"key":"797_CR11","unstructured":"Askell, A., Bai, Y., Chen, A., Drain, D., Ganguli, D., Henighan, T., Jones, A., Joseph, N., Mann, B., DasSarma, N., et\u00a0al.: A general language assistant as a laboratory for alignment. arXiv preprint arXiv:2112.00861, (2021)"},{"key":"797_CR12","doi-asserted-by":"crossref","unstructured":"Atallah, M\u00a0J., Raskin, V., Hempelmann, C\u2020., Karahan, M., Sion, R., Topkara, U., Triezenberg, K\u2026.: Natural language watermarking and tamperproofing. In International Workshop on Information Hiding, pages 196\u2013212, (2002)","DOI":"10.1007\/3-540-36415-3_13"},{"key":"797_CR13","doi-asserted-by":"crossref","unstructured":"Attanasio, G., Nozza, D., Hovy, D., Baralis, E.: Entropy-based attention regularization frees unintended bias mitigation from lists. In Findings of the Association for Computational Linguistics: ACL 2022, pages 1105\u20131119, Dublin, Ireland, (May 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.findings-acl.88"},{"issue":"1","key":"797_CR14","doi-asserted-by":"crossref","first-page":"120","DOI":"10.1186\/s13054-023-04393-x","volume":"27","author":"R Azamfirei","year":"2023","unstructured":"Azamfirei, R., Kudchadkar, S.R., Fackler, J.: Large language models and the perils of their hallucinations. Crit. Care 27(1), 120 (2023)","journal-title":"Crit. Care"},{"key":"797_CR15","doi-asserted-by":"crossref","unstructured":"Baek, J, Aji, A\u00a0F, Saffari, A: Knowledge-augmented language model prompting for zero-shot knowledge graph question answering. arXiv:2306.04136, (2023)","DOI":"10.18653\/v1\/2023.nlrse-1.7"},{"key":"797_CR16","doi-asserted-by":"crossref","unstructured":"Bai, J., Gao, K., Min, S., Xia, S.-T., Li, Z.: and Wei Liu. Trigger-aware prompt learning for backdoor attacks on clip. In CVPR, Badclip (2024)","DOI":"10.1109\/CVPR52733.2024.02288"},{"key":"797_CR17","unstructured":"Bai, Y., Jones, A., Ndousse, K., Askell, A., Chen, A., DasSarma, N., Drain, D., Fort, S., Ganguli, D., Henighan, T., et\u00a0al.: Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862, (2022)"},{"key":"797_CR18","unstructured":"Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., Chen, A., Goldie, A., Mirhoseini, A., McKinnon, C., et\u00a0al.: Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073 (2022)"},{"key":"797_CR19","first-page":"7641","volume":"35","author":"M Balunovic","year":"2022","unstructured":"Balunovic, M., Dimitrov, D., Jovanovi\u0107, N., Vechev, M.: Lamp: Extracting text from gradients with language model priors. Adv. Neural. Inf. Process. Syst. 35, 7641\u20137654 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR20","unstructured":"Barocas, S., Hardt, M., Narayanan, A.: Fairness and Machine Learning: Limitations and Opportunities. fairmlbook.org, 2019. http:\/\/www.fairmlbook.org"},{"key":"797_CR21","doi-asserted-by":"crossref","unstructured":"Bender, E\u00a0M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pages 610\u2013623, (2021)","DOI":"10.1145\/3442188.3445922"},{"key":"797_CR22","doi-asserted-by":"crossref","first-page":"547","DOI":"10.1147\/sj.393.0547","volume":"39","author":"W Bender","year":"2000","unstructured":"Bender, W., Butera, W., Gruhl, D.F., Hwang, R., Paiz, F.J., Pogreb, S.: Applications for data hiding. IBM Syst. J. 39, 547\u2013568 (2000)","journal-title":"IBM Syst. J."},{"key":"797_CR23","volume-title":"Morimoto","author":"W Bender","year":"1995","unstructured":"Bender, W., Gruhl, D.F.: Morimoto. Techniques for data hiding. In Electronic imaging, Norishige, Lu, Anthony (1995)"},{"key":"797_CR24","unstructured":"Bergmair, R.: Towards linguistic steganography: A systematic investigation of approaches, systems, and issues. (2004)"},{"key":"797_CR25","unstructured":"Berk, R., Heidari, H., Jabbari, S., Joseph, M., Kearns, M., Morgenstern, J., Neel, S., Roth, A.: A convex framework for fair regression. arXiv preprint arXiv:1706.02409, (2017)"},{"key":"797_CR26","unstructured":"Bertuzzi, L.: Ai act: Eu parliament\u2019s crunch time on high-risk categorisation, prohibited practices, (2023)"},{"key":"797_CR27","unstructured":"Bertuzzi, L.: Ai act: Meps close in on rules for general purpose ai, foundation models, (2023)"},{"key":"797_CR28","unstructured":"Bertuzzi, L.: Meps seal the deal on artificial intelligence act, (2023)"},{"key":"797_CR29","doi-asserted-by":"crossref","first-page":"1","DOI":"10.12840\/issn.2255-4165.017","volume":"7","author":"CJ Beukeboom","year":"2019","unstructured":"Beukeboom, C.J.: Burgers, Christian: How stereotypes are shared through language: a review and introduction of the aocial categories and stereotypes communication (scsc) framework. Rev. Commun. Res. 7, 1\u201337 (2019)","journal-title":"Rev. Commun. Res."},{"key":"797_CR30","unstructured":"Biderman, S., Schoelkopf, H., Anthony, Q\u00a0G., Bradley, H., O\u2019Brien, K., Hallahan, E., Khan, M\u00a0A., Purohit, S., Prashanth, U\u00a0S., Raff, E., et\u00a0al.: Pythia: A suite for analyzing large language models across training and scaling. In International Conference on Machine Learning, pages 2397\u20132430. PMLR, (2023)"},{"key":"797_CR31","doi-asserted-by":"crossref","unstructured":"Bishop, M., Cummins, J., Peisert, S., Singh, A., Bhumiratana, B., Agarwal, D., Frincke, D., Hogarth, M.: Relationships and data sanitization: A study in scarlet. In Proceedings of the 2010 New Security Paradigms Workshop, pages 151\u2013164, (2010)","DOI":"10.1145\/1900546.1900567"},{"key":"797_CR32","doi-asserted-by":"crossref","unstructured":"Blodgett, S\u00a0L., Barocas, S., Daum\u00e9\u00a0III, H., Wallach, H.: Language (technology) is power: A critical survey of \u201cbias\u201d in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5454\u20135476, Online, July 2020. Association for Computational Linguistics","DOI":"10.18653\/v1\/2020.acl-main.485"},{"key":"797_CR33","unstructured":"Blodgett, S\u00a0L., O\u2019Connor, B.: Racial disparity in natural language processing: A case study of social media african-american english. arXiv preprint arXiv:1707.00061, (2017)"},{"key":"797_CR34","unstructured":"Bommasani, R., Klyman, K., Zhang, D., Liang, P.: Do foundation model providers comply with the eu ai act?, (2023)"},{"key":"797_CR35","doi-asserted-by":"crossref","unstructured":"Bommasani, R., Liang, P., Lee, T.: Holistic evaluation of language models. Annals of the New York Academy of Sciences, (2023)","DOI":"10.1111\/nyas.15007"},{"key":"797_CR36","doi-asserted-by":"crossref","first-page":"1181","DOI":"10.1109\/5.771071","volume":"87","author":"J Brassil","year":"1999","unstructured":"Brassil, J., Low, S.H., Maxemchuk, N.F.: Copyright protection for the electronic distribution of text documents. Proc. IEEE 87, 1181\u20131196 (1999)","journal-title":"Proc. IEEE"},{"issue":"8","key":"797_CR37","doi-asserted-by":"crossref","first-page":"1495","DOI":"10.1109\/49.464718","volume":"13","author":"JT Brassil","year":"1995","unstructured":"Brassil, J.T., Low, S., Maxemchuk, N.F., O\u2019Gorman, L.: Electronic marking and identification techniques to discourage document copying. IEEE J. Sel. Areas Commun. 13(8), 1495\u20131504 (1995)","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"797_CR38","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"1","key":"797_CR39","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1017\/err.2019.8","volume":"10","author":"MC Buiten","year":"2019","unstructured":"Buiten, M.C.: Towards intelligent regulation of artificial intelligence. Eur. J. Risk Regul. 10(1), 41\u201359 (2019)","journal-title":"Eur. J. Risk Regul."},{"issue":"6334","key":"797_CR40","doi-asserted-by":"crossref","first-page":"183","DOI":"10.1126\/science.aal4230","volume":"356","author":"A Caliskan","year":"2017","unstructured":"Caliskan, A., Bryson, J.J., Narayanan, A.: Semantics derived automatically from language corpora contain human-like biases. Science 356(6334), 183\u2013186 (2017)","journal-title":"Science"},{"key":"797_CR41","doi-asserted-by":"crossref","unstructured":"Cao, Y., Zhou, L., Lee, S., Cabello, L., Chen, M., Hershcovich, D.: Assessing cross-cultural alignment between chatgpt and human societies: An empirical study. arXiv preprint arXiv:2303.17466 (2023)","DOI":"10.18653\/v1\/2023.c3nlp-1.7"},{"key":"797_CR42","doi-asserted-by":"crossref","unstructured":"Carlini, N., Chien, S., Nasr, M., Song, S., Terzis, A., Tramer, F.: Membership inference attacks from first principles. In 2022 IEEE Symposium on Security and Privacy (SP), pages 1897\u20131914. IEEE, (2022)","DOI":"10.1109\/SP46214.2022.9833649"},{"key":"797_CR43","unstructured":"Carlini, N., Tramer, F., Wallace, E., Jagielski, M., Herbert-Voss, A., Lee, K., Roberts, A., Brown, T., Song, D., Erlingsson, U., et\u00a0al.: Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633\u20132650, (2021)"},{"key":"797_CR44","unstructured":"Carlini, N., Hayes, J., Nasr, M., Jagielski, M., Sehwag, V., Tramer, F., Balle, B., Ippolito, D., Wallace, E.: Extracting training data from diffusion models. In 32nd USENIX Security Symposium (USENIX Security 23), pages 5253\u20135270, (2023)"},{"key":"797_CR45","doi-asserted-by":"crossref","first-page":"501","DOI":"10.1007\/s11948-019-00151-x","volume":"26","author":"J-A Cervantes","year":"2020","unstructured":"Cervantes, J.-A., L\u00f3pez, S., Rodr\u00edguez, L.-F., Cervantes, S., Cervantes, F., Ramos, F.: Artificial moral agents: a survey of the current status. Sci. Eng. Ethics 26, 501\u2013532 (2020)","journal-title":"Sci. Eng. Ethics"},{"key":"797_CR46","unstructured":"Chang, Y., Wang, X., Wang, J., Wu, Y., Zhu, K., Chen, H., Yang, L., Yi, X., Wang, C., Wang, Y., et\u00a0al.: A survey on evaluation of large language models. arXiv preprint arXiv:2307.03109 (2023)"},{"key":"797_CR47","unstructured":"Chen, C., Wang, X., Jin, Y.: Victor Ye Dong, Li Dong, Jie Cao, Yi Liu, and Rui Yan. Semi-offline reinforcement learning for optimized text generation, In ICML (2023)"},{"key":"797_CR48","doi-asserted-by":"crossref","unstructured":"Chen, D., Zhang, Y., Kundu, S., Li, C., Beerel, P\u00a0A.: Rna-vit: Reduced-dimension approximate normalized attention vision transformers for latency efficient private inference. In 2023 IEEE\/ACM International Conference on Computer Aided Design (ICCAD), pages 1\u20139. IEEE, (2023)","DOI":"10.1109\/ICCAD57390.2023.10323702"},{"key":"797_CR49","doi-asserted-by":"crossref","unstructured":"Chen, T., Bao, H., Huang, S., Dong, L., Jiao, B., Jiang, D., Zhou, H., Li, J., Wei, F.: The-x: Privacy-preserving transformer inference with homomorphic encryption. arXiv preprintarXiv:2206.00216, (2022)","DOI":"10.18653\/v1\/2022.findings-acl.277"},{"key":"797_CR50","unstructured":"Chen, X., Liu, C., Li, B., Lu, K., Song, D.: Targeted backdoor attacks on deep learning systems using data poisoning. arXiv preprint arXiv:1712.05526, (2017)"},{"key":"797_CR51","doi-asserted-by":"crossref","unstructured":"Chen, Y., Wang, R., Jiang, H., Shi, S., Xu, R.: Exploring the use of large language models for reference-free text quality evaluation: A preliminary empirical study. arXiv preprint arXiv:2304.00723, (2023)","DOI":"10.18653\/v1\/2023.findings-ijcnlp.32"},{"key":"797_CR52","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1613\/jair.1.12814","volume":"71","author":"L Cheng","year":"2021","unstructured":"Cheng, L., Varshney, K.R., Liu, H.: Socially responsible ai algorithms: Issues, purposes, and challenges. J. Artif. Intell. Res. 71, 1137\u20131181 (2021)","journal-title":"J. Artif. Intell. Res."},{"issue":"2","key":"797_CR53","doi-asserted-by":"crossref","first-page":"153","DOI":"10.1089\/big.2016.0047","volume":"5","author":"A Chouldechova","year":"2017","unstructured":"Chouldechova, A.: Fair prediction with disparate impact: a study of bias in recidivism prediction instruments. Big data 5(2), 153\u2013163 (2017)","journal-title":"Big data"},{"key":"797_CR54","unstructured":"Christ, M., Gunn, S., Zamir, O.: Undetectable watermarks for language models. ArXiv, abs\/2306.09194, (2023)"},{"key":"797_CR55","unstructured":"Christiano, P\u00a0F., Leike, J., Brown, T., Martic, M., Legg, S., Amodei, D.: Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, (2017)"},{"key":"797_CR56","unstructured":"Chu, H-M., Geiping, J., Fowl, L\u00a0H., Goldblum, M., Goldstein, T.: Panning for gold in federated learning: Targeted text extraction under arbitrarily large-scale aggregation. In The Eleventh International Conference on Learning Representations, (2022)"},{"issue":"1","key":"797_CR57","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1145\/3682112.3682117","volume":"26","author":"Z Chu","year":"2024","unstructured":"Chu, Z., Wang, Z., Zhang, W.: Fairness in large language models: a taxonomic survey. ACM SIGKDD Explorations Newsl 26(1), 34\u201348 (2024)","journal-title":"ACM SIGKDD Explorations Newsl"},{"key":"797_CR58","unstructured":"Chung, H\u00a0W., Hou, L., Longpre, S., Zoph, B., Tay, Y., Fedus, W., Li, E., Wang, X., Dehghani, M., Brahma, S., et\u00a0al.: Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416, (2022)"},{"key":"797_CR59","doi-asserted-by":"crossref","unstructured":"Chung, J., Kamar, E., Amershi, S.: Increasing diversity while maintaining accuracy: Text data generation with large language models and human interventions. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 575\u2013593, Toronto, Canada, (July 2023) Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.acl-long.34"},{"key":"797_CR60","unstructured":"Sumathi, C.P., Santanam, T.: Graduate\u00a0School of\u00a0Science, Sdnb Vaishnav College\u00a0For Women, Chennai, Indian\u00a0Institute of\u00a0Science, DG\u00a0Vaishnav College\u00a0For Men, and India. A study of various steganographic techniques used for information hiding. ArXiv, abs\/1401.5561, (2013)"},{"key":"797_CR61","doi-asserted-by":"crossref","unstructured":"Cramer, R., Damg\u00e5rd, I\u00a0B., et\u00a0al.: Secure multiparty computation. Cambridge University Press, (2015)","DOI":"10.1017\/CBO9781107337756"},{"key":"797_CR62","unstructured":"Crawford, K.: The trouble with bias, Keynote at NeurIPS (2017)"},{"key":"797_CR63","first-page":"5009","volume":"35","author":"G Cui","year":"2022","unstructured":"Cui, G., Yuan, L., He, B., Chen, Y., Liu, Z., Sun, M.: A unified evaluation of textual backdoor learning: Frameworks and benchmarks. Adv. Neural. Inf. Process. Syst. 35, 5009\u20135023 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR64","doi-asserted-by":"crossref","unstructured":"Cummings, R., Desfontaines, D., Evans, D., Geambasu, R., Huang, Y., Jagielski, M., Kairouz, P., Kamath, G., Oh, S., Ohrimenko, O., et\u00a0al.: Advancing differential privacy: Where we are now and future directions for real-world deployment. arXiv preprint arXiv:2304.06929, (2023)","DOI":"10.1162\/99608f92.d3197524"},{"key":"797_CR65","doi-asserted-by":"crossref","unstructured":"Dai, S., Xu, C., Xu, S., Pang, L., Dong, Z., Xu, J.: Bias and unfairness in information retrieval systems: New challenges in the llm era. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 6437\u20136447, (2024)","DOI":"10.1145\/3637528.3671458"},{"issue":"2","key":"797_CR66","doi-asserted-by":"crossref","first-page":"120","DOI":"10.1089\/big.2016.0048","volume":"5","author":"B d\u2019Alessandro","year":"2017","unstructured":"d\u2019Alessandro, B., O\u2019Neil, C., LaGatta, T.: Conscientious classification: a data scientist\u2019s guide to discrimination-aware classification. Big data 5(2), 120\u2013134 (2017)","journal-title":"Big data"},{"key":"797_CR67","doi-asserted-by":"crossref","unstructured":"Delobelle, P., Berendt, B.: Fairdistillation: mitigating stereotyping in language models. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 638\u2013654. Springer, (2022)","DOI":"10.1007\/978-3-031-26390-3_37"},{"key":"797_CR68","first-page":"7659","volume":"34","author":"S Dev","year":"2020","unstructured":"Dev, S., Li, T., Phillips, J.M., Srikumar, V.: On measuring and mitigating biased inferences of word embeddings. Proc. AAAI Conf. Artif. Intell. 34, 7659\u20137666 (2020)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"797_CR69","doi-asserted-by":"crossref","unstructured":"Dhamala, J., Sun, T., Kumar, V., Krishna, S., Pruksachatkun, Y., Chang, K-W., Gupta, R.: Bold: Dataset and metrics for measuring biases in open-ended language generation. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT \u201921, page 862\u2013872, New York, NY, USA, (2021) Association for Computing Machinery","DOI":"10.1145\/3442188.3445924"},{"key":"797_CR70","unstructured":"Dhingra, H., Jayashanker, P., Moghe, S., Strubell, E.: Queer people are people first: Deconstructing sexual identity stereotypes in large language models. arXiv preprint arXiv:2307.00101, (2023)"},{"key":"797_CR71","unstructured":"Doan, T\u00a0V., Chu, Z., Wang, Z., Zhang, W.: Fairness definitions in language models explained. arXiv preprint arXiv:2407.18454, (2024)"},{"key":"797_CR72","doi-asserted-by":"crossref","unstructured":"Dolci, T., Azzalini, F., Tanelli, M.: Improving gender-related fairness in sentence encoders: A semantics-based approach. Data Science and Engineering, pages 1\u201319, (2023)","DOI":"10.1007\/s41019-023-00211-0"},{"key":"797_CR73","unstructured":"Dong, Y., Lu, W-j., Zheng, Y., Wu, H., Zhao, D., Tan, J., Huang, Z., Hong, C., Wei, T., Cheng, W.: Puma: Secure inference of llama-7b in five minutes. arXiv preprint arXiv:2307.12533, (2023)"},{"key":"797_CR74","doi-asserted-by":"crossref","unstructured":"Du, M., Yue, X., Chow, S\u00a0S.M., Wang, T., Huang, C., Sun, H.: Dp-forward: Fine-tuning and inference on language models with differential privacy in forward pass. In Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, pages 2665\u20132679, (2023)","DOI":"10.1145\/3576915.3616592"},{"key":"797_CR75","doi-asserted-by":"crossref","unstructured":"Dwork, C., Hardt, M., Pitassi, T., Reingold, O., Zemel, R.: Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, ITCS \u201912, page 214\u2013226, New York, NY, USA, (2012) Association for Computing Machinery","DOI":"10.1145\/2090236.2090255"},{"key":"797_CR76","doi-asserted-by":"crossref","unstructured":"Dwork, C., Roth, A.: The algorithmic foundations of differential privacy. Foundations and Trends\u00ae in Theoretical Computer Science, 9(3\u20134):211\u2013407, (2014)","DOI":"10.1561\/0400000042"},{"key":"797_CR77","doi-asserted-by":"crossref","unstructured":"Fereidooni, H., Marchal, S., Miettinen, M.,Mirhoseini, A., M\u00f6llering, H., Nguyen, T\u00a0D., Rieger, P., Sadeghi, A-R., Schneider, T., Yalame, H., et\u00a0al.: Safelearn: Secure aggregation for private federated learning. In 2021 IEEE Security and Privacy Workshops (SPW), pages 56\u201362. IEEE, (2021)","DOI":"10.1109\/SPW53761.2021.00017"},{"key":"797_CR78","doi-asserted-by":"crossref","unstructured":"Ferrara, E.: Should chatgpt be biased? challenges and risks of bias in large language models. arXiv preprint arXiv:2304.03738, (2023)","DOI":"10.2139\/ssrn.4627814"},{"key":"797_CR79","unstructured":"Fowl, L., Geiping, J., Reich, S., Wen, Y., Czaja, W., Oldblum, G.,: Decepticons: Corrupted transformers breach privacy in federated learning for language models. arXiv preprint arXiv:2201.12675, (2022)"},{"key":"797_CR80","unstructured":"Fu, Y., Xiong, D., Dong, Y.: Watermarking conditional text generation for ai detection: Unveiling challenges and a semantic-aware watermark remedy. ArXiv, abs\/2307.13808, (2023)"},{"key":"797_CR81","doi-asserted-by":"crossref","unstructured":"Gaci, Y., Benattallah, B., Casati, F., Benabdeslem, K.: Debiasing Pretrained Text Encoders by Paying Attention to Paying Attention. In 2022 Conference on Empirical Methods in Natural Language Processing, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9582\u20139602, Abu Dhabi, United Arab Emirates, (December 2022) Association for Computational Linguistics, Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.emnlp-main.651"},{"key":"797_CR82","unstructured":"Gallegos, I\u00a0O., Rossi, R\u00a0A., Barrow, J., Tanjim, M\u00a0M., Kim, S., Dernoncourt, F., Yu, T., Zhang, R., Ahmed, N\u00a0K.: Bias and fairness in large language models: A survey. arXiv preprintarXiv:2309.00770, (2023)"},{"key":"797_CR83","doi-asserted-by":"crossref","unstructured":"Gallegos, I.O., Rossi, R.A., Joe Barrow, Md., Tanjim, M., Kim, S., Dernoncourt, F., Tong, Yu., Zhang, R., Ahmed, N.K.: Bias and fairness in large language models: A survey. Comput. Linguist. 50(3), 1097\u20131179 (2024)","DOI":"10.1162\/coli_a_00524"},{"key":"797_CR84","doi-asserted-by":"crossref","first-page":"1267","DOI":"10.1109\/TIFS.2023.3333687","volume":"19","author":"K Gao","year":"2023","unstructured":"Gao, K., Bai, J., Baoyuan, W., Ya, M., Xia, S.-T.: Imperceptible and robust backdoor attack in 3d point cloud. IEEE Trans. Inf. Forensics Secur. 19, 1267\u20131282 (2023)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"797_CR85","unstructured":"Gao, K., Bai, Y., Gu, J., Xia, S-T., Torr, P., Li, Z., Liu, W.: Inducing high energy-latency of large vision-language models with verbose images. In International Conference on Learning Representations, (2024)"},{"key":"797_CR86","doi-asserted-by":"crossref","unstructured":"Gao, K., Bai, Y., Gu, J., Yang, Y., Xia, S-T.: Backdoor defense via adaptively splitting poisoned dataset. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pages 4005\u20134014, (2023)","DOI":"10.1109\/CVPR52729.2023.00390"},{"key":"797_CR87","unstructured":"Gao, K., Gu, J., Bai, Y., Xia, S-T., Torr, P., Liu, W., Li, Z.: Energy-latency manipulation of multi-modal large language models via verbose samples. arXiv preprint arXiv:2404.16557, (2024)"},{"key":"797_CR88","doi-asserted-by":"crossref","unstructured":"Gehrmann, S., Strobelt, H., Rush, A\u00a0M.: Gltr: Statistical detection and visualization of generated text. arXiv preprint arXiv:1906.04043, (2019)","DOI":"10.18653\/v1\/P19-3019"},{"key":"797_CR89","unstructured":"Geng, M., He, S., Trotta, R.: Are large language models chameleons? arXiv preprint arXiv:2405.19323, (2024)"},{"issue":"2","key":"797_CR90","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3715073.3715075","volume":"26","author":"C Geren","year":"2025","unstructured":"Geren, C., Board, A., Dagher, G.G., Andersen, T., Zhuang, J.: Blockchain for large language model security and safety: A holistic survey. ACM SIGKDD Explorations Newsl 26(2), 1\u201320 (2025)","journal-title":"ACM SIGKDD Explorations Newsl"},{"key":"797_CR91","doi-asserted-by":"crossref","unstructured":"Ghanbarzadeh, S., Huang, Y., Palangi, H., Cruz\u00a0Moreno, R., Khanpour, H.: Gender-tuning: Empowering fine-tuning for debiasing pre-trained language models. In Findings of the Association for Computational Linguistics: ACL 2023, pages 5448\u20135458, Toronto, Canada, (July 2023) Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.findings-acl.336"},{"key":"797_CR92","doi-asserted-by":"crossref","unstructured":"Gillespie, N., Lockey, S., Curtis, C., Pool, J., Akbari, A.: Trust in artificial intelligence: A global study. (2023)","DOI":"10.14264\/00d3c94"},{"key":"797_CR93","unstructured":"Goldreich, O.: Secure multi-party computation. Manuscript. Preliminary version, 78(110), (1998)"},{"key":"797_CR94","unstructured":"Gu, C., Huang, C., Zheng, X., Chang, K-W., Hsieh, C-J.: Watermarking pre-trained language models with backdooring. arXiv preprint arXiv:2210.07543, (2022)"},{"key":"797_CR95","doi-asserted-by":"crossref","first-page":"47230","DOI":"10.1109\/ACCESS.2019.2909068","volume":"7","author":"G Tianyu","year":"2019","unstructured":"Tianyu, G., Liu, K., Dolan-Gavitt, B., Garg, S.: Badnets: Evaluating backdooring attacks on deep neural networks. IEEE Access 7, 47230\u201347244 (2019)","journal-title":"IEEE Access"},{"key":"797_CR96","unstructured":"Gunasekar, S., Zhang, Y., Aneja, J., Mendes, C\u00a0T., Giorno, Allie Del, Gopi, Sivakanth, Javaheripi, Mojan, Kauffmann, Piero, de\u00a0Rosa, Gustavo, Saarikivi, Olli, et\u00a0al.: Textbooks are all you need. arXiv preprint arXiv:2306.11644, (2023)"},{"key":"797_CR97","doi-asserted-by":"crossref","unstructured":"Guo, W., Caliskan, A.: Detecting emergent intersectional biases: Contextualized word embeddings contain a distribution of human-like biases. In Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society, pages 122\u2013133, (2021)","DOI":"10.1145\/3461702.3462536"},{"key":"797_CR98","doi-asserted-by":"crossref","unstructured":"Gupta, K., Jawalkar, N., Mukherjee, A., Chandran, N., Gupta, D., Panwar, A., Sharma, R.: Sigma: secure gpt inference with function secret sharing. Cryptology ePrint Archive, (2023)","DOI":"10.56553\/popets-2024-0107"},{"key":"797_CR99","first-page":"8130","volume":"35","author":"S Gupta","year":"2022","unstructured":"Gupta, S., Huang, Y., Zhong, Z., Gao, T., Li, K., Chen, D.: Recovering private text in federated learning of language models. Adv. Neural. Inf. Process. Syst. 35, 8130\u20138143 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR100","doi-asserted-by":"crossref","unstructured":"Gupta, U., Dhamala, J., Kumar, V., Verma, A., Pruksachatkun, Y., Krishna, S., Gupta, R., Chang, K-W., Ver\u00a0Steeg, G., Galstyan, A.: Mitigating gender bias in distilled language models via counterfactual role reversal. In Findings of the Association for Computational Linguistics: ACL 2022, pages 658\u2013678, Dublin, Ireland, (May 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.findings-acl.55"},{"key":"797_CR101","doi-asserted-by":"crossref","unstructured":"Hacker, P., Engel, A., Mauer, M.: Regulating chatgpt and other large generative ai models. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pages 1112\u20131123, (2023)","DOI":"10.1145\/3593013.3594067"},{"key":"797_CR102","doi-asserted-by":"crossref","unstructured":"Hallinan, S., Liu, A., Choi, Y., Sap, M.: Detoxifying text with marco: Controllable revision with experts and anti-experts. arXiv preprint arXiv:2212.10543, (2022)","DOI":"10.18653\/v1\/2023.acl-short.21"},{"key":"797_CR103","doi-asserted-by":"crossref","unstructured":"Han, X., Baldwin, T., Cohn, T.: Diverse adversaries for mitigating bias in training. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 2760\u20132765, Online, (April 2021) Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.eacl-main.239"},{"key":"797_CR104","doi-asserted-by":"crossref","unstructured":"Han, X., Baldwin, T., Cohn, T.: Balancing out bias: Achieving fairness through balanced training. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11335\u201311350, Abu Dhabi, United Arab Emirates, (December 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.emnlp-main.779"},{"key":"797_CR105","first-page":"15718","volume":"35","author":"M Hao","year":"2022","unstructured":"Hao, M., Li, H., Chen, H., Xing, P., Guowen, X., Zhang, T.: Iron: Private inference on transformers. Adv. Neural. Inf. Process. Syst. 35, 15718\u201315731 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR106","unstructured":"Hardt, M., Price, E., Srebro, N.: Equality of opportunity in supervised learning. Advances in neural information processing systems, 29, (2016)"},{"key":"797_CR107","doi-asserted-by":"crossref","unstructured":"Hauzenberger, L., Masoudian, S., Kumar, D., Schedl, M., Rekabsaz, N.: Modular and on-demand bias mitigation with attribute-removal subnetworks. In Findings of the Association for Computational Linguistics: ACL 2023, 6192\u20136214 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.386"},{"key":"797_CR108","doi-asserted-by":"crossref","unstructured":"He, S., Han, S., Miao, F.: Robust electric vehicle balancing of autonomous mobility-on-demand system: A multi-agent reinforcement learning approach. In 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 5471\u20135478. IEEE, (2023)","DOI":"10.1109\/IROS55552.2023.10342263"},{"key":"797_CR109","doi-asserted-by":"crossref","unstructured":"He, S., Pepin, L., Wang, G., Zhang, D., Miao, F.: Data-driven distributionally robust electric vehicle balancing for mobility-on-demand systems under demand and supply uncertainties. In 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 2165\u20132172. IEEE, (2020)","DOI":"10.1109\/IROS45743.2020.9341481"},{"key":"797_CR110","doi-asserted-by":"crossref","unstructured":"He, S., Wang, Y., Han, S., Zou, S., Miao, F.: A robust and constrained multi-agent reinforcement learning framework for electric vehicle amod systems. arXiv preprint arXiv:2209.08230, (2022)","DOI":"10.1109\/IROS55552.2023.10342342"},{"key":"797_CR111","doi-asserted-by":"crossref","unstructured":"He, S., Wang, Y., Han, S., Zou, S., Miao, F.: A robust and constrained multi-agent reinforcement learning electric vehicle rebalancing method in amod systems. In 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 5637\u20135644. IEEE, (2023)","DOI":"10.1109\/IROS55552.2023.10342342"},{"key":"797_CR112","doi-asserted-by":"crossref","unstructured":"He, S., Zhang, Z., Han, S., Pepin, L., Wang, G., Zhang, D., Stankovic, J\u00a0A., Miao, F.: Data-driven distributionally robust electric vehicle balancing for autonomous mobility-on-demand systems under demand and supply uncertainties. IEEE Transactions on Intelligent Transportation Systems, (2023)","DOI":"10.1109\/TITS.2023.3237804"},{"key":"797_CR113","doi-asserted-by":"crossref","unstructured":"He, X., Qiongkai, X., Lyu, L., Fangzhao, W., Wang, C.: Protecting intellectual property of language generation apis with lexical watermark. In Proceedings of the AAAI Conference on Artificial Intelligence 36, 10758\u201310766 (2022)","DOI":"10.1609\/aaai.v36i10.21321"},{"key":"797_CR114","doi-asserted-by":"crossref","unstructured":"He, Z., Majumder, B\u00a0P., McAuley, J.: Detect and perturb: Neutral rewriting of biased and sensitive text via gradient-based decoding. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 4173\u20134181, Punta Cana, Dominican Republic, (November 2021) Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.findings-emnlp.352"},{"key":"797_CR115","doi-asserted-by":"crossref","unstructured":"He, Z., Wang, Y., McAuley, J., Majumder, B\u00a0P.: Controlling bias exposure for fair interpretable predictions. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 5854\u20135866, Abu Dhabi, United Arab Emirates, (December 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.findings-emnlp.431"},{"key":"797_CR116","volume-title":"Vinyals, Oriol, Dean","author":"G Hinton","year":"2014","unstructured":"Hinton, G.: Vinyals, Oriol, Dean. Distilling the knowledge in a neural network. In NeurIPS Workshop, Jeff (2014)"},{"key":"797_CR117","doi-asserted-by":"crossref","first-page":"49","DOI":"10.1162\/tacl_a_00299","volume":"8","author":"S Hisamoto","year":"2020","unstructured":"Hisamoto, S.: Post, Matt, Duh, Kevin: Membership inference attacks on sequence-to-sequence models: Is my data in your machine translation system? Trans. Assoc. Comput. Linguistics 8, 49\u201363 (2020)","journal-title":"Trans. Assoc. Comput. Linguistics"},{"key":"797_CR118","doi-asserted-by":"crossref","unstructured":"Hoory, S., Feder, A., Tendler, A., Erell, S., Peled-Cohen, A., Laish, I., Nakhost, H., Stemmer, U., Benjamini, A., Hassidim, A., et al.: Learning and evaluating a differentially private pre-trained language model. In Findings of the Association for Computational Linguistics: EMNLP 2021, 1178\u20131189 (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.102"},{"key":"797_CR119","unstructured":"Hou, A\u00a0B., Zhang, J., He, T., Wang, Y., Chuang, Y-S., Wang, H, Shen, L., Durme, B\u00a0V., Khashabi, D., Tsvetkov, Y.: Semstamp: A semantic watermark with paraphrastic robustness for text generation. ArXiv, abs\/2310.03991, (2023)"},{"key":"797_CR120","doi-asserted-by":"crossref","unstructured":"Hou, A\u00a0B., Zhang, J., Wang, Y., Khashabi, D., He, T.: k-semstamp: A clustering-based semantic watermark for detection of machine-generated text. arXiv preprint arXiv:2402.11399, (2024)","DOI":"10.18653\/v1\/2024.findings-acl.98"},{"key":"797_CR121","unstructured":"Hou, C., Zhan, H., Shrivastava, A., Wang, S., Livshits, S., Fanti, G., Lazar, D.: Privately customizing prefinetuning to better match user data in federated learning. arXiv preprint arXiv:2302.09042, (2023)"},{"key":"797_CR122","unstructured":"Hou, X., Liu, J., Li, J., Li, Y., Wen-jie, L., Hong, C.: and Kui Ren. Secure two-party gpt inference. Cryptology ePrint Archive, Ciphergpt (2023)"},{"key":"797_CR123","unstructured":"Houlsby, N., Giurgiu, A., Jastrzebski, S., Morrone, B., De\u00a0Laroussilhe, Q., Gesmundo, A., Attariyan, M., Gelly, S.: Parameter-efficient transfer learning for nlp. In International Conference on Machine Learning, pages 2790\u20132799. PMLR, (2019)"},{"key":"797_CR124","doi-asserted-by":"crossref","unstructured":"Hu, T., Zhu, W., Yan, Y.: Artificial intelligence aspect of transportation analysis using large scale systems. In 2023 6th Artificial Intelligence and Cloud Computing Conference (AICCC), pages 54\u201359, (2023)","DOI":"10.1145\/3639592.3639600"},{"key":"797_CR125","doi-asserted-by":"crossref","unstructured":"Huang, J., Chang, KC-C.: Towards reasoning in large language models: A survey. arXiv preprint arXiv:2212.10403, (2022)","DOI":"10.18653\/v1\/2023.findings-acl.67"},{"key":"797_CR126","unstructured":"Huang, L., Yu, W., Ma, W., Zhong, W., Feng, Z., Wang, H., Chen, Q., Peng, W., Feng, X., Qin, B., et\u00a0al.: A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. arXiv preprintarXiv:2311.05232, (2023)"},{"key":"797_CR127","doi-asserted-by":"crossref","unstructured":"Huang, P-S, Zhang, H, Jiang, R, Stanforth, R, Welbl, J, Rae, Jack, M, Vishal, Y, Dani, K, Pushmeet: Reducing sentiment bias in language models via counterfactual evaluation. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 65\u201383, Online, November 2020. Association for Computational Linguistics","DOI":"10.18653\/v1\/2020.findings-emnlp.7"},{"key":"797_CR128","unstructured":"Huang, Y, Zhang, Q, Sun, L, et\u00a0al. Trustgpt: A benchmark for trustworthy and responsible large language models. arXiv preprint arXiv:2306.11507, (2023)"},{"key":"797_CR129","unstructured":"Ishibashi, Y, Shimodaira, H: Knowledge sanitization of large language models. arXiv preprint arXiv:2309.11852, (2023)"},{"key":"797_CR130","doi-asserted-by":"crossref","unstructured":"Iskander, S, Radinsky, K, Belinkov, Y: Shielded representations: Protecting sensitive attributes through iterative gradient-based projection. In Findings of the Association for Computational Linguistics: ACL 2023, pages 5961\u20135977, Toronto, Canada, (July 2023) Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.findings-acl.369"},{"key":"797_CR131","doi-asserted-by":"crossref","unstructured":"Jaeger, L, Dacorogna, M: Artificial intelligence from its origins via today to the future: Significant progress in understanding, replicating, and changing us humans or solely technological advances contained to optimising certain processes? In Where Is Science Leading Us? And What Can We Do to Steer It?, pages 207\u2013235. Springer, (2024)","DOI":"10.1007\/978-3-031-47138-4_10"},{"key":"797_CR132","doi-asserted-by":"crossref","unstructured":"Jain, N, Popovi\u0107, M, Groves, D, Vanmassenhove, E: Generating gender augmented data for NLP. In Proceedings of the 3rd Workshop on Gender Bias in Natural Language Processing, pages 93\u2013102, Online, (August 2021) Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.gebnlp-1.11"},{"key":"797_CR133","doi-asserted-by":"crossref","unstructured":"Jalil, Z, Mirza, A\u00a0M., Jabeen, H: Word length based zero-watermarking algorithm for tamper detection in text documents. In 2010 2nd International Conference on Computer Engineering and Technology, volume\u00a06, pages V6\u2013378. IEEE, (2010)","DOI":"10.1109\/ICCET.2010.5486185"},{"key":"797_CR134","unstructured":"Jalil, Z, Mirza, A\u00a0M., Sabir, M: Content based zero-watermarking algorithm for authentication of text documents. arXiv preprint arXiv:1003.1796, (2010)"},{"key":"797_CR135","unstructured":"Ji, J, Liu, M, Dai, J, Pan, X, Zhang, C, Bian, C, Sun, R, Wang, Y, Yang, Y: Beavertails: Towards improved safety alignment of llm via a human-preference dataset. arXiv preprint arXiv:2307.04657, (2023)"},{"key":"797_CR136","unstructured":"Jiang, J, Liu, X, Fan, C: Low-parameter federated learning with large language models. arXiv preprint arXiv:2307.13896, (2023)"},{"key":"797_CR137","unstructured":"Jin, H, Hu, L, Li, X, Zhang, P, Chen, C, Zhuang, J, Wang, H: Jailbreakzoo: Survey, landscapes, and horizons in jailbreaking large language and vision-language models. arXiv preprint arXiv:2407.01599, (2024)"},{"key":"797_CR138","unstructured":"Jin, X, Larson, J, Yang, W, Lin, Z: Binary code summarization: Benchmarking chatgpt\/gpt-4 and other large language models. arXiv preprint arXiv:2312.09601, (2023)"},{"key":"797_CR139","doi-asserted-by":"crossref","unstructured":"Jin, Y, Chandra, M, Verma, G, Hu, Y, De\u00a0Choudhury, M, Kumar, S: Better to ask in english: Cross-lingual evaluation of large language models for healthcare queries. arXiv:2310.13132, (2023)","DOI":"10.1145\/3589334.3645643"},{"key":"797_CR140","doi-asserted-by":"crossref","unstructured":"Jin, Y, Choi, M, Verma, G, Wang, J, Kumar, S: Mm-soc: Benchmarking multimodal large language models in social media platforms. arXiv:2402.14154, (2024)","DOI":"10.18653\/v1\/2024.findings-acl.370"},{"key":"797_CR141","doi-asserted-by":"crossref","unstructured":"Jin, Y., Wang, X., Yang, R., Sun, Y., Wang, W., Liao, H., Xie, X.: Towards fine-grained reasoning for fake news detection. In Proceedings of the AAAI Conference on Artificial Intelligence 36, 5746\u20135754 (2022)","DOI":"10.1609\/aaai.v36i5.20517"},{"key":"797_CR142","doi-asserted-by":"crossref","unstructured":"Kairouz, P., Brendan MH., Avent, B., Bellet, A., Bennis, M., Bhagoji, A\u00a0N., Bonawitz, K., Charles, Z., Cormode, G., Cummings, R., et\u00a0al: Advances and open problems in federated learning. Foundations and Trends\u00ae in Machine Learning, 14(1\u20132):1\u2013210, (2021)","DOI":"10.1561\/2200000083"},{"issue":"1","key":"797_CR143","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s10115-011-0463-8","volume":"33","author":"F Kamiran","year":"2012","unstructured":"Kamiran, F., Calders, T.: Data preprocessing techniques for classification without discrimination. Knowl. Inf. Syst. 33(1), 1\u201333 (2012)","journal-title":"Knowl. Inf. Syst."},{"key":"797_CR144","unstructured":"Kandpal, N., Jagielski, M., Tram\u00e8r, F., Carlini, N.: Backdoor attacks for in-context learning with language models. arXiv preprint arXiv:2307.14692, (2023)"},{"key":"797_CR145","unstructured":"Kandpal, N., Wallace, E., Raffel, C.: Deduplicating training data mitigates privacy risks in language models. In International Conference on Machine Learning, pages 10697\u201310707. PMLR, (2022)"},{"key":"797_CR146","first-page":"11954","volume":"36","author":"M Kaneko","year":"2022","unstructured":"Kaneko, M., Bollegala, D.: Unmasking the mask-evaluating social biases in masked language models. Proc. AAAI Conf. Artif. Intell. 36, 11954\u201311962 (2022)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"key":"797_CR147","doi-asserted-by":"crossref","unstructured":"Kang, C., Choi, J.: Impact of co-occurrence on factual knowledge of large language models. arXiv preprint arXiv:2310.08256, (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.518"},{"key":"797_CR148","doi-asserted-by":"crossref","unstructured":"Kharitonov, E., Lee, A., Polyak, A., Adi, Y., Copet, J., Lakhotia, K., Nguyen, T-A., Rivi\u00e8re, M., Mohamed, A., Dupoux, E., et\u00a0al. Text-free prosody-aware generative spoken language modeling. arXiv preprint arXiv:2109.03264, (2021)","DOI":"10.18653\/v1\/2022.acl-long.593"},{"key":"797_CR149","doi-asserted-by":"crossref","unstructured":"Kim, H., Yu, Y., Jiang, L., Lu, X., Khashabi, D., Kim, Gunhee, C., Yejin, S., M.: ProsocialDialog: A prosocial backbone for conversational agents. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 4005\u20134029, Abu Dhabi, United Arab Emirates, (December 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.emnlp-main.267"},{"key":"797_CR150","doi-asserted-by":"crossref","unstructured":"Kim, M\u00a0P., Ghorbani, A., Zou, J.: Multiaccuracy: Black-box post-processing for fairness in classification. In Proceedings of the 2019 AAAI\/ACM Conference on AI, Ethics, and Society, pages 247\u2013254, (2019)","DOI":"10.1145\/3306618.3314287"},{"key":"797_CR151","unstructured":"Kirchenbauer, J., Geiping, J., Wen, Y., Katz, J., Miers, I., Goldstein, T.: A watermark for large language models. arXiv preprint arXiv:2301.10226, (2023)"},{"key":"797_CR152","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1016\/j.csbj.2014.11.005","volume":"13","author":"K Kourou","year":"2015","unstructured":"Kourou, K., Exarchos, T.P., Exarchos, K.P., Karamouzis, M.V., Fotiadis, D.I.: Machine learning applications in cancer prognosis and prediction. Comput. Struct. Biotechnol. J. 13, 8\u201317 (2015)","journal-title":"Comput. Struct. Biotechnol. J."},{"key":"797_CR153","doi-asserted-by":"crossref","unstructured":"Kuang, W., Qian, B., Li, Z., Chen, D., Gao, D., Pan, X., Xie, Y., Li, Y., Ding, B., Zhou, J.: Federatedscope-llm: A comprehensive package for fine-tuning large language models in federated learning. arXiv preprint arXiv:2309.00363 (2023)","DOI":"10.1145\/3637528.3671573"},{"key":"797_CR154","unstructured":"Kuditipudi, R., Thickstun, J., Hashimoto, T., Liang, P.: Robust distortion-free watermarks for language models. ArXiv, abs\/2307.15593, (2023)"},{"key":"797_CR155","doi-asserted-by":"crossref","first-page":"1125083","DOI":"10.3389\/fdata.2023.1125083","volume":"6","author":"S Kumar","year":"2023","unstructured":"Kumar, S.: Advances in ai for web integrity, equity, and well-being. Front. Big Data 6, 1125083 (2023)","journal-title":"Front. Big Data"},{"key":"797_CR156","doi-asserted-by":"crossref","unstructured":"Lai, J., Gan, W., Wu, J., Qi, Z., Yu, P\u00a0S.: Large language models in law: A survey. arXiv preprint arXiv:2312.03718, (2023)","DOI":"10.1016\/j.aiopen.2024.09.002"},{"key":"797_CR157","doi-asserted-by":"crossref","unstructured":"Lee, K., Ippolito, D., Nystrom, A., Zhang, C., Eck, D., Callison-Burch, C., Carlini, N.: Deduplicating training data makes language models better. arXiv preprint arXiv:2107.06499, (2021)","DOI":"10.18653\/v1\/2022.acl-long.577"},{"key":"797_CR158","first-page":"34586","volume":"35","author":"N Lee","year":"2022","unstructured":"Lee, N., Ping, W., Xu, P., Patwary, M., Fung, P.N., Shoeybi, M., Catanzaro, B.: Factuality enhanced language models for open-ended text generation. Adv. Neural. Inf. Process. Syst. 35, 34586\u201334599 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR159","unstructured":"Li, D., Shao, R., Wang, H., Guo, H., Xing, E\u00a0P., Zhang, H.: Mpcformer: fast, performant and private transformer inference with mpc. arXiv preprint arXiv:2211.01452, (2022)"},{"key":"797_CR160","doi-asserted-by":"crossref","unstructured":"Li, H., Feng, Y., Xue, S., Liu, X., Zeng, B., Li, S., Liu, B., Liu, J., Han, S., Zhang, B.: Uv-idm: Identity-conditioned latent diffusion model for face uv-texture generation. In CVPR, pages 10585\u201310595, (2024)","DOI":"10.1109\/CVPR52733.2024.01007"},{"key":"797_CR161","unstructured":"Li, J., Sun, S., Yuan, W., Fan, R-Z., Zhao, H., Liu, P.: Generative judge for evaluating alignment. arXiv preprintarXiv:2310.05470, (2023)"},{"key":"797_CR162","doi-asserted-by":"crossref","unstructured":"Li, J., Cheng, X., Zhao, W\u00a0X., Nie, J-Y., Wen, J-R.: Halueval: A large-scale hallucination evaluation benchmark for large language models. In EMNLP, pages 6449\u20136464, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.397"},{"key":"797_CR163","doi-asserted-by":"crossref","unstructured":"Li, M., Wang, J., Wang, J., Neel, S.: Mope: Model perturbation-based privacy attacks on language models. arXiv preprint arXiv:2310.14369, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.842"},{"key":"797_CR164","doi-asserted-by":"crossref","unstructured":"Li, X\u00a0L., Liang, P.: Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582\u20134597, Online, (August 2021) Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"797_CR165","unstructured":"Li, X., Tramer, F., Liang, P., Hashimoto, T.: Large language models can be strong differentially private learners. arXiv preprint arXiv:2110.05679, (2021)"},{"key":"797_CR166","unstructured":"Li, Y., Tan, Z., Liu, Y.: Privacy-preserving prompt tuning for large language model services. arXiv preprint arXiv:2305.06212, (2023)"},{"key":"797_CR167","unstructured":"Li, Y., Jiang, Y., Li, Z., and Shu-Tao X.: A survey. IEEE Transactions on Neural Networks and Learning Systems, Backdoor learning (2022)"},{"key":"797_CR168","doi-asserted-by":"crossref","unstructured":"Li, Y., Du, M., Wang, X., Wang, Y.: Prompt tuning pushes farther, contrastive learning pulls closer: A two-stage approach to mitigate social biases. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 14254\u201314267, Toronto, Canada, July 2023. Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.acl-long.797"},{"key":"797_CR169","unstructured":"Li, Z., Zhang, S., Zhao, H., Yang, Y., Yang, D.: Batgpt: A bidirectional autoregessive talker from generative pre-trained transformer. arXiv preprint arXiv:2307.00360, (2023)"},{"key":"797_CR170","doi-asserted-by":"crossref","DOI":"10.1016\/j.techsoc.2023.102442","volume":"76","author":"Y Lian","year":"2024","unstructured":"Lian, Y., Tang, H., Xiang, M., Dong, X.: Public attitudes and sentiments toward chatgpt in china: A text mining analysis based on social media. Technol. Soc. 76, 102442 (2024)","journal-title":"Technol. Soc."},{"key":"797_CR171","unstructured":"Liang, P\u00a0P., Wu, C., Morency, L-P., Salakhutdinov, R.: Towards understanding and mitigating social biases in language models. In International Conference on Machine Learning, pages 6565\u20136576. PMLR, (2021)"},{"key":"797_CR172","unstructured":"Liang, P., Bommasani, R., Lee, T., Tsipras, D., Soylu, D., Yasunaga, M., Zhang, Y., Narayanan, D., Wu, Y., Kumar, A., et\u00a0al. Holistic evaluation of language models. arXiv preprint arXiv:2211.09110, 2022"},{"key":"797_CR173","unstructured":"Liang, Z., Wang, P., Zhang, R., Xu, N., Zhang, S.: Merge: Fast private text generation. arXiv preprint arXiv:2305.15769, (2023)"},{"key":"797_CR174","unstructured":"Liu, A., Pan, L, Hu, X., Li, S., Wen, L., King, I., Yu, P\u00a0S.: A private watermark for large language models. ArXiv, abs\/2307.16230, (2023)"},{"key":"797_CR175","unstructured":"Liu, F., Guan, T., Li, Z., Chen, L., Yacoob, Y., Manocha, D., Zhou, T.: Hallusionbench: You see what you think? or you think what you see? an image-context reasoning benchmark challenging for gpt-4v (ision), llava-1.5, and other multi-modality models. arXiv:2310.14566, (2023)"},{"key":"797_CR176","unstructured":"Liu, F., Lin, K., Li, L., Wang, J., Yacoob, Y., Wang, L.: Mitigating hallucination in large multi-modal models via robust instruction tuning. In The Twelfth International Conference on Learning Representations, (2023)"},{"key":"797_CR177","doi-asserted-by":"crossref","unstructured":"Liu, F., Wang, X., Yao, W., Chen, J., Song, K., Cho, S., Yacoob, Y., Yu, D.: Mmc: Advancing multimodal chart understanding with large-scale instruction tuning. arXiv preprint arXiv:2311.10774, (2023)","DOI":"10.18653\/v1\/2024.naacl-long.70"},{"key":"797_CR178","doi-asserted-by":"crossref","unstructured":"Liu, F., Yacoob, Y., Shrivastava, A.: Covid-vts: Fact extraction and verification on short video platforms. arXiv preprint arXiv:2302.07919, (2023)","DOI":"10.18653\/v1\/2023.eacl-main.14"},{"key":"797_CR179","doi-asserted-by":"crossref","first-page":"1024","DOI":"10.1109\/TIFS.2022.3155921","volume":"17","author":"G Liu","year":"2022","unstructured":"Liu, G., Tianlong, X., Ma, X., Wang, C.: Your model trains on my data? protecting intellectual property of training data via membership fingerprint authentication. IEEE Trans. Inf. Forensics Secur. 17, 1024\u20131037 (2022)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"797_CR180","doi-asserted-by":"crossref","unstructured":"Liu, H., Dacon, J., Fan, W., Liu, H., Liu, Z., Tang, J.: Does gender matter? towards fairness in dialogue systems. arXiv preprint arXiv:1910.10486, (2019)","DOI":"10.18653\/v1\/2020.coling-main.390"},{"key":"797_CR181","doi-asserted-by":"crossref","unstructured":"Liu, R., Jia, C., Wei, J., Guangxuan, X., Wang, L., Vosoughi, S.: Mitigating political bias in language models through reinforced calibration. In Proceedings of the AAAI Conference on Artificial Intelligence 35, 14857\u201314866 (2021)","DOI":"10.1609\/aaai.v35i17.17744"},{"key":"797_CR182","doi-asserted-by":"crossref","unstructured":"Liu, S., Liu, X., Wang, Y., Cheng, Z., Li, C., Zhang, Z., Lan, Y., Shen, C.: Does[CDATA[\\backslash ]]$$\\backslash $$textsc $$\\{$$[CDATA[\\{]]DetectGPT$$\\}$$[CDATA[\\}]] fully utilize perturbation? selective perturbation on model-based contrastive learning detector would be better. arXiv preprint arXiv:2402.00263, (2024)","DOI":"10.18653\/v1\/2024.acl-long.103"},{"key":"797_CR183","doi-asserted-by":"crossref","unstructured":"Liu, X., Zheng, Y., Zhengxiao D., Ming D., Zhilin Y., and Jie T.: Gpt understands, too. AI Open, Yujie Qian (2023)","DOI":"10.1016\/j.aiopen.2023.08.012"},{"key":"797_CR184","doi-asserted-by":"crossref","unstructured":"Liu, X., Zhang, Z., Wang, Y., Pu, H., Lan, Y., Shen, C.: Coco: Coherence-enhanced machine-generated text detection under low resource with contrastive learning. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 16167\u201316188, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.1005"},{"key":"797_CR185","unstructured":"Liu, Y., Yao, Y., Ton, J-F., Zhang, X., Cheng, RG\u00a0H., Klochkov, Y., Taufiq, M\u00a0F., Li, H.: Trustworthy llms: a survey and guideline for evaluating large language models\u2019 alignment. arXiv preprint arXiv:2308.05374, (2023)"},{"key":"797_CR186","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: RoBERTa: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, (2019)"},{"key":"797_CR187","unstructured":"Liu, Y., Hu, H., Zhang, X., Sun, L.: Watermarking text data on large language models for dataset copyright protection. arXiv preprint arXiv:2305.13257, (2023)"},{"key":"797_CR188","doi-asserted-by":"crossref","unstructured":"Liu, Y., Li, K., Liu, Z., Wen, B., Ke, X., Wang, W., Zhao, W., Li, Q.: Provenance of training without training data: Towards privacy-preserving dnn model ownership verification. In Proceedings of the ACM web conference 2023, 1980\u20131990 (2023)","DOI":"10.1145\/3543507.3583198"},{"key":"797_CR189","doi-asserted-by":"crossref","unstructured":"Liu, Y., Jia, J., Liu, H., Gong, N\u00a0Z.: Stolenencoder: stealing pre-trained encoders in self-supervised learning. In Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, pages 2115\u20132128, (2022)","DOI":"10.1145\/3548606.3560586"},{"key":"797_CR190","doi-asserted-by":"crossref","unstructured":"Liu, Z., Dou, G., Tian, Y., Zhang, C., Chien, E., Zhu, Z.: Breaking the trilemma of privacy, utility, efficiency via controllable machine unlearning. arXiv preprint arXiv:2310.18574, (2023)","DOI":"10.1145\/3589334.3645669"},{"key":"797_CR191","first-page":"27591","volume":"35","author":"L Ximing","year":"2022","unstructured":"Ximing, L., Welleck, S., Hessel, J., Jiang, L., Qin, L., West, P., Ammanabrolu, P., Choi, Y.: Quark: Controllable text generation with reinforced unlearning. Adv. Neural. Inf. Process. Syst. 35, 27591\u201327609 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR192","doi-asserted-by":"crossref","unstructured":"Lucas, E., Havens, T.: Gpts don\u2019t keep secrets: Searching for backdoor watermark triggers in autoregressive language models. In Proceedings of the 3rd Workshop on Trustworthy Natural Language Processing (TrustNLP 2023), pages 242\u2013248, (2023)","DOI":"10.18653\/v1\/2023.trustnlp-1.21"},{"key":"797_CR193","doi-asserted-by":"crossref","unstructured":"Lyu, W., Lin, X., Zheng, S., Pang, L., Ling, H., Jha, S., Chen, C.: Task-agnostic detector for insertion-based backdoor attacks. arXiv preprint arXiv:2403.17155, (2024)","DOI":"10.18653\/v1\/2024.findings-naacl.179"},{"key":"797_CR194","unstructured":"Lyu, W., Zheng, S., Ling, H., Chen, C.: Backdoor attacks against transformers with attention enhancement. In ICLR 2023 Workshop on Backdoor Attacks and Defenses in Machine Learning, (2023)"},{"key":"797_CR195","doi-asserted-by":"crossref","unstructured":"Lyu, W., Zheng, S., Ma, T., Chen, C.: A study of the attention abnormality in trojaned berts. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4727\u20134741, (2022)","DOI":"10.18653\/v1\/2022.naacl-main.348"},{"key":"797_CR196","doi-asserted-by":"crossref","unstructured":"Lyu, W., Songzhu Zheng, L., Pang, H.L., Chen, C.: Attention-enhancing backdoor attacks against bert-based models. In Findings of the Association for Computational Linguistics: EMNLP 2023, 10672\u201310690 (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.716"},{"key":"797_CR197","unstructured":"Mai, P., Yan, R., Huang, Z., Yang, Y., Pang, Y.: Split-and-denoise: Protect large language model inference with local differential privacy. arXiv preprint arXiv:2310.09130, (2023)"},{"key":"797_CR198","unstructured":"Maini, P., Yaghini, M., Papernot, N.: Dataset inference: Ownership resolution in machine learning. arXiv preprint arXiv:2104.10706, (2021)"},{"key":"797_CR199","unstructured":"Majmudar, J., Dupuy, C., Peris, C., Smaili, S., Gupta, R., Zemel, R.: Differentially private decoding in large language models. arXiv preprint arXiv:2205.13621, (2022)"},{"key":"797_CR200","unstructured":"Mao, Y., Wang, W., Du, H., Guan, N., Xue, C\u00a0J.: On the compressibility of quantized large language models. arXiv preprint arXiv:2403.01384, (2024)"},{"key":"797_CR201","doi-asserted-by":"crossref","unstructured":"Mattern, J., Mireshghallah, F., Jin, Z., Sch\u00f6lkopf, B., Sachan, M., Berg-Kirkpatrick, T.: Membership inference attacks against language models via neighbourhood comparison. arXiv preprint arXiv:2305.18462, (2023)","DOI":"10.18653\/v1\/2023.findings-acl.719"},{"key":"797_CR202","unstructured":"Maudslay, R\u00a0H., Gonen, H., Cotterell, R., Teufel, S.: It\u2019s all in the name: Mitigating gender bias with name-based counterfactual data substitution. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5267\u20135275, (2019)"},{"key":"797_CR203","doi-asserted-by":"crossref","unstructured":"May, C., Wang, A., Bordia, S., Bowman, S\u00a0R., Rudinger, R.; On measuring social biases in sentence encoders. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 622\u2013628, Minneapolis, Minnesota, (June 2019) Association for Computational Linguistics","DOI":"10.18653\/v1\/N19-1063"},{"key":"797_CR204","doi-asserted-by":"crossref","unstructured":"Thomas McCoy, R., Smolensky, P., Linzen, T., Gao, J., Celikyilmaz, A.: How much do language models copy from their training data? evaluating linguistic novelty in text generation using raven. Transactions of the Association for Computational Linguistics 11, 652\u2013670 (2023)","DOI":"10.1162\/tacl_a_00567"},{"key":"797_CR205","unstructured":"McMahan, H\u00a0B., Ramage, D., Talwar, K., Zhang, L.: Learning differentially private recurrent language models. arXiv preprint arXiv:1710.06963, (2017)"},{"key":"797_CR206","doi-asserted-by":"crossref","unstructured":"Meade, N., Gella, S., Hazarika, D., Gupta, P., Jin, D., Reddy, S., Liu, Y., Hakkani-T\u00fcr, D.: Using in-context learning to improve dialogue safety. arXiv preprint arXiv:2302.00871, (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.796"},{"key":"797_CR207","doi-asserted-by":"crossref","unstructured":"M\u011bchura, M.: A taxonomy of bias-causing ambiguities in machine translation. In Proceedings of the 4th Workshop on Gender Bias in Natural Language Processing (GeBNLP), pages 168\u2013173, Seattle, Washington, (July 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.gebnlp-1.18"},{"issue":"7","key":"797_CR208","doi-asserted-by":"crossref","first-page":"1658","DOI":"10.1038\/s41591-023-02439-9","volume":"29","author":"M Meeus","year":"2023","unstructured":"Meeus, M., Jain, S., de Montjoye, Y.-A.: Concerns about using a digital mask to safeguard patient privacy. Nat. Med. 29(7), 1658\u20131659 (2023)","journal-title":"Nat. Med."},{"issue":"6","key":"797_CR209","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3457607","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., Galstyan, A.: A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 54(6), 1\u201335 (2021)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"797_CR210","doi-asserted-by":"crossref","unstructured":"Melis, L., Song, C., De\u00a0Cristofaro, E., Shmatikov, V.: Exploiting unintended feature leakage in collaborative learning. In 2019 IEEE symposium on security and privacy (SP), pages 691\u2013706. IEEE, (2019).","DOI":"10.1109\/SP.2019.00029"},{"key":"797_CR211","doi-asserted-by":"crossref","unstructured":"Mesk\u00f3, B., Topol, E.J.: The imperative for regulatory oversight of large language models (or generative ai) in healthcare. npj Digital Medicine, 6(1):120, (2023)","DOI":"10.1038\/s41746-023-00873-0"},{"key":"797_CR212","doi-asserted-by":"crossref","unstructured":"Minssen, T., Vayena, E., Cohen, I.G.: The challenges for regulating medical use of chatgpt and other large language models. Jama, (2023)","DOI":"10.1001\/jama.2023.9651"},{"key":"797_CR213","doi-asserted-by":"crossref","first-page":"648","DOI":"10.1016\/j.chb.2013.07.040","volume":"30","author":"N Mir","year":"2014","unstructured":"Mir, N.: Copyright for web content using invisible text watermarking. Comput. Hum. Behav. 30, 648\u2013653 (2014)","journal-title":"Comput. Hum. Behav."},{"key":"797_CR214","doi-asserted-by":"crossref","unstructured":"Mireshghallah, F., Goyal, K., Uniyal, A., Berg-Kirkpatrick, T., Shokri, R.: Quantifying privacy risks of masked language models using membership inference attacks. arXiv preprint arXiv:2203.03929, (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.570"},{"key":"797_CR215","unstructured":"Mitchell, E., Lee, Y., Khazatsky, A., Manning, C.D., Finn, C.: Detectgpt: Zero-shot machine-generated text detection using probability curvature. In International Conference on Machine Learning, pages 24950\u201324962. PMLR, (2023)"},{"key":"797_CR216","unstructured":"Mitchell, E., Lin, C., Bosselut, A., Manning, C.D., Finn, C.: Memory-based model editing at scale. In International Conference on Machine Learning, pages 15817\u201315831. PMLR, (2022)"},{"key":"797_CR217","unstructured":"Mo, Y., Qin, H., Dong, Y., Zhu, Z., Li, Z.: Large language model (llm) ai text generation detection based on transformer deep learning algorithm. arXiv preprint arXiv:2405.06652, (2024)"},{"issue":"8","key":"797_CR218","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0237861","volume":"15","author":"M Mozafari","year":"2020","unstructured":"Mozafari, M., Farahbakhsh, R., Crespi, N.: Hate speech detection and racial bias mitigation in social media based on bert model. PLoS ONE 15(8), e0237861 (2020)","journal-title":"PLoS ONE"},{"key":"797_CR219","unstructured":"Mugunthan, V., Polychroniadou, A., Byrd, D., Balch, T.H.: Smpai: Secure multi-party computation for federated learning. In Proceedings of the NeurIPS 2019 Workshop on Robust AI in Financial Services, pages 1\u20139. MIT Press Cambridge, MA, USA, (2019)"},{"key":"797_CR220","doi-asserted-by":"crossref","unstructured":"Nangia, N., Vania, C., Bhalerao, R., Bowman, S\u00a0R.: CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pages 1953\u20131967, Online, (November 2020) Association for Computational Linguistics","DOI":"10.18653\/v1\/2020.emnlp-main.154"},{"key":"797_CR221","doi-asserted-by":"crossref","unstructured":"Venkit, PN., Gautam, S., Panchanadikar, R., Huang, T-H., Wilson, S.: Nationality bias in text generation. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 116\u2013122, Dubrovnik, Croatia, (May 2023) Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.eacl-main.9"},{"key":"797_CR222","unstructured":"Nasr, M., Carlini, N., Hayase, J., Jagielski, M., Cooper, A\u00a0F., Ippolito, D., Choquette-Choo, C\u00a0A., Wallace, E., Tram\u00e8r, F., Lee, K.: Scalable extraction of training data from (production) language models. arXiv preprint arXiv:2311.17035, (2023)"},{"key":"797_CR223","doi-asserted-by":"crossref","unstructured":"Neo, NK\u00a0N., Lee, Y-C., Jin, Y., Kim, S-W., Kumar, S.: Towards fair graph anomaly detection: Problem, new datasets, and evaluation. arXiv:2402.15988, (2024)","DOI":"10.1145\/3627673.3679754"},{"key":"797_CR224","unstructured":"Nguyen, T.A.: Tran. Input-aware dynamic backdoor attack. In Advances in Neural Information Processing Systems, Anh (2020)"},{"key":"797_CR225","doi-asserted-by":"crossref","unstructured":"Nordemann, J\u00a0B., Pukas, J.: Copyright exceptions for ai training data\u2013will there be an international level playing field?, (2022)","DOI":"10.1093\/jiplp\/jpac106"},{"key":"797_CR226","unstructured":"Nori, H., King, N., McKinney, S\u00a0M., Dean C., and Eric H.: Capabilities of gpt-4 on medical challenge problems. arXiv:2303.13375, (2023)"},{"key":"797_CR227","doi-asserted-by":"crossref","unstructured":"Nozza, D., Bianchi, F., Hovy, D.: HONEST: Measuring hurtful sentence completion in language models. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2398\u20132406, Online, (June 2021) Association for Computational Linguistics, Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.naacl-main.191"},{"key":"797_CR228","doi-asserted-by":"crossref","unstructured":"Oh, C., Won, H., So, J., Kim, T., Kim, Y., Choi, H., Song, K.: Learning fair representation via distributional contrastive disentanglement. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 1295\u20131305, (2022)","DOI":"10.1145\/3534678.3539232"},{"key":"797_CR229","doi-asserted-by":"crossref","unstructured":"Orgad, H., Belinkov, Y.: BLIND: Bias removal with no demographics. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8801\u20138821, Toronto, Canada, (July 2023) Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.acl-long.490"},{"key":"797_CR230","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang, L., Jeffrey, W., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., et al.: Training language models to follow instructions with human feedback. Adv. Neural. Inf. Process. Syst. 35, 27730\u201327744 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR231","doi-asserted-by":"crossref","unstructured":"Pan, L., Saxon, M., Xu, W., Nathani, D., Wang, X., Wang, W\u00a0Y.: Automatically correcting large language models: Surveying the landscape of diverse self-correction strategies. arXiv preprint arXiv:2308.03188, (2023)","DOI":"10.1162\/tacl_a_00660"},{"key":"797_CR232","doi-asserted-by":"crossref","unstructured":"Park, S., Choi, K., Yu, H., Ko, Y.: Never too late to learn: Regularizing gender bias in coreference resolution. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, WSDM \u201923, page 15\u201323, New York, NY, USA, (2023). Association for Computing Machinery","DOI":"10.1145\/3539597.3570473"},{"key":"797_CR233","doi-asserted-by":"crossref","unstructured":"Parrish, A., Chen, A., Nangia, N., Padmakumar, V., Phang, J., Thompson, J., Htut, P\u00a0M., Bowman, S.: BBQ: A hand-built bias benchmark for question answering. In Findings of the Association for Computational Linguistics: ACL 2022, pages 2086\u20132105, Dublin, Ireland, (May 2022) Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.findings-acl.165"},{"key":"797_CR234","doi-asserted-by":"crossref","unstructured":"Patroc\u00ednio, J., Santana, D.B., de Andrade.: Artificial intelligence, algorithmic recommendation and decision-making in european union law: analysis of the regulatory challenge and legal certainty. Latin American Center of European Studies 3(2), 136\u2013179 (2023)","DOI":"10.51799\/2763-8685v3n2005"},{"key":"797_CR235","unstructured":"Pawelczyk, M., Neel, S., Lakkaraju, H.: In-context unlearning: Language models as few shot unlearners. arXiv preprint arXiv:2310.07579, (2023)"},{"key":"797_CR236","unstructured":"Peng, B., Galley, M., He, P., Cheng, H., Xie, Y., Hu, Y., Huang, Q., Liden, L., Yu, Z., Chen, W., et\u00a0al.: Check your facts and try again: Improving large language models with external knowledge and automated feedback. arXiv preprint arXiv:2302.12813, (2023)"},{"key":"797_CR237","doi-asserted-by":"crossref","unstructured":"Peng, W., Yi, J., Wu, F., Wu, S., Zhu, B., Lyu, L., Jiao, B., Xu, T., Sun, G., Xie, X.: Are you copying my model? protecting the copyright of large language models for eaas via backdoor watermark. arXiv preprint arXiv:2305.10036, (2023)","DOI":"10.18653\/v1\/2023.acl-long.423"},{"issue":"2","key":"797_CR238","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3639368","volume":"15","author":"D Pessach","year":"2024","unstructured":"Pessach, D., Tassa, T., Shmueli, E.: Fairness-driven private collaborative machine learning. ACM Trans. Intell. Syst. Technol. 15(2), 1\u201330 (2024)","journal-title":"ACM Trans. Intell. Syst. Technol."},{"key":"797_CR239","first-page":"25944","volume":"34","author":"F Petersen","year":"2021","unstructured":"Petersen, F., Mukherjee, D., Sun, Y., Yurochkin, M.: Post-processing for individual fairness. Adv. Neural. Inf. Process. Syst. 34, 25944\u201325955 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"7","key":"797_CR240","doi-asserted-by":"crossref","first-page":"1062","DOI":"10.1109\/5.771065","volume":"87","author":"F Petitcolas","year":"1999","unstructured":"Petitcolas, F., Anderson, R.J., Kuhn, M.G.: Information hiding-a survey. Proc. IEEE 87(7), 1062\u20131078 (1999)","journal-title":"Proc. IEEE"},{"key":"797_CR241","unstructured":"Pollina, E., Mukherjee, S.: Italy curbs chatgpt, starts probe over privacy concerns. Reuters, available at: https:\/\/www. reuters. com\/technology\/italy-data-protectionagency-opens-chatgpt-probe-privacyconcerns-2023-03-31\/(accessed 2nd October, 2024), (2023)"},{"key":"797_CR242","doi-asserted-by":"crossref","unstructured":"Qian, R., Ross, C., Fernandes, J., Smith, E\u00a0M., Kiela, D., Williams, A.: Perturbation augmentation for fairer NLP. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9496\u20139521, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.emnlp-main.646"},{"key":"797_CR243","doi-asserted-by":"crossref","unstructured":"Qian, Y., Muaz, U., Zhang, B., Hyun, J\u00a0W.: Reducing gender bias in word-level language models with a gender-equalizing loss function. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, pages 223\u2013228, Florence, Italy, July 2019. Association for Computational Linguistics","DOI":"10.18653\/v1\/P19-2031"},{"key":"797_CR244","unstructured":"Radford, A., Karthik N., Ilya S., et\u00a0al. Improving language understanding by generative pre-training, Tim Salimans (2018)"},{"issue":"8","key":"797_CR245","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Jeffrey, W., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI blog 1(8), 9 (2019)","journal-title":"OpenAI blog"},{"key":"797_CR246","doi-asserted-by":"crossref","unstructured":"Raji, I\u00a0D., Xu, P., Honigsberg, C., Ho, D.: Outsider oversight: Designing a third party audit ecosystem for ai governance. In Proceedings of the 2022 AAAI\/ACM Conference on AI, Ethics, and Society, pages 557\u2013571, (2022)","DOI":"10.1145\/3514094.3534181"},{"key":"797_CR247","doi-asserted-by":"crossref","unstructured":"Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: SQuAD: 100,000+ questions for machine comprehension of text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2383\u20132392, Austin, Texas, (November 2016) Association for Computational Linguistics","DOI":"10.18653\/v1\/D16-1264"},{"key":"797_CR248","doi-asserted-by":"crossref","first-page":"1316","DOI":"10.1162\/tacl_a_00605","volume":"11","author":"O Ram","year":"2023","unstructured":"Ram, O., Levine, Y., Dalmedigos, I., Muhlgay, D., Shashua, A., Leyton-Brown, K., Shoham, Y.: In-context retrieval-augmented language models. Trans. Assoc. Comput. Linguistics 11, 1316\u20131331 (2023)","journal-title":"Trans. Assoc. Comput. Linguistics"},{"key":"797_CR249","doi-asserted-by":"crossref","unstructured":"Ramesh, K., Chavan, A., Pandit, S., Sitaram, S.: A comparative study on the impact of model compression techniques on fairness in language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15762\u201315782, (2023)","DOI":"10.18653\/v1\/2023.acl-long.878"},{"key":"797_CR250","unstructured":"Rashid, M\u00a0R\u00a0U., Dasu, V\u00a0A., Gu, K., Sultana, N., Mehnaz, S.: Fltrojan: Privacy leakage attacks against federated language models through selective weight tampering. arXiv preprint arXiv:2310.16152, (2023)"},{"key":"797_CR251","doi-asserted-by":"crossref","unstructured":"Ravfogel, S., Elazar, Y., Gonen, H., Twiton, M., Goldberg, Y.: Null it out: Guarding protected attributes by iterative nullspace projection. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7237\u20137256, Online, (July 2020) Association for Computational Linguistics","DOI":"10.18653\/v1\/2020.acl-main.647"},{"key":"797_CR252","unstructured":"General Data\u00a0Protection Regulation. General data protection regulation (gdpr). Intersoft Consulting, Accessed in October, 24(1), (2018)"},{"key":"797_CR253","doi-asserted-by":"crossref","unstructured":"Rekabsaz, N., Kopeinik, S., Schedl, M.: Societal biases in retrieved contents: Measurement framework and adversarial mitigation of bert rankers. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 306\u2013316, (2021)","DOI":"10.1145\/3404835.3462949"},{"key":"797_CR254","doi-asserted-by":"crossref","unstructured":"Rekabsaz, N., Schedl, M.: Do neural ranking models intensify gender bias? In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR \u201920, page 2065\u20132068, New York, NY, USA, (2020). Association for Computing Machinery","DOI":"10.1145\/3397271.3401280"},{"key":"797_CR255","unstructured":"Ren, A\u00a0Z., Dixit, A., Bodrova, A., Singh, S., Tu, S., Brown, N., Xu, P., Takayama, L., Xia, F., Varley, J., et\u00a0al.: Robots that ask for help: Uncertainty alignment for large language model planners. arXiv preprint arXiv:2307.01928, (2023)"},{"key":"797_CR256","unstructured":"Ren, J., Xu, H., Liu, Y., Cui, Y., Wang, S., Yin, D., Tang, J.: A robust semantics-based watermark for large language model against paraphrasing. ArXiv, abs\/2311.08721, (2023)"},{"key":"797_CR257","doi-asserted-by":"crossref","first-page":"59","DOI":"10.1007\/s00146-020-00992-2","volume":"36","author":"H Roberts","year":"2021","unstructured":"Roberts, H., Cowls, J., Morley, J., Taddeo, M., Wang, V., Floridi, L.: The chinese approach to artificial intelligence: an analysis of policy, ethics, and regulation. AI Soc. 36, 59\u201377 (2021)","journal-title":"AI Soc."},{"key":"797_CR258","volume-title":"Breaking down the visual barrier: Designing data interactions for the visually impaired in informal learning settings","author":"J Roberts","year":"2023","unstructured":"Roberts, J., Lowy, R., Li, H., Bellona, J., Smith, L., Bower, A.: Breaking down the visual barrier: Designing data interactions for the visually impaired in informal learning settings. International Society of the Learning Sciences, In CSCL (2023)"},{"key":"797_CR259","doi-asserted-by":"crossref","unstructured":"Ruan, K., He, X., Wang, J., Zhou, X., Feng, H., Kebarighotbi, A.: S2e: Towards an end-to-end entity resolution solution from acoustic signal. In ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 10441\u201310445. IEEE, (2024)","DOI":"10.1109\/ICASSP48485.2024.10446126"},{"key":"797_CR260","unstructured":"Sah, C\u00a0K., Xiaoli, L., Islam, M\u00a0M.: Unveiling bias in fairness evaluations of large language models: A critical literature review of music and movie recommendation systems. arXiv preprint arXiv:2401.04057, (2024)"},{"key":"797_CR261","doi-asserted-by":"crossref","unstructured":"Saunders, D., Sallis, R., Byrne, B.: First the worst: Finding better gender translations during beam search. In Findings of the Association for Computational Linguistics: ACL 2022, pages 3814\u20133823, Dublin, Ireland, Association for Computational Linguistics (May 2022)","DOI":"10.18653\/v1\/2022.findings-acl.301"},{"key":"797_CR262","unstructured":"Shi, W., Cui, A., Li, E., Jia, R., Yu, Z.: Selective differential privacy for language modeling. arXiv preprint arXiv:2108.12944, (2021)"},{"key":"797_CR263","doi-asserted-by":"crossref","unstructured":"Shoaib, M\u00a0R., Wang, Z., Ahvanooey, M\u00a0T., Zhao, J.: Deepfakes, misinformation, and disinformation in the era of frontier ai, generative ai, and large ai models. In ICCA, pages 1\u20137. IEEE, (2023)","DOI":"10.1109\/ICCA59364.2023.10401723"},{"key":"797_CR264","doi-asserted-by":"crossref","unstructured":"Shokri, R., Stronati, M., Song, C., Shmatikov, V.: Membership inference attacks against machine learning models. In 2017 IEEE symposium on security and privacy (SP), pages 3\u201318. IEEE, (2017)","DOI":"10.1109\/SP.2017.41"},{"key":"797_CR265","volume-title":"Chadha","author":"P Singh","year":"2013","unstructured":"Singh, P.: Chadha. A survey of digital watermarking techniques, applications and attacks, Ramneet Singh (2013)"},{"key":"797_CR266","doi-asserted-by":"crossref","unstructured":"Smith, E\u00a0M., Hall, M., Kambadur, M., Presani, E., Williams, A.: \u201cI\u2019m sorry to hear that\u201d:Finding new biases in language models with a holistic descriptor dataset. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9180\u20139211, Abu Dhabi, United Arab Emirates, Association for Computational Linguistics (December 2022)","DOI":"10.18653\/v1\/2022.emnlp-main.625"},{"key":"797_CR267","doi-asserted-by":"crossref","unstructured":"Smith, E.\u00a0M., Hall, M., Kambadur, M., Presani, E., Williams, A.:\u201ci\u2019m sorry to hear that\u201d:Finding new biases in language models with a holistic descriptor dataset. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 9180\u20139211, (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.625"},{"key":"797_CR268","doi-asserted-by":"crossref","unstructured":"Song, C., Raghunathan, A.: Information leakage in embedding models. In Proceedings of the 2020 ACM SIGSAC conference on computer and communications security, pages 377\u2013390, (2020)","DOI":"10.1145\/3372297.3417270"},{"key":"797_CR269","unstructured":"Song, F., Yu, B., Li, M., Yu, H., Huang, F., Li, Y., Wang, H.: Preference ranking optimization for human alignment. arXiv preprint arXiv:2306.17492, (2023)"},{"key":"797_CR270","unstructured":"Staab, R., Vero, M., Balunovi\u0107, M., Vechev, M.: Beyond memorization: Violating privacy via inference with large language models. arXiv preprint arXiv:2310.07298, (2023)"},{"key":"797_CR271","unstructured":"Su, J., Jiang, C., Jin, X., Qiao, Y., Xiao, T., Ma, Hongda, W., Rong, J., Zhi, X., Jiajun, L.: Large language models for forecasting and anomaly detection: A systematic literature review. arXiv preprint arXiv:2402.10350, (2024)"},{"key":"797_CR272","doi-asserted-by":"crossref","unstructured":"Sun, H., Zhang, Z., Mi, F., Wang, Y., Liu, W., Cui, J., Wang, B., Liu, Q., Huang, M.: MoralDial: A framework to train and evaluate moral dialogue systems via moral discussions. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2213\u20132230, Toronto, Canada, July 2023. Association for Computational Linguistics","DOI":"10.18653\/v1\/2023.acl-long.123"},{"key":"797_CR273","unstructured":"Sun, L., Huang, Y., Wang, H., Wu, S., Zhang, Q., Gao, C., Huang, Y., Lyu, W., Zhang, Y., Li, X., et\u00a0al.: Trustllm: Trustworthiness in large language models. arXiv preprint arXiv:2401.05561, (2024)"},{"key":"797_CR274","unstructured":"Sun, T., Webster, K., Shah, A., Wang, W\u00a0Y., Johnson, M.: They, them, theirs: Rewriting with gender-neutral english. arXiv preprint arXiv:2102.06788, (2021)"},{"key":"797_CR275","unstructured":"Sun, Z., Shen, Y., Zhou, Q., Zhang, H., Chen, Z., Cox, D., Yang, Y., Gan, C.: Principle-driven self-alignment of language models from scratch with minimal human supervision. arXiv preprint arXiv:2305.03047, (2023)"},{"key":"797_CR276","doi-asserted-by":"crossref","unstructured":"Suresh, H., Guttag, J.: A framework for understanding sources of harm throughout the machine learning life cycle. Equity and access in algorithms, mechanisms, and optimization, pages 1\u20139, (2021)","DOI":"10.1145\/3465416.3483305"},{"key":"797_CR277","unstructured":"Tang, X., Shin, R., Inan, H\u00a0A., Manoel, A., Mireshghallah, F., Lin, Z., Gopi, S., Kulkarni, J., Sim, R.: Privacy-preserving in-context learning with differentially private few-shot generation. arXiv preprint arXiv:2309.11765, (2023)"},{"key":"797_CR278","unstructured":"Taori, R., Gulrajani, I., Zhang, T., Dubois, Y., Li, X., Guestrin, C., Liang, P.: and Tatsunori B Hashimoto. An instruction-following llama model, Stanford alpaca (2023)"},{"key":"797_CR279","volume-title":"Nitesh V Chawla, and Panpan Xu","author":"Y Tian","year":"2024","unstructured":"Tian, Y., Song, H., Wang, Z., Wang, H., Ziqing, H., Wang, F.: Nitesh V Chawla, and Panpan Xu. Graph neural prompting with large language models, In AAAI (2024)"},{"key":"797_CR280","first-page":"38274","volume":"35","author":"K Tirumala","year":"2022","unstructured":"Tirumala, K., Markosyan, A., Zettlemoyer, L., Aghajanyan, A.: Memorization without overfitting: Analyzing the training dynamics of large language models. Adv. Neural. Inf. Process. Syst. 35, 38274\u201338290 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"797_CR281","doi-asserted-by":"crossref","unstructured":"Tokpo, E\u00a0K., Calders, T.: Text style transfer for bias mitigation using masked language modeling. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Student Research Workshop, pages 163\u2013171, Hybrid: Seattle, Washington + Online, July 2022. Association for Computational Linguistics","DOI":"10.18653\/v1\/2022.naacl-srw.21"},{"key":"797_CR282","unstructured":"Tong, M., Chen, K., Qi, Y., Zhang, J., Zhang, W., Yu, N.: Privinfer: Privacy-preserving inference for black-box large language model. arXiv preprint arXiv:2310.12214, (2023)"},{"key":"797_CR283","doi-asserted-by":"crossref","unstructured":"Topkara, M., Taskiran, C\u00a0M., Delp, E\u00a0J.: Natural language watermarking. In IS &T\/SPIE Electronic Imaging, (2005)","DOI":"10.1117\/12.593790"},{"key":"797_CR284","unstructured":"Tople, S.: Brockschmidt, Marc, K\u00f6pf, Boris, Ohrimenko, Olga, Zanella-B\u00e9guelin. Analyzing privacy loss in updates of natural language models, Santiago (2019)"},{"key":"797_CR285","unstructured":"Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M-A., Lacroix, T., Rozi\u00e8re, B., Goyal, N., Hambro, E., Azhar, F., et\u00a0al.: Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, (2023)"},{"key":"797_CR286","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et\u00a0al.: Llama 2: Open foundation and fine-tuned chat models, 2023. arxiv:2307.09288, (2023)"},{"key":"797_CR287","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.102135","volume":"103","author":"I Triguero","year":"2024","unstructured":"Triguero, I., Molina, D., Poyatos, J., Del Ser, J., Herrera, F.: General purpose artificial intelligence systems (gpais): Properties, definition, taxonomy, societal implications and responsible governance. Inf. Fusion 103, 102135 (2024)","journal-title":"Inf. Fusion"},{"key":"797_CR288","doi-asserted-by":"crossref","unstructured":"Tripto, N\u00a0I., Venkatraman, S., Macko, D., Moro, R., Srba, I., Uchendu, A., Le, T., Lee, D.: A ship of theseus: Curious cases of paraphrasing in llm-generated texts. arXiv preprint arXiv:2311.08374, (2023)","DOI":"10.18653\/v1\/2024.acl-long.357"},{"key":"797_CR289","doi-asserted-by":"crossref","unstructured":"Truex, S., Baracaldo, N., Anwar, A., Steinke, T., Ludwig, H., Zhang, R., Zhou, Y.: A hybrid approach to privacy-preserving federated learning. In Proceedings of the 12th ACM workshop on artificial intelligence and security, pages 1\u201311, (2019)","DOI":"10.1145\/3338501.3357370"},{"key":"797_CR290","doi-asserted-by":"crossref","unstructured":"Vanmassenhove, E., Emmery, C., Shterionov, D.: NeuTral Rewriter: A rule-based and neural approach to automatic rewriting into gender neutral alternatives. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 8940\u20138948, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics","DOI":"10.18653\/v1\/2021.emnlp-main.704"},{"key":"797_CR291","unstructured":"Verma, G., Choi, M., Sharma, K., Watson-Daniels, J., Oh, S., Kumar, S.: Mysterious projections: Multimodal llms gain domain-specific visual capabilities without richer cross-modal projections. arXiv:2402.16832, (2024)"},{"key":"797_CR292","doi-asserted-by":"crossref","unstructured":"Voigt, P., Bussche, AV\u00a0d.: The eu general data protection regulation (gdpr). A Practical Guide, 1st Ed., Cham: Springer International Publishing, 10(3152676):10\u20135555, (2017)","DOI":"10.1007\/978-3-319-57959-7"},{"key":"797_CR293","doi-asserted-by":"crossref","unstructured":"Wan, D., Liu, M., McKeown, K., Dreyer, M., Bansal, M.: Faithfulness-aware decoding strategies for abstractive summarization. arXiv preprint arXiv:2303.03278, (2023)","DOI":"10.18653\/v1\/2023.eacl-main.210"},{"key":"797_CR294","doi-asserted-by":"crossref","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S\u00a0R.: Glue: A multi-task benchmark and analysis platform for natural language understanding. arXiv preprint arXiv:1804.07461, (2018)","DOI":"10.18653\/v1\/W18-5446"},{"key":"797_CR295","unstructured":"Wang, B., Chen, W., Pei, H., Xie, C., Kang, M., Zhang, C., Xu, C., Xiong, Z., Dutta, R., Schaeffer, R., et\u00a0al.: Decodingtrust: A comprehensive assessment of trustworthiness in gpt models. arXiv preprint arXiv:2306.11698, (2023)"},{"key":"797_CR296","unstructured":"Wang, B., Xu, C., Wang, S., Gan, Z., Cheng, Y., Gao, J., Awadallah, A\u00a0H., Li, B.: Adversarial glue: A multi-task benchmark for robustness evaluation of language models. arXiv preprint arXiv:2111.02840, (2021)"},{"key":"797_CR297","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Y., Sun, R., Li, B.: Global convergence of maml and theory-inspired neural architecture search for few-shot learning. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pages 9797\u20139808, (2022)","DOI":"10.1109\/CVPR52688.2022.00957"},{"key":"797_CR298","unstructured":"Wang, J., Hu, X., Hou, W., Chen, H., Zheng, R., Wang, Y., Yang, L., Huang, H., Ye, W., Geng, X., et\u00a0al.: On the robustness of chatgpt: An adversarial and out-of-distribution perspective. arXiv preprint arXiv:2302.12095, (2023)"},{"key":"797_CR299","unstructured":"Wang, J., Zhou, Y., Xu, G., Shi, P., Zhao, C., Xu, H., Ye, Q., Yan, M., Zhang, J., Zhu, J., et\u00a0al.: Evaluation and analysis of hallucination in large vision-language models. arXiv:2308.15126, (2023)"},{"key":"797_CR300","unstructured":"Wang, L., Yang, W., Chen, D., Zhou, H., Lin, Y., Meng, F., Zhou, J., Sun, X.: Towards codable text watermarking for large language models. ArXiv, abs\/2307.15992, (2023)"},{"key":"797_CR301","doi-asserted-by":"crossref","unstructured":"Wang, X., Zhou, Y., Liu, X., Lu, H., Xu, Y., He, F., Yoon, J., Lu, T., Bertasius, G., Bansal, M., et\u00a0al.: Mementos: A comprehensive benchmark for multimodal large language model reasoning over image sequences. arXiv preprint arXiv:2401.10529, (2024)","DOI":"10.18653\/v1\/2024.acl-long.25"},{"key":"797_CR302","doi-asserted-by":"crossref","unstructured":"Wang, Y., Feng, S., Hou, A\u00a0B., Pu, X., Shen, C., Liu, X., Tsvetkov, Y., He, T.: Stumbling blocks: Stress testing the robustness of machine-generated text detectors under attacks. arXiv preprint arXiv:2402.11638, (2024)","DOI":"10.18653\/v1\/2024.acl-long.160"},{"key":"797_CR303","unstructured":"Wang, Y., Li, D., Sun, R.: Ntk-sap: Improving neural network pruning by aligning training dynamics. In The Eleventh International Conference on Learning Representations, (2022)"},{"key":"797_CR304","unstructured":"Wang, Y., Su, J., Lu, H., Xie, C., Liu, Tianyi, Y., Jianbo, L., Haibin, S., Ruoyu, Y., Hongxi: Lemon: Lossless model expansion. In The Twelfth International Conference on Learning Representations, (2023)"},{"key":"797_CR305","unstructured":"Wang, Y., Wu, J., Hovakimyan, N., Sun, R.: Balanced training for sparse gans. Advances in Neural Information Processing Systems, 36, (2024)"},{"key":"797_CR306","doi-asserted-by":"crossref","unstructured":"Weaver, J\u00a0F.: Regulation of artificial intelligence in the united states. In Research Handbook on the Law of Artificial Intelligence, pages 155\u2013212. Edward Elgar Publishing, (2018)","DOI":"10.4337\/9781786439055.00018"},{"key":"797_CR307","unstructured":"Webster, K., Wang, X., Tenney, I., Beutel, A., Pitler, E., Pavlick, E., Chen, J., Chi, E., Petrov, S.: Measuring and reducing gendered correlations in pre-trained models. arXiv preprint arXiv:2010.06032, (2020)"},{"key":"797_CR308","unstructured":"Weidinger, L., Mellor, J., Rauh, M., Griffin, C., Uesato, J., Huang, P-S., Cheng, M., Glaese, M., Balle, B., Kasirzadeh, A., et\u00a0al.: Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359, 2021"},{"key":"797_CR309","doi-asserted-by":"crossref","unstructured":"Weinberger, K\u00a0Q., Sha, F., Zhu, Q., Saul, L.: Graph laplacian regularization for large-scale semidefinite programming. Advances in neural information processing systems, 19, (2006)","DOI":"10.7551\/mitpress\/7503.003.0191"},{"key":"797_CR310","doi-asserted-by":"crossref","unstructured":"Welbl, J., Glaese, A., Uesato, J., Dathathri, S., Mellor, J., Hendricks, L\u00a0A., Anderson, K., Kohli, P., Coppin, B., Huang, P-S.: Challenges in detoxifying language models. arXiv preprint arXiv:2109.07445, (2021)","DOI":"10.18653\/v1\/2021.findings-emnlp.210"},{"key":"797_CR311","doi-asserted-by":"crossref","unstructured":"Wen, J., Ke, P., Sun, H., Zhang, Z., Li, C., Bai, J., Huang, M.: Unveiling the implicit toxicity in large language models. arXiv preprint arXiv:2311.17391, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.84"},{"key":"797_CR312","unstructured":"Winograd, A.: Loose-lipped large language models spill your secrets: The privacy implications of large language models. Harvard Journal of Law & Technology, 36(2), (2023)"},{"key":"797_CR313","unstructured":"Wolf, Y., Wies, N., Levine, Y., Shashua, A.: Fundamental limitations of alignment in large language models. arXiv preprint arXiv:2304.11082, (2023)"},{"key":"797_CR314","unstructured":"Wu, T., Zhu, B., Zhang, R., Wen, Z., Ramchandran, K., Jiao, J.; Pairwise proximal policy optimization: Harnessing relative feedback for llm alignment. arXiv preprint arXiv:2310.00212, (2023)"},{"issue":"5","key":"797_CR315","doi-asserted-by":"crossref","first-page":"1122","DOI":"10.1109\/JAS.2023.123618","volume":"10","author":"W Tianyu","year":"2023","unstructured":"Tianyu, W., He, S., Liu, J., Sun, S., Liu, K., Han, Q.-L., Tang, Y.: A brief overview of chatgpt: The history, status quo and potential future development. IEEE\/CAA J. Automatica Sinica 10(5), 1122\u20131136 (2023)","journal-title":"IEEE\/CAA J. Automatica Sinica"},{"key":"797_CR316","unstructured":"Wu, T., Panda, A., Wang, J\u00a0T., Mittal, P.: Privacy-preserving in-context learning for large language models. arXiv e-prints, pages arXiv\u20132305, (2023)"},{"key":"797_CR317","unstructured":"Xiang, T., Xie, C., Guo, S., Li, Jiwei, Z., Tianwei: Protecting your nlg models with semantic and robust watermarks. arXiv preprint arXiv:2112.05428, (2021)"},{"key":"797_CR318","unstructured":"Xiang, Z., Jiang, F., Xiong, Z., Ramasubramanian, B., Radha P., and Bo L,: Backdoor chain-of-thought prompting for large language models. In ICLR, Badchain (2024)"},{"key":"797_CR319","doi-asserted-by":"crossref","unstructured":"Xiao, Y., Jin, Y., Bai, Y., Wu, Y., Yang, X., Luo, X., Yu, W., Zhao, X., Liu, Y., Chen, H., et\u00a0al.: Large language models can be good privacy protection learners. arXiv preprint arXiv:2310.02469, (2023)","DOI":"10.18653\/v1\/2024.emnlp-main.785"},{"key":"797_CR320","unstructured":"Xu, H., Liu, X., Li, Y., Jain, A., Tang J.: To be robust or to be fair: Towards fairness in adversarial training. In International conference on machine learning, pages 11492\u201311501. PMLR, (2021)"},{"key":"797_CR321","doi-asserted-by":"crossref","unstructured":"Xu, M., Song, C., Tian, Y., Agrawal, N., Granqvist, F., van D., Rogier, Z., Xiao, A., Arturo, H., Shiyi, D., Yaqiao, et\u00a0al.: Training large-vocabulary neural language models by private federated learning for resource-constrained devices. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1\u20135. IEEE, (2023)","DOI":"10.1109\/ICASSP49357.2023.10096570"},{"key":"797_CR322","first-page":"17238","volume":"36","author":"H Yan","year":"2023","unstructured":"Yan, H., Li, C., Long, R., Yan, C., Zhao, J., Zhuang, W., Yin, J., Zhang, P., Han, W., Sun, H., et al.: A comprehensive study on text-attributed graphs: Benchmarking and rethinking. NeurIPS 36, 17238\u201317264 (2023)","journal-title":"NeurIPS"},{"issue":"5","key":"797_CR323","doi-asserted-by":"crossref","first-page":"283","DOI":"10.1038\/s42256-020-0180-7","volume":"2","author":"L Yan","year":"2020","unstructured":"Yan, L., Zhang, H.-T., Goncalves, J., Xiao, Y., Wang, M., Guo, Y., Sun, C., Tang, X., Jing, L., Zhang, M., et al.: An interpretable mortality prediction model for covid-19 patients. Nature Mach. Intell. 2(5), 283\u2013288 (2020)","journal-title":"Nature Mach. Intell."},{"key":"797_CR324","unstructured":"Yang, Y., Chern, E., Qiu, X., Neubig, G., Liu, P.: Alignment for honesty. arXiv preprint arXiv:2312.07000, (2023)"},{"key":"797_CR325","unstructured":"Yang, Z., Yi, X., Li, P., Liu, Y., Xie, X: Unified detoxifying and debiasing in language generation via inference-time adaptive optimization. arXiv preprint arXiv:2210.04492, (2022)"},{"key":"797_CR326","doi-asserted-by":"crossref","unstructured":"Yao, H., Lou, J., Ren, K., Qin, Z.: Promptcare: Prompt copyright protection by watermark injection and verification. arXiv preprint arXiv:2308.02816, (2023)","DOI":"10.1109\/SP54263.2024.00209"},{"key":"797_CR327","unstructured":"Yao, Y., Xu, X., Liu, Y.: Large language model unlearning. arXiv preprint arXiv:2310.10683, (2023)"},{"key":"797_CR328","doi-asserted-by":"crossref","unstructured":"Yao, Y., Wang, P., Tian, B., Cheng, S., Li, Z., Deng, S., Chen, H., Zhang, N.: Editing large language models: Problems, methods, and opportunities. arXiv preprintarXiv:2305.13172, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.632"},{"key":"797_CR329","doi-asserted-by":"crossref","unstructured":"Ye, J., Kang, H., Wang, H., Altaleb, S., Heidari, E., Asadizanjani, N., S., Volker\u00a0J., Dalir, H.: Multiplexed oam beams classification via fourier optical convolutional neural network. In 2023 IEEE Photonics Conference (IPC), pages 1\u20132. IEEE, (2023)","DOI":"10.1109\/IPC57732.2023.10360629"},{"key":"797_CR330","doi-asserted-by":"crossref","unstructured":"Ye, J., Kang, H., Wang, H., Altaleb, S., Heidari, E., Asadizanjani, N., Sorger, V.\u00a0J., Dalir, H.: Oam beams multiplexing and classification under atmospheric turbulence via fourier convolutional neural network. In Frontiers in Optics, pages JTu4A\u201373. Optica Publishing Group, (2023)","DOI":"10.1364\/FIO.2023.JTu4A.73"},{"key":"797_CR331","doi-asserted-by":"crossref","unstructured":"Ye, J., Kang, H., Wang, H., Shen, C., Jahannia, B., Heidari, E., Asadizanjani, N., Miri, M-A., Sorger, V\u00a0J., Dalir, H.: Demultiplexing oam beams via fourier optical convolutional neural network. In Laser Beam Shaping XXIII, volume 12667, pages 16\u201333. SPIE, (2023)","DOI":"10.1117\/12.2682108"},{"key":"797_CR332","doi-asserted-by":"crossref","unstructured":"Ye, J., Solyanik, M., Hu, Z., Dalir, H., Nouri, B\u00a0M., Sorger, V\u00a0J.: Free-space optical multiplexed orbital angular momentum beam identification system using fourier optical convolutional layer based on 4f system. In Complex Light and Optical Forces XVII, volume 12436, pages 70\u201380. SPIE, (2023)","DOI":"10.1117\/12.2649451"},{"key":"797_CR333","unstructured":"Yin, Z., Wang, Z., Xu, W., Zhuang, J., Mozumder, P., Smith, A., Zhang, W.: Digital forensics in the age of large language models. arXiv preprint arXiv:2504.02963, (2025)"},{"key":"797_CR334","doi-asserted-by":"crossref","unstructured":"Yoo, K., Ahn, W., Jang, J., Kwak, N\u00a0J.: Robust multi-bit natural language watermarking through invariant features. In Annual Meeting of the Association for Computational Linguistics, (2023)","DOI":"10.18653\/v1\/2023.acl-long.117"},{"key":"797_CR335","unstructured":"Charles, Y., Jeoung, S., Kasi, A., Pengfei, Y., Ji, H.: Unlearning bias in language models by partitioning gradients. In Findings of the Association for Computational Linguistics: ACL 2023, 6032\u20136048 (2023)"},{"key":"797_CR336","unstructured":"Yu, D., Naik, S., Backurs, A., Gopi, S., Inan, H\u00a0A., Kamath, G., Kulkarni, J., Lee, Y\u00a0T, Manoel, A., Wutschitz, L., et\u00a0al.: Differentially private fine-tuning of language models. arXiv preprint arXiv:2110.06500, (2021)"},{"key":"797_CR337","doi-asserted-by":"crossref","unstructured":"Yu, L., Mao, Y., Wu, J., Zhou, F.: Mixup-based unified framework to overcome gender bias resurgence. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR \u201923, page 1755\u20131759, New York, NY, USA, (2023). Association for Computing Machinery","DOI":"10.1145\/3539618.3591938"},{"key":"797_CR338","unstructured":"Yu, W., Pang, T., Liu, Q., Du, C., Kang, B., Huang, Y., Lin, M., Yan, S.: Bag of tricks for training data extraction from language models. arXiv preprint arXiv:2302.04460, (2023)"},{"key":"797_CR339","unstructured":"Zayed, A., Mordido, G., Shabanian, S., Chandar, S.: Should we attend more or less? modulating attention for fairness. arXiv preprint arXiv:2305.13088, (2023)"},{"key":"797_CR340","doi-asserted-by":"crossref","unstructured":"Zayed, A., Parthasarathi, P., Mordido, G., Palangi, H., Shabanian, S., Chandar, S.: Deep learning on a healthy data diet: Finding important examples for fairness. In Proceedings of the AAAI Conference on Artificial Intelligence 37, 14593\u201314601 (2023)","DOI":"10.1609\/aaai.v37i12.26706"},{"key":"797_CR341","unstructured":"Zeng, A., Liu, X., Du, Z., Wang, Z., Lai, H., Ding, M., Yang, Z., Xu, Y., Zheng, W., Xia, X., et\u00a0al.: Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414, (2022)"},{"key":"797_CR342","doi-asserted-by":"crossref","unstructured":"Zeng, W., Li, M., Xiong, W., Lu, W., Tan, J., Wang, R., Huang, R.: Mpcvit: Searching for mpc-friendly vision transformer with heterogeneous attention. arXiv preprint arXiv:2211.13955, (2022)","DOI":"10.1109\/ICCV51070.2023.00466"},{"key":"797_CR343","doi-asserted-by":"crossref","unstructured":"Chen Z., Yu., XHB., Bin, Y., Li, W., Gao, Y.: A survey on federated learning. Knowl.-Based Syst. 216, 106775 (2021)","DOI":"10.1016\/j.knosys.2021.106775"},{"key":"797_CR344","unstructured":"Zhang, C., Ekanut, S., Zhen, L., Li, Z.: Augmented multi-party computation against gradient leakage in federated learning. IEEE Transactions on Big Data, (2022)"},{"key":"797_CR345","unstructured":"Zhang, M., Zhao, H., Ebron, S., Xie, R., Yang, K.: Multi-criteria client selection and scheduling with fairness guarantee for federated learning service. arXiv preprint arXiv:2312.14941, (2023)"},{"key":"797_CR346","doi-asserted-by":"crossref","unstructured":"Zhang, P., Li, C., Kang, L., Huang, F., Wang, S., Xie, X., Kim, S.: High-frequency-aware hierarchical contrastive selective coding for representation learning on text-attributed graphs. arXiv:2402.16240, (2024)","DOI":"10.1145\/3589334.3645614"},{"key":"797_CR347","unstructured":"Zhang, P., Liu, H., Li, C., Xie, X., Kim, S.: and Haohan Wang. Robust image model evaluation with pretrained models. In ICLR, Foundation model-oriented robustness (2023)"},{"key":"797_CR348","unstructured":"Zhang, S., Roller, S., Goyal, N., Artetxe, M., Chen, M., Chen, S., Dewan, C., Diab, M., Li, X., Lin, X\u00a0V., et\u00a0al.: Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068, (2022)"},{"issue":"2","key":"797_CR349","first-page":"46","volume":"2","author":"Y Zhang","year":"2024","unstructured":"Zhang, Y., Gui, K., Zhu, M., Hao, Y., Sun, H.: Unlocking personalized anime recommendations: Langchain and llm at the forefront. J. Indus. Eng. Appl. Sci. 2(2), 46\u201353 (2024)","journal-title":"J. Indus. Eng. Appl. Sci."},{"key":"797_CR350","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Sun, S., Galley, M., Chen, Y-C., Brockett, C., Gao, X., Gao, J., Liu, J., Dolan, B.: Dialogpt: Large-scale generative pre-training for conversational response generation. arXiv preprint arXiv:1911.00536, (2019)","DOI":"10.18653\/v1\/2020.acl-demos.30"},{"key":"797_CR351","unstructured":"Zhang, Y., Li, Y., Cui, L., Cai, D., Liu, L., Fu, T., Huang, X., Zhao, E., Zhang, Y., Chen, Y., et\u00a0al.: Siren\u2019s song in the ai ocean: A survey on hallucination in large language models. arXiv:2309.01219, (2023)"},{"key":"797_CR352","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Chen, D., Kundu, S., Liu, H., Peng, R., Beerel, P\u00a0A.: C2pi: An efficient crypto-clear two-party neural network private inference. arXiv preprint arXiv:2304.13266, (2023)","DOI":"10.1109\/DAC56929.2023.10247682"},{"key":"797_CR353","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Wen, J., Huang, M.: Ethicist: Targeted training data extraction through loss smoothed soft prompting and calibrated confidence estimation. arXiv preprint arXiv:2307.04401, (2023)","DOI":"10.18653\/v1\/2023.acl-long.709"},{"key":"797_CR354","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Yang, Y., Dai, Yong, W., Qifan, Y., Yue, Q., Lizhen, X., Zenglin: Fedpetuning: When federated learning meets the parameter-efficient tuning methods of pre-trained language models. In Annual Meeting of the Association of Computational Linguistics 2023, pages 9963\u20139977. Association for Computational Linguistics (ACL), (2023)","DOI":"10.18653\/v1\/2023.findings-acl.632"},{"key":"797_CR355","doi-asserted-by":"crossref","unstructured":"Zhao, H., Chen, H., Yang, F., Liu, N., Deng, H., Cai, H., Wang, S., Yin, D., Mengnan D. A survey. ACM Transactions on Intelligent Systems and Technology, Explainability for large language models (2023)","DOI":"10.1145\/3639372"},{"key":"797_CR356","unstructured":"Zhao, Q., Wang, J., Zhang, Y., Jin, Y., Zhu, K., Chen, H., Xie, X.: Competeai: Understanding the competition behaviors in large language model-based agents. arXiv preprint arXiv:2310.17512, (2023)"},{"key":"797_CR357","unstructured":"Zheng, L., Chiang, W-L., Sheng, Y., Zhuang, S., Wu, Z., Zhuang, Z\u00a0L., Yonghao, L., Zhuohan, L., Dacheng, X., Eric, et\u00a0al.: Judging llm-as-a-judge with mt-bench and chatbot arena. arXiv preprint arXiv:2306.05685, (2023)"},{"key":"797_CR358","doi-asserted-by":"crossref","unstructured":"Zheng, M., Lou, Q., Jiang, L.: Primer: Fast private transformer inference on encrypted data. arXiv preprint arXiv:2303.13679, (2023)","DOI":"10.1109\/DAC56929.2023.10247719"},{"key":"797_CR359","unstructured":"Zheng, S., Huang, J., Chang, KC-C.: Why does chatgpt fall short in answering questions faithfully? arXiv preprint arXiv:2304.10513, (2023)"},{"key":"797_CR360","unstructured":"Zhou, C., Liu, P., Xu, P., Iyer, S., Sun, J., Mao, Y., Ma, X., Efrat, A., Yu, P., Yu, L., et\u00a0al.: Lima: Less is more for alignment. arXiv preprint arXiv:2305.11206, (2023)"},{"key":"797_CR361","doi-asserted-by":"crossref","unstructured":"Zhou, K., Jurafsky, D., Hashimoto, T.: Navigating the grey area: Expressions of overconfidence and uncertainty in language models. arXiv preprint arXiv:2302.13439, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.335"},{"key":"797_CR362","doi-asserted-by":"crossref","unstructured":"Zhu, K., Wang, J., Zhou, J., Wang, Z., Chen, H., Wang, Y., Yang, L., Ye, W., Gong, N\u00a0Z., Zhang, Y., et\u00a0al.: Promptbench: Towards evaluating the robustness of large language models on adversarial prompts. arXiv preprint arXiv:2306.04528, (2023)","DOI":"10.1145\/3689217.3690621"},{"key":"797_CR363","volume-title":"International Conference on Cloud Computing, Internet of Things, and Computer Applications (CICA 2022), volume 12303, page 1230306","author":"W Zhu","year":"2022","unstructured":"Zhu, W.: Optimizing distributed networking with big data scheduling and cloud computing. In: Powell, W., Tolba, A. (eds.) International Conference on Cloud Computing, Internet of Things, and Computer Applications (CICA 2022), volume 12303, page 1230306. International Society for Optics and Photonics, SPIE (2022)"},{"key":"797_CR364","doi-asserted-by":"crossref","unstructured":"Zhu, W., Hu, T.: Twitter sentiment analysis of covid vaccines. In 2021 5th International Conference on Artificial Intelligence and Virtual Reality (AIVR), AIVR 2021, page 118\u2013122, New York, NY, USA, (2021). Association for Computing Machinery","DOI":"10.1145\/3480433.3480442"},{"key":"797_CR365","first-page":"1486","volume":"2024","author":"J Zhuang","year":"2024","unstructured":"Zhuang, J.: Robust data-centric graph structure learning for text classification. Compan. Proc. ACM Web Conf. 2024, 1486\u20131495 (2024)","journal-title":"Compan. Proc. ACM Web Conf."},{"key":"797_CR366","unstructured":"Zhuang, J., Guan, C.: Large language models can help mitigate barren plateaus. arXiv preprint arXiv:2502.13166, (2025)"},{"key":"797_CR367","doi-asserted-by":"crossref","unstructured":"Zhuang, J., Kennington, C.: Understanding survey paper taxonomy about large language models via graph representation learning. arXiv preprint arXiv:2402.10409, (2024)","DOI":"10.18653\/v1\/2024.sdp-1.6"},{"key":"797_CR368","unstructured":"Zou, A., Wang, Z., Kolter, J\u00a0Z., Fredrikson, M: Universal and transferable adversarial attacks on aligned language models. arXiv preprint arXiv:2307.15043, (2023)"},{"key":"797_CR369","doi-asserted-by":"crossref","unstructured":"Zou, H., Caragea, C.: Jointmatch: A unified approach for diverse and collaborative pseudo-labeling to semi-supervised text classification. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 7290\u20137301, (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.451"},{"key":"797_CR370","doi-asserted-by":"crossref","unstructured":"Zou, Henry\u00a0Peng, Yu, Gavin\u00a0Heqing, Fan, Ziwei, Bu, Dan, Liu, Han, Dai, Peng, Jia, Dongmei, Caragea, Cornelia: Eiven: Efficient implicit attribute value extraction using multimodal llm. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Track, (2024)","DOI":"10.18653\/v1\/2024.naacl-industry.40"},{"key":"797_CR371","doi-asserted-by":"crossref","unstructured":"Zou, H\u00a0P., Zhou, Y., Zhang, W., Caragea, C.: Decrisismb: Debiased semi-supervised learning for crisis tweet classification via memory bank. In The 2023 Conference on Empirical Methods in Natural Language Processing, (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.406"}],"container-title":["AI and Ethics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-025-00797-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s43681-025-00797-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-025-00797-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,19]],"date-time":"2025-09-19T08:40:00Z","timestamp":1758271200000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s43681-025-00797-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,13]]},"references-count":371,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["797"],"URL":"https:\/\/doi.org\/10.1007\/s43681-025-00797-3","relation":{},"ISSN":["2730-5953","2730-5961"],"issn-type":[{"value":"2730-5953","type":"print"},{"value":"2730-5961","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,8,13]]},"assertion":[{"value":"14 November 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 July 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 August 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The ethical requirement is fulfilled.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}