Distinguishing Human From Machine: A Review of Advances and Challenges in AI-Generated Text Detection

Authors

  • Serena Fariello University of Salerno image/svg+xml
  • Giuseppe Fenza University of Salerno
  • Flavia Forte University of Salerno
  • Mariacristina Gallo University of Salerno
  • Martina Marotta University of Salerno

DOI:

https://doi.org/10.9781/ijimai.2024.12.002

Keywords:

Generated-Text Detection, AI-Detection, Large Language Models, Literature Review, Survey
Supporting Agencies
This work was partially supported by project SERICS (PE00000014) under the MUR National Recovery and Resilience Plan funded by the European Union - NextGenerationEU.

Abstract

The rise of Large Language Models (LLMs) has dramatically altered the generation and spreading of textual content. This advancement offers benefits in various domains, including medicine, education, law, coding, and journalism, but also has negative implications, mainly related to ethical concerns. Preventing measures to mitigate negative implications pass through solutions that distinguish machine-generated text from humanwritten text. This study aims to provide a comprehensive review of existing literature for detecting LLMgenerated texts. Emerging techniques are categorized into five categories: watermarking, feature-based, neural-based, hybrid, and human-aided methods. For each introduced category, strengths and limitations are discussed, providing insights into their effectiveness and potential for future improvements. Moreover, available datasets and tools are introduced. Results demonstrate that, despite the good delimited performance, the multitude of languages to recognize, hybrid texts, the continuous improvement of algorithms for text generation and the lack of regulation require additional efforts for efficient detection.

Downloads

Download data is not yet available.

References

F. García-Peñalvo, A. Vázquez-Ingelmo, et al., “What do we mean by genai? a systematic mapping of the evolution, trends, and techniques involved in generative ai,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 8, no. 4, pp. 7–16, 2023.

J. Oluwaseyi, K. Potter, “Exploring natural language generation (nlg) methods for generating human-like text from structured or unstructured data,” Journal of Machine to Machine Communications, 12 2023.

A. Tlili, B. Shehata, M. A. Adarkwah, A. Bozkurt, D. T. Hickey, R. Huang, B. Agyemang, “What if the devil is my guardian angel: Chatgpt as a case study of using chatbots in education,” Smart learning environments, vol. 10, no. 1, p. 15, 2023.

M. Alier, F. J. García-Peñalvo, J. D. Camba, “Generative Artificial Intelligence in education: From deceptive to disruptive,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 8, no. 5, Special issue on Generative Artificial Intelligence in Education, pp. 5–14, 2024.

J. Wu, S. Yang, R. Zhan, Y. Yuan, D. F. Wong, L. S. Chao, “A survey on llm-generated text detection: Necessity, methods, and future directions,” ArXiv, vol. abs/2310.14724, 2023.

A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin, “Attention is all you need,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Red Hook, NY, USA, 2017, p. 6000–6010, Curran Associates Inc.

Z. Yang, “Xlnet: Generalized autoregressive pretraining for language understanding,” arXiv preprint arXiv:1906.08237, 2019.

S. Grassini, “Shaping the future of education: exploring the potential and consequences of ai and chatgpt in educational settings,” Education Sciences, vol. 13, no. 7, p. 692, 2023.

J. Jeon, S. Lee, “Large language models in education: A focus on the complementary relationship between human teachers and chatgpt,” Education and Information Technologies, 05 2023, doi: 10.1007/s10639-023-11834-1.

V. Parra, P. Sureda, A. Corica, S. Schiaffino, D. Godoy, “Can Generative AI solve Geometry Problems? Strengths and Weaknesses of LLMs for Geometric Reasoning in Spanish,” International Journal of Interactive Multimedia and Artificial Intelligence, vol. 8, no. 5, Special issue on Generative Artificial Intelligence in Education, pp. 65–74, 2024, doi: 10.9781/ijimai.2024.02.009.

A. Thirunavukarasu, D. Ting, K. Elangovan, L. Gutierrez Sinisterra, T. Tan, D. Ting, “Large language models in medicine,” Nature Medicine, vol. 29, 07 2023, doi: 10.1038/s41591-023-02448-8.

S. Hajijama, D. Juneja, P. Nasa, “Large language model in critical care medicine: Opportunities and challenges,” Indian Journal of Critical Care Medicine, vol. 28, no. 6, pp. 523–525, 2024.

Y. Feng, S. Vanam, M. Cherukupally, W. Zheng, M. Qiu, H. Chen, “Investigating code generation performance of chatgpt with crowdsourcing social data,” in 2023 IEEE 47th Annual Computers, Software, and Applications Conference (COMPSAC), 2023, pp. 876–885.

R. Khoury, A. R. Avila, J. Brunelle, B. M. Camara, “How secure is code generated by chatgpt?,” in 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), 2023, pp. 2445–2451.

V. C. C. Kwok-Yan Lam, Z. K. Yeong, “Applying large language models for enhancing contract drafting,” in Proceedings of the Third International Workshop on Artificial Intelligence and Intelligent Assistance for Legal Professionals in the Digital Workspace (LegalAIIA 2023), 2023.

J. H. Choi, “How to use large language models for empirical legal research,” Journal of Institutional and Theoretical Economics (Forthcoming), 2023.

J. Cui, Z. Li, Y. Yan, B. Chen, L. Yuan, “Chatlaw: Open- source legal large language model with integrated external knowledge bases,” arXiv preprint arXiv:2306.16092, 2023.

J. Tan, H. Westermann, K. Benyekhlef, “Chatgpt as an artificial lawyer?,” in AI4AJ@ ICAIL, 2023.

X. Fang, S. Che, M. Mao, H. Zhang, M. Zhao, X. Zhao, “Bias of ai-generated content: an examination of news produced by large language models,” Scientific Reports, vol. 14, no. 1, p. 5224, 2024.

C. Novelli, F. Casolari, P. Hacker, G. Spedicato, L. Floridi, “Generative ai in eu law: Liability, privacy, intellectual property, and cybersecurity,” EU Law: Liability, Privacy, Intellectual Property, and Cybersecurity (January 14, 2024), 2024.

L. Tang, Y.-S. Su, “Ethical Implications and Principles of Using Artificial Intelligence Models in the Classroom: A Systematic Literature Review,” International Journal of Interactive Multimedia & Artificial Intelligence, vol. 8, no. 5, 2024.

I. Ulnicane, “Governance fix? power and politics in controversies about governing generative ai,” Policy and Society, p. puae022, 2024.

S. S. Ghosal, S. Chakraborty, J. Geiping, F. Huang, D. Manocha, A. Bedi, “A survey on the possibilities & impossibilities of ai-generated text detection,” Transactions on Machine Learning Research, 2023.

D. Beresneva, “Computer-generated text detection using machine learning: A systematic review,” in Natural Language Processing and Information Systems: 21st International Conference on Applications of Natural Language to Information Systems, NLDB 2016, Salford, UK, June 22-24, 2016, Proceedings 21, 2016, pp. 421–426, Springer.

G. Jawahar, M. Abdul-Mageed, V. Laks Lakshmanan, “Automatic detection of machine generated text: A critical survey,” in Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 2296–2309.

M. Dhaini, W. Poelman, E. Erdogan, “Detecting chatgpt: A survey of the state of detecting chatgpt-generated text,” in Proceedings of the 8th Student Research Workshop associated with the International Conference Recent Advances in Natural Language Processing, 2023, pp. 1–12.

E. Crothers, N. Japkowicz, H. Viktor, “Machine-generated text: A comprehensive survey of threat models and detection methods,” IEEE Access, vol. PP, pp. 1–1, 01 2023, doi: 10.1109/ACCESS.2023.3294090.

R. Tang, Y.-N. Chuang, X. Hu, “The science of detecting llm-generated text,” Communications of the ACM, vol. 67, p. 50–59, mar 2024, doi: 10.1145/3624725.

X. Yang, L. Pan, X. Zhao, H. Chen, L. R. Petzold, W. Y. Wang, W. Cheng, “A survey on detection of llms-generated content,” ArXiv, vol. abs/2310.15654, 2023.

A. Uchendu, T. Le, D. Lee, “Attribution and obfuscation of neural text authorship: A data mining perspective,” ACM SIGKDD Explorations Newsletter, vol. 25, no. 1, pp. 1–18, 2023.

Chenxi Gu and Chengsong Huang and Xiaoqing Zheng and Kai-Wei Chang and Cho-Jui Hsieh, “Watermarking Pre-trained Language Models with Backdooring,” ArXiv, vol. abs/2210.07543, 2022.

E. Lucas, T. Havens, “GPTs don’t keep secrets: Searching for backdoor watermark triggers in autoregressive language models,” in Proceedings of the 3rd Workshop on Trustworthy Natural Language Processing (TrustNLP 2023), Toronto, Canada, July 2023, pp. 242–248, Association for Computational Linguistics.

R. Tang, Q. Feng, N. Liu, F. Yang, X. Hu, “Did you train on my dataset? towards public dataset protection with cleanlabel backdoor watermarking,” ACM SIGKDD Explorations Newsletter, vol. 25, p. 43–53, jul 2023, doi: 10.1145/3606274.3606279.

J. Kirchenbauer, J. Geiping, Y. Wen, J. Katz, I. Miers, T. Goldstein, “A watermark for large language models,” in Proceedings of the 40th International Conference on Machine Learning, vol. 202 of Proceedings of Machine Learning Research, 23–29 Jul 2023, pp. 17061–17084, PMLR.

J. Kirchenbauer, J. Geiping, Y. Wen, M. Shu, K. Saifullah, K. Kong, K. Fernando, A. Saha, M. Goldblum, T. Goldstein, “On the reliability of watermarks for large language models,” in The Twelfth International Conference on Learning Representations, 2024.

A. Liu, L. Pan, X. Hu, S. Li, L. Wen, I. King, S. Y. Philip, “An unforgeable publicly verifiable watermark for large language models,” in The Twelfth International Conference on Learning Representations, 2023.

A. Hou, J. Zhang, T. He, Y. Wang, Y.-S. Chuang, H. Wang, L. Shen, B. Van Durme, D. Khashabi, Y. Tsvetkov, “Semstamp: A semantic watermark with paraphrastic robustness for text generation,” in Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024, pp. 4067–4082.

Por, Lip Yee and Wong, KokSheik and Chee, Kok Onn, “UniSpaCh: A text-based data hiding method using Unicode space characters,” Journal of Systems and Software, vol. 85, pp. 1075–1082, may 2012, doi: 10.1016/j.jss.2011.12.023.

U. Topkara, M. Topkara, M. J. Atallah, “The hiding virtues of ambiguity: quantifiably resilient watermarking of natural language text through synonym substitutions,” in Proceedings of the 8th workshop on Multimedia and security, 2006, pp. 164–174.

Munyer, Travis and Tanvir, Abdullah and Das, Arjon and Zhong, Xin, “DeepTextMark: A Deep Learning- Driven Text Watermarking Approach for Identifying Large Language Model Generated Text,” IEEE Access, vol. PP, pp. 1–1, 01 2024, doi: 10.1109/ACCESS.2024.3376693.

X. Yang, J. Zhang, K. Chen, W. Zhang, Z. Ma, F. Wang, N. Yu, “Tracing text provenance via context-aware lexical substitution,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, 2022, pp. 11613–11621.

K. Yoo, W. Ahn, J. Jang, N. Kwak, “Robust multi-bit natural language watermarking through invariant features,” in The 61st Annual Meeting Of The Association For Computational Linguistics, 2023.

X. Yang, K. Chen, W. Zhang, C. Liu, Y. Qi, J. Zhang, H. Fang, N. Yu, “Watermarking text generated by black-box language models,” arXiv preprint arXiv:2305.08883, 2023.

S. Abdelnabi, M. Fritz, “Adversarial watermarking transformer: Towards tracing text provenance with data hiding,” in 2021 IEEE Symposium on Security and Privacy (SP), 2021, pp. 121–140, IEEE.

A. Modupe, T. Celik, V. Marivate, O. Olugbara, “Post- authorship attribution using regularized deep neural network,” Applied Sciences, vol. 12, p. 7518, 07 2022, doi: 10.3390/app12157518.

T. Schuster, R. Schuster, D. Shah, R. Barzilay, “The limitations of stylometry for detecting machine-generated fake news,” Computational Linguistics, vol. 46, pp. 1–18, 03 2020, doi: 10.1162/COLI_a_00380.

T. Kumarage, J. Garland, A. Bhattacharjee, K. Trapeznikov, S. Ruston, H. Liu, “Stylometric detection of ai-generated text in twitter timelines,” arXiv preprint arXiv:2303.03697, 2023.

L. Frohling, A. Zubiaga, “Feature-based detection of automated language models: tackling gpt-2, gpt-3 and grover,” PeerJ Computer Science, vol. 7, p. e443, 04 2021, doi: 10.7717/peerj-cs.443.

A. Hans, A. Schwarzschild, V. Cherepanova, H. Kazemi, A. Saha, M. Goldblum, J. Geiping, T. Goldstein, “Spotting llms with binoculars: Zero-shot detection of machine- generated text,” arXiv preprint arXiv:2401.12070, 2024.

E. Mitchell, Y. Lee, A. Khazatsky, C. D. Manning, C. Finn, “Detectgpt: zero-shot machine-generated text detection using probability curvature,” in Proceedings of the 40th International Conference on Machine Learning, ICML’23, 2023, JMLR.org.

S. Venkatraman, A. Uchendu, D. Lee, “Gpt-who: An information density-based machine-generated text detector,” in Findings of the Association for Computational Linguistics: NAACL 2024, 2024, pp. 103–115.

A. Aich, S. Bhattacharya, N. Parde, “Demystifying neural fake news via linguistic feature-based interpretation,” in Proceedings of the 29th International Conference on Computational Linguistics, 2022, pp. 6586–6599.

J. Su, T. Zhuo, D. Wang, P. Nakov, “Detectllm: Leveraging log rank information for zero-shot detection of machine-generated text,” in Findings of the Association for Computational Linguistics: EMNLP 2023, 2023, pp. 12395–12412.

A. Bakhtin, S. Gross, M. Ott, Y. Deng, M. Ranzato, A. Szlam, “Real or fake? learning to discriminate machine from human generated text,” arXiv preprint arXiv:1906.03351, 2019.

J. D. Rodriguez, T. Hay, D. Gros, Z. Shamsi, R. Srinivasan, “Cross-domain detection of gpt-2-generated technical text,” in Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: human language technologies, 2022, pp. 1213–1233.

I. Solaiman, M. Brundage, J. Clark, A. Askell, A. Herbert- Voss, J. Wu, A. Radford, G. Krueger, J. W. Kim, S. Kreps, et al., “Release strategies and the social impacts of language models,” arXiv preprint arXiv:1908.09203, 2019.

R. Zellers, A. Holtzman, H. Rashkin, Y. Bisk, A. Farhadi, F. Roesner, Y. Choi, “Defending against neural fake news,” Advances in neural information processing systems, vol. 32, 2019.

E. Crothers, N. Japkowicz, H. Viktor, P. Branco, “Adversarial robustness of neural-statistical features in detection of generative transformers,” in 2022 International Joint Conference on Neural Networks (IJCNN), 2022, pp. 1–8, IEEE.

S. Gehrmann, H. Strobelt, A. M. Rush, “Gltr: Statistical detection and visualization of generated text,” in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 2019, pp. 111–116.

X. Hu, P.-Y. Chen, T.-Y. Ho, “Radar: Robust ai-text detection via adversarial learning,” Advances in Neural Information Processing Systems, vol. 36, pp. 15077–15095, 2023.

E. Clark, T. August, S. Serrano, N. Haduong, S. Gururangan, N. A. Smith, “All that’s ‘human’is not gold: Evaluating human evaluation of generated text,” in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 2021, pp. 7282–7296.

D. Ippolito, D. Duckworth, D. Eck, “Automatic detection of generated text is easiest when humans are fooled,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 1808–1822.

A. Uchendu, Z. Ma, T. Le, R. Zhang, D. Lee, “Turingbench: A benchmark environment for turing test in the age of neural text generation,” in Findings of the Association for Computational Linguistics: EMNLP 2021, 2021, pp. 2001–2016.

Y. Dou, M. Forbes, R. Koncel-Kedziorski, N. A. Smith, Y. Choi, “Is gpt-3 text indistinguishable from human text? scarecrow: A framework for scrutinizing machine text,” in Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022, pp. 7250–7274.

L. Kushnareva, D. Cherniavskii, V. Mikhailov, E. Artemova, S. Barannikov, A. Bernstein, I. Piontkovskaya, D. Piontkovski, E. Burnaev, “Artificial text detection via examining the topology of attention maps,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021, pp. 635–649.

X. Liu, Z. Zhang, Y. Wang, H. Pu, Y. Lan, C. Shen, “Coco: Coherence-enhanced machine-generated text detection under low resource with contrastive learning,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023, pp. 16167–16188.

R. Tan, B. Plummer, K. Saenko, “Detecting cross- modal inconsistency to defend against neural fake news,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 2081–2106.

L. Wang, W. Yang, D. Chen, H. Zhou, Y. Lin, F. Meng, J. Zhou, X. Sun, “Towards codable text watermarking for large language models,” arXiv preprint arXiv:2307.15992, 2023.

R. Kuditipudi, J. Thickstun, T. Hashimoto, P. Liang “Robust Distortion-free Watermarks for Language Models,” Transactions on Machine Learning Research, 2024.

J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, “BERT: Pre- training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, Minnesota, June 2019, Association for Computational Linguistics.

A. Muñoz-Ortiz, C. Gómez-Rodríguez, D. Vilares, “Contrasting linguistic patterns in human and llm- generated text,” arXiv preprint arXiv:2308.09067, 2023.

R. Corizzo, S. Leal-Arenas, “One-class learning for ai- generated essay detection,” Applied Sciences, vol. 13, no. 13, 2023, doi: 10.3390/app13137901.

A. Shah, P. Ranka, U. Dedhia, S. Prasad, S. Muni, K. Bhowmick, “Detecting and unmasking ai-generated texts through explainable artificial intelligence using stylistic features,” International Journal of Advanced Computer Science and Applications, vol. 14, no. 10, 2023.

Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, V. Stoyanov, “Roberta: A robustly optimized bert pretraining approach,” arXiv preprint arXiv:1907.11692, 2019.

D. Macko, R. Moro, A. Uchendu, J. Lucas, M. Yamashita, M. Pikuliak, I. Srba, T. Le, D. Lee, J. Simko, et al., “Multitude: Large-scale multilingual machine-generated text detection benchmark,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023, pp. 9960–9987.

R. Gaggar, A. Bhagchandani, H. Oza, “Machine-generated text detection using deep learning,” 2023. [Online]. Available: https://arxiv.org/abs/2311.15425.

J. A. Guerrero, “Detecting ai generated text using neural networks,” Master’s thesis, Texas A&M University, 2023.

Y. Zhang, Q. Leng, M. Zhu, R. Ding, Y. Wu, J. Song, Y. Gong, “Enhancing text authenticity: A novel hybrid approach for ai-generated text detection,” in 2024 IEEE 4th International Conference on Electronic Technology, Communication and Information (ICETCI), 2024, pp. 433–438.

C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, P. J. Liu, “Exploring the limits of transfer learning with a unified text-to-text transformer,” Journal of machine learning research, vol. 21, no. 140, pp. 1–67, 2020.

A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, I. Sutskever, et al., “Language models are unsupervised multitask learners,” OpenAI blog, vol. 1, no. 8, p. 9, 2019.

S. Merity, C. Xiong, J. Bradbury, R. Socher, “Pointer sentinel mixture models,” arXiv preprint arXiv:1609.07843, 2016.

A. Maas, R. E. Daly, P. T. Pham, D. Huang, A. Y. Ng, C. Potts, “Learning word vectors for sentiment analysis,” in Proceedings of the 49th annual meeting of the association for computational linguistics: Human language technologies, 2011, pp. 142–150.

X. Zhang, J. Zhao, Y. LeCun, “Character-level convolutional networks for text classification,” Advances in neural information processing systems, vol. 28, 2015.

P. Rajpurkar, J. Zhang, K. Lopyrev, P. Liang, “Squad: 100,000+ questions for machine comprehension of text,” arXiv preprint arXiv:1606.05250, 2016.

A. Fan, M. Lewis, Y. Dauphin, “Hierarchical neural story generation,” arXiv preprint arXiv:1805.04833, 2018.

M. Lee, P. Liang, Q. Yang, “Coauthor: Designing a human-ai collaborative writing dataset for exploring language model capabilities,” in Proceedings of the 2022 CHI conference on human factors in computing systems, 2022, pp. 1–19.

A. Richburg, C. Bao, M. Carpuat, “Automatic authorship analysis in human-ai collaborative writing,” in Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC- COLING 2024), 2024, pp. 1845–1855.

Z. Zeng, S. Liu, L. Sha, Z. Li, K. Yang, S. Liu, D. Gavsevi’c, G. Chen, “Detecting ai-generated sentences in human- ai collaborative hybrid texts: Challenges, strategies, and insights,” arXiv preprint arXiv:2403.03506v4, 2024.

P. Sen, G. Namata, M. Bilgic, L. Getoor, B. Galligher, T. Eliassi-Rad, “Collective classification in network data,” AI magazine, vol. 29, no. 3, pp. 93–93, 2008.

O. Bojar, R. Chatterjee, C. Federmann, Y. Graham, B. Haddow, M. Huck, A. J. Yepes, P. Koehn, V. Logacheva, C. Monz, et al., “Findings of the 2016 conference on machine translation (wmt16),” in First conference on machine translation, 2016, pp. 131–198, Association for Computational Linguistics.

Q. Jin, B. Dhingra, Z. Liu, W. Cohen, X. Lu, “Pubmedqa: A dataset for biomedical research question answering,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP- IJCNLP), 2019, pp. 2567–2577.

M. Bień, M. Gilski, M. Maciejewska, W. Taisner, D. Wisniewski, A. Lawrynowicz, “Recipenlg: A cooking recipes dataset for semi-structured text generation,” in Proceedings of the 13th International Conference on Natural Language Generation, 2020, pp. 22–28.

J. M. Patel, J. M. Patel, “Introduction to common crawl datasets,” Getting structured data from the internet: running web crawlers/scrapers on a big data production scale, pp. 277–324, 2020.

R. Tan, B. Plummer, K. Saenko, “Detecting cross- modal inconsistency to defend against neural fake news,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 2081–2106.

Y. Chen, Y. Liu, L. Chen, Y. Zhang, “Dialogsum: A real-life scenario dialogue summarization dataset,” in Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2021, pp. 5062–5074.

S. Auer, C. Bizer, G. Kobilarov, J. Lehmann, R. Cyganiak, Z. Ives, “Dbpedia: A nucleus for a web of open data,” in international semantic web conference, 2007, pp. 722–735, Springer.

B. Guo, X. Zhang, Z. Wang, M. Jiang, J. Nie, Y. Ding, J. Yue, Y. Wu, “How close is chatgpt to human experts? comparison corpus, evaluation, and detection,” arXiv preprint arXiv:2301.07597, 2023.

A. Wang, A. Singh, J. Michael, F. Hill, O. Levy, S. R. Bowman, “Glue: A multi-task benchmark and analysis platform for natural language understanding,” in 7th International Conference on Learning Representations, ICLR 2019, 2019.

S. Bowman, G. Angeli, C. Potts, C. D. Manning, “A large annotated corpus for learning natural language inference,” in Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 2015, pp. 632–642.

W. H. Walters, “The effectiveness of software designed to detect ai-generated writing: A comparison of 16 ai text detectors,” Open Information Science, vol. 7, no. 1, p. 20220158, 2023.

N. Ladha, K. Yadav, P. Rathore, “Ai-generated content detectors: Boon or bane for scientific writing,” Indian Journal of Science and Technology, vol. 16, no. 39, pp. 3435–3439, 2023.

G.-A. Odri, D. J. Y. Yoon, “Detecting generative artificial intelligence in scientific articles: evasion techniques and implications for scientific integrity,” Orthopaedics & Traumatology: Surgery & Research, vol. 109, no. 8, p. 103706, 2023.

M. Perkins, J. Roe, B. H. Vu, D. Postma, D. Hickerson, J. McGaughran, H. Q. Khuat, “Genai detection tools, adversarial techniques and implications for inclusivity in higher education,” arXiv preprint arXiv:2403.19148, 2024.

W. Liang, M. Yuksekgonul, Y. Mao, E. Wu, J. Zou, “Gpt detectors are biased against non-native english writers,” Patterns, vol. 4, no. 7, p. 100779, 2023, doi: https://doi.org/10.1016/j.patter.2023.100779.

D. Macko, R. Moro, A. Uchendu, J. Lucas, M. Yamashita, M. Pikuliak, I. Srba, T. Le, D. Lee, J. Simko, M. Bielikova, “MULTITuDE: Large-scale multilingual machine- generated text detection benchmark,” in Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Singapore, Dec. 2023, pp. 9960–9987, Association for Computational Linguistics.

Y. Wang, J. Mansurov, P. Ivanov, J. Su, A. Shelmanov, A. Tsvigun, O. M. Afzal, T. Mahmoud, G. Puccetti, T. Arnold, et al., “Semeval-2024 task 8: Multidomain, multimodel and multilingual machine-generated text detection,” arXiv preprint arXiv:2404.14183, 2024.

A. Bhattacharjee, R. Moraffah, J. Garland, H. Liu, “Eagle: A domain generalization framework for ai-generated text detection,” arXiv preprint arXiv:2403.15690, 2024.

O. J. of the European Union, “Regulation (eu) 2024/1689 of the european parliament and of the council.” https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32024R1689. Last accessed: 23 September 2024.

F. C. Commission, “Telephone consumer protection act.” https://docs.fcc.gov/public/attachments/DOC-404036A1.pdf. Last accessed: 23 September 2024.

Downloads

Published

2025-06-01
Metrics
Views/Downloads
  • Abstract
    357
  • PDF
    84

How to Cite

Fariello, S., Fenza, G., Forte, F., Gallo, M., and Marotta, M. (2025). Distinguishing Human From Machine: A Review of Advances and Challenges in AI-Generated Text Detection. International Journal of Interactive Multimedia and Artificial Intelligence, 9(3), 6–18. https://doi.org/10.9781/ijimai.2024.12.002