According to a recent report in Nature, the use of generative AI in academic writing has seen explosive growth. Research shows that 10% of abstracts in PubMed, the largest biomedical database, are suspected to involve AI writing, equivalent to about 150,000 papers annually.
A study from the Berlin University of Applied Sciences found that mainstream AI-generated content detection tools have an average accuracy of only 50%, and often misidentify human-written content as AI-generated. Many AI-generated papers can easily evade detection through paraphrasing and synonym substitution. Additionally, AI tool usage by native English speakers is harder to detect.
While AI tools have been widely used in academia before, using generative AI to directly output papers or ghostwrite remains controversial. AI tools make plagiarism easier and may lead to copyright infringement.
AI-assisted writing is not without merit. Many scholars have avoided the trouble of publishing papers in unfamiliar languages by using generative AI, allowing them to focus on research itself. Many journals now allow the use of generative AI tools but require authors to disclose usage details in the paper.
The study from the University of Tübingen analyzed 14 million abstracts in PubMed from 2010-2024. They found an abnormal surge in the use of certain modifying stylistic words after the emergence of generative AI tools like ChatGPT. The frequency of these words was used to estimate the proportion of AI-written abstracts.
The researchers also found differences in AI tool usage across countries. Their data showed papers from countries like China and South Korea used AI writing tools more frequently than those from English-speaking countries. However, usage by authors from English-speaking countries may be harder to detect.
The use of generative AI in academic writing has raised two major issues. First, plagiarism has become easier, as plagiarists can use AI to paraphrase others' research in academic journal style, making it difficult to detect. Second, AI models may output copyrighted content without attribution, as seen in the lawsuit by The New York Times against OpenAI.
To address the proliferation of AI tool usage, many companies have launched AI-generated content detection tools. However, these tools have largely failed in the "cat and mouse game" with generative AI. A study from the Berlin University of Applied Sciences found that only 5 out of 14 commonly used academic AI detection tools achieved over 70% accuracy, with an average accuracy of only 50-60%.
These detection tools perform even worse on AI-generated content that has been manually edited or machine-paraphrased. Simple operations like synonym replacement and sentence restructuring can reduce the accuracy of detection tools to below 50%. The study concluded that the overall detection accuracy of these tools is only about 50%.
The detection tools show high accuracy in identifying human-written papers. However, if an author writes an original paper in their native language and then uses translation software to translate it to another language, it may be misidentified as AI-generated. This could severely damage the academic reputation of scholars and students.
However, generative AI tools have indeed brought convenience to some researchers. Hend Al-Khalifa, an IT researcher at King Saud University, shared that before generative AI tools, many colleagues not proficient in English faced significant obstacles in paper writing. Now, these scholars can focus on the research itself without spending too much time on writing.
The boundary between AI-assisted writing and academic misconduct is difficult to define. Soheil Feizi, a computer scientist at the University of Maryland, believes that using generative AI to paraphrase existing paper content is clearly plagiarism. However, using AI tools to assist in expressing ideas should not be punished. Researchers can use detailed prompts to generate text or use AI tools to edit drafts, provided they actively disclose the use of AI tools.
Many journals have regulated the use of AI tools in academic writing without outright banning them. Science stipulates that AI cannot be listed as a co-author, and authors should disclose the AI systems and prompts used, and be responsible for content accuracy and potential plagiarism. Nature requires researchers to record the use of generative AI tools in the "Research Methods" section. As of October 2023, 87 out of the top 100 ranked journals had established guidelines for using generative AI tools.
Taking an antagonistic stance towards generative AI tools in academic research may not solve the problem at its root. Scholars from the Berlin University of Applied Sciences emphasized that the misuse of AI in academic writing is difficult to address through AI detection alone. Adjusting the academic climate that focuses on papers and results is key to solving this problem.