How to Use LLMs Ethically in Academic Writing?
DOI:
https://doi.org/10.70088/50wkze06Keywords:
large language models, academic writing, control experiment, qualitative content analysisAbstract
This paper presents an experimental study based on selected Large Language Models (LLMs) and Artificial Intelligence Generated Content (AIGC) detection systems, conducted within a mixed-methods research paradigm that combines empirical validation and Qualitative Content Analysis (QCA). The empirical validation process consists of both a condition optimization experiment and the main experiment, while the materials for qualitative content analysis are directly derived from these experimental outputs. In the experiments, six LLMs are evaluated using four different AIGC detectors. Through the analysis of the contents generated by these LLMs, the existing theoretical framework, which is referred to as the authors’ checklist, for the application of LLMs in academic writing is revised. The updated framework refines the checklist step for assessing and amending the accuracy of AI-generated content. The updated framework contains five steps, Intellectual Contribution, Accuracy of Conceptions, Accuracy of Demonstrations, Academic Competency, and Transparency, for authors’ academic writing with the assistance of LLMs. Additionally, it emphasizes the importance of authors’ innovation and proficiency in prompting LLMs when ethically using LLMs in academic writing.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Tiantian Yu (Author)

This work is licensed under a Creative Commons Attribution 4.0 International License.