Abstract

The past few years has seen an explosion in the availability and usage of artificial intelligence (AI) in all fields. Medicine, and more specifically orthopaedics, has been no exception. AI using machine learning has the potential to help researchers analyze huge datasets to develop predictive models for injury and improve treatment plans. The use of large language models (LLMs) like Chat GPT-4 in scientific production has also opened new avenues for efficiency in content generation and summarization. 3 However, their integration into the writing and peer review processes in medical journals raises concerns surrounding ethics, legality, copyright, and the integrity of peer review.3,5,11 This editorial will examine the benefits, risks and dangers of the “wild-west” of LLMs in orthopaedic publications (Figure 1). 3

Image created by Dall-e based on content of this editorial.
The Benefits of LLMs
Chat GPT-4 and other LLMs can assist with the production of a manuscript. 3 Generative AI can be utilized to quickly identify background articles and help create a bibliography. In the writing process, LLMs can assist in formatting an article, 13 correcting grammatical mistakes, improving style, and making manuscripts more readable. 12 Most orthopaedic journals are published in the English language. For non-native speakers, LLMs can quickly translate articles from their native language into English.
Editors and publishers can also use AI for their benefit. 15 LLMs can ensure compliance with submission guidelines, scan references, and detect duplicate submissions or potential plagiarism. 14 AI can also assist in identifying potential reviewers and provide timely copyediting. LLMs can help expedite the manuscript processing system and allow more timely publication of orthopaedic research. 4
The Risks of LLMs
LLMs can also be used to generate ideas and texts. A recent study showed an increase in the use of LLMs in the writing or articles submitted to an orthopaedic journal.2,9 The use of chatbots to produce content comes with risks. It is well documented that generative AI can produce hallucinations, creating facts that have not been proven.1,7 Additionally, since LLMs synthesize responses based on existing data patterns, there is a risk that their outputs could closely resemble existing literature, leading to accusations of plagiarism. Additionally, the risk of data privacy violations is significant if patient information or sensitive clinical data is inadvertently disclosed in published articles. The World Association of Medical Editors (WAME) has stated that chatbots cannot be authors as they don't have the ability to give “final approval” of a manuscript or understand conflict of interest statements. 18 WAME has additionally recommended that authors must disclose the use of AI and that they are responsible for all material generated by chatbots.13,17
For the American Journal of Sports Medicine (AJSM) and the Orthopaedic Journal of Sports Medicine (OJSM), authors are required to declare any use of AI and specify the use. AI should only be used to correct language, not to generate text. Fabrication of research by any means is a major violation of ethics.
Some investigators have explored whether LLMs can be used in the peer review system. 6 Chat GPT-4—generated reviews have been shown to be somewhat useful in analyzing research papers. 4 However, AI will not replace the expertise of human peer review given that AI is based on existing knowledge while research involves the exploration of innovative ideas. 15 We have identified the use of LLMs in some reviews submitted to AJSM. LLM-generated reviews could unintentionally perpetuate biases or overlook methodological flaws due to their reliance on patterns rather than a deep understanding of the research context. 5 Reviewers relying on these models to evaluate submissions might produce assessments that lack the nuanced, critical judgment required to validate research findings. A further concern is that using an Open AI system to generate peer-review breaches the confidentiality of the submitted papers. 16 Authors retain the copyright of their work and uploading manuscripts to a chatbot violates that copyright. It is the policy of AJSM/OJSM that reviewers should not use AI to write a review.
The Dangers of LLMs
Medical publishing requires stringent ethical standards to ensure the quality and reliability of research. LLMs could inadvertently compromise these standards by propagating biases embedded in their training data, potentially affecting how research is interpreted and applied in clinical practice. 11 Additionally, the risk of data privacy violations is significant if patient information or sensitive clinical data is inadvertently disclosed in published articles. 8 Medical publishers must navigate strict data protection. If LLMs introduce sensitive information into manuscripts, journals and authors could be exposed to liability. 16
Even more concerning is the ability of artificial intelligence to be used to create intentionally fraudulent manuscripts. Chat GPT-4 has been shown to be capable of creating a highly convincing fraudulent manuscript in approximately 1 hour. 8 Paper mills are businesses that sell authorship of fake or poor-quality manuscripts. It has been estimated that 3% of all medical publications in 2022 resembled paper mill productions. 17 The ability of generative AI to quickly produce fraudulent research papers will only make this problem worse. Authors, reviewers, editors, and readers must be vigilant in identifying fraudulent manuscripts. 10
Conclusion
LLMs offer transformative potential for medical publishing, enhancing productivity and access to information. However, their use in manuscript drafting and peer review processes brings ethical, legal, copyright, and review integrity issues to the forefront. Editors and publishers of orthopaedic journals must develop clear guidelines on the use of LLMs to ensure transparency in authorship, originality of content, and fairness in peer review. Ultimately, fostering collaboration between researchers, publishers, and AI developers will help address these challenges while maximizing the benefits that LLMs can bring to scientific advancement. Until then, the use of artificial intelligence in orthopaedic publishing is in a “wild-west” phase. For the time being, it is critical that authors, reviewers, and editors are aware of the good, the bad, and the ugly of these LLMs.
Footnotes
This editorial has been copublished in The American Journal of Sports Medicine.
