Administrative task



Artificial intelligence (AI) is rapidly being used in medicine due to its advanced capabilities in image and videorecognition, clinical decision support, neurosurgical education, and administrative task automation. Large language models such as OpenAI's Generative Pretrained Transformer (GPT)-4 and Google's Bard have particularly revolutionized text generation, offering substantial benefits for the academic surgeon, including aiding in manuscript and grant writing. However, integrating AI into academic surgery necessitates addressing ethical concerns such as bias, transparency, and intellectual property. This paper provides guidelines and recommendations based on current literature around the opportunities and ethical challenges of AI in academic surgery. We discuss the underlying mechanisms of large language models, their potential biases, and the importance of responsible usage. Furthermore, they explore the ethical implications of AI in clinical documentation, highlighting improved efficiency and necessary privacy concerns. This review also addresses the critical issue of intellectual property dilemmas posed by AI-generated innovations in university settings. Finally, we propose guidelines for the responsible adoption of AI in academic and clinical environments, stressing the need for transparency, ethical training, and robust governance frameworks to ensure AI enhances, rather than undermines, academic integrity and patient care 1)

1)
Robinson JR, Stey A, Schneider DF, Kothari AN, Lindeman B, Kaafarani HM, Haines KL. Generative Artificial Intelligence in Academic Surgery: Ethical Implications and Transformative Potential. J Surg Res. 2025 Feb 10:S0022-4804(25)00021-6. doi: 10.1016/j.jss.2024.12.059. Epub ahead of print. PMID: 39934059.