TY - JOUR
T1 - Medical Teacher’s first ChatGPT’s referencing hallucinations
T2 - Lessons for editors, reviewers, and teachers
AU - Masters, Ken
N1 - Funding Information:
The author(s) reported there is no funding associated with the work featured in this article.
Publisher Copyright:
© 2023 Informa UK Limited, trading as Taylor & Francis Group.
PY - 2023
Y1 - 2023
N2 - Students’ inappropriate use of ChatGPT is a concern. There is also, however, the potential for academics to use ChatGPT inappropriately. After explaining ChatGPT’s “hallucinations” regarding citing and referencing, this commentary illustrates the problem by describing the detection of the first known Medical Teacher submission using ChatGPT inappropriately, the lessons that can be drawn from it for journal editors, reviewers, and teachers, and then the wider implications if this problem is left unchecked.
AB - Students’ inappropriate use of ChatGPT is a concern. There is also, however, the potential for academics to use ChatGPT inappropriately. After explaining ChatGPT’s “hallucinations” regarding citing and referencing, this commentary illustrates the problem by describing the detection of the first known Medical Teacher submission using ChatGPT inappropriately, the lessons that can be drawn from it for journal editors, reviewers, and teachers, and then the wider implications if this problem is left unchecked.
KW - ChatGPT
KW - Ethics
KW - GPT-4
KW - artificial intelligence
KW - hallucinations
KW - medical education
UR - http://www.scopus.com/inward/record.url?scp=85159265919&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85159265919&partnerID=8YFLogxK
U2 - 10.1080/0142159X.2023.2208731
DO - 10.1080/0142159X.2023.2208731
M3 - Comment/debate
C2 - 37183932
AN - SCOPUS:85159265919
SN - 0142-159X
VL - 45
SP - 673
EP - 675
JO - Medical Teacher
JF - Medical Teacher
IS - 7
ER -