Skip to main content

Bibliometric analysis of ChatGPT in medicine

Abstract

Introduction

The emergence of artificial intelligence (AI) chat programs has opened two distinct paths, one enhancing interaction and another potentially replacing personal understanding. Ethical and legal concerns arise due to the rapid development of these programs. This paper investigates academic discussions on AI in medicine, analyzing the context, frequency, and reasons behind these conversations.

Methods

The study collected data from the Web of Science database on articles containing the keyword “ChatGPT” published from January to September 2023, resulting in 786 medically related journal articles. The inclusion criteria were peer-reviewed articles in English related to medicine.

Results

The United States led in publications (38.1%), followed by India (15.5%) and China (7.0%). Keywords such as “patient” (16.7%), “research” (12%), and “performance” (10.6%) were prevalent. The Cureus Journal of Medical Science (11.8%) had the most publications, followed by the Annals of Biomedical Engineering (8.3%). August 2023 had the highest number of publications (29.3%), with significant growth between February to March and April to May. Medical General Internal (21.0%) was the most common category, followed by Surgery (15.4%) and Radiology (7.9%).

Discussion

The prominence of India in ChatGPT research, despite lower research funding, indicates the platform’s popularity and highlights the importance of monitoring its use for potential medical misinformation. China’s interest in ChatGPT research suggests a focus on Natural Language Processing (NLP) AI applications, despite public bans on the platform. Cureus’ success in publishing ChatGPT articles can be attributed to its open-access, rapid publication model. The study identifies research trends in plastic surgery, radiology, and obstetric gynecology, emphasizing the need for ethical considerations and reliability assessments in the application of ChatGPT in medical practice.

Conclusion

ChatGPT’s presence in medical literature is growing rapidly across various specialties, but concerns related to safety, privacy, and accuracy persist. More research is needed to assess its suitability for patient care and implications for non-medical use. Skepticism and thorough review of research are essential, as current studies may face retraction as more information emerges.

Introduction

The emergence of AI chat programs presents two paths: one in which it is used to enhance and optimize the way we interact with queries and problems, becoming a spark that ushers in a new era of rapid academic and technological development, and another in which it is used to replace the need for one’s personal understanding. Additionally, it poses a multitude of ethical considerations and has even created certain legal gray areas, indicative of its development surpassing the speed of societies. This paper will serve as a means to better understand the context of how academic circles, countries, and institutions are discussing AI within the realm of medicine, how much they are discussing it, and conjecturing into why these conversations are occurring using supporting research.

ChatGPT is a Natural Language Processing (NLP) AI software that generates responses to any query a user may input [1]. It provides quick and clear responses and, as a result, is used widely in the same capacity as a search engine but with a more dynamic ability to interpret complicated questions, compile relevant information, and respond. Holders of professional titles such as PhDs are predicted to be affected by this; they may be at risk of decreasing importance due to AI’s ability to generate the same accurate and precise reports, curtailing the novelty of such research [2].

Although only existed for less than a year among the public, ChatGPT has made a significant impact on higher education and a variety of academic disciplines, including medicine. ChatGPT’s potential use in medicine arises from its success in aiding with diagnosis and decision-making due to its efficiency, timeliness, and access to a vast wealth of research and information. This allows it to compare medical knowledge between institutions globally, enhance communication among patients and hospital workers, and even assist in answering questions, whether they be medical queries, dosing information, or even medical exams [2, 3]. . Another recent use of the platform, which has been considered to simplify the process of medical writing, is its ability to extract medical information and perform searches to create research drafts [1].

Methods

The data set collected was obtained using an advanced search on the Web of Science database for the keyword “ChatGPT,” not case sensitive, resulting in 1440 articles published from January 1st, 2023 to September 30th, 2023. Web of Science was used because of its unique positioning as an interdisciplinary hub for global research boasting its representation of over 256 disciplines and 15 million researchers, and its ability to portray a general climate of research. The Web of Science database also allows filtering by index and category, which was performed to only allow articles from the Science Citation Index Expanded (SCI-EXPANDED) and Emerging Sources Citation (ESCI) due to their medical focus. The categories were also limited to those pertaining to the medical field; all others were excluded. Analysis was performed on the remaining 786 medically related journal articles containing the keyword “ChatGPT”.

The criterion for inclusion was that the articles be peer-reviewed, pertaining to the medical field, within the data range analyzed, and in the English language. The title, abstract, and keywords of the included studies all included “ChatGPT”.

The Web of Science also allows for data collection of bibliometric information, which can later be imputed into specialized software for interpretation. This interpretation was performed using VOSviewer 1.6.19, a software specialized in analyzing articles bibliometrically producing visual and analytic findings for speculation. The metadata acquired was then examined using 4 different categorical distinctions -- country of origin, journal, month of publication, and keywords -- and mapped using the same software.

Results

There were 786 documents retrieved from the Web of Science Core Collection, after filtering for medical-related research that used the term ChatGPT in 2023. The country leading in total number of publications was the US at 38.1% [Fig. 1]. Following the US was India and the People’s Republic of China at 15.5% and 7.0% respectively. England and Australia were the fourth and fifth top contributors with 7.5% and 6.5% of total publications [Fig. 2].

Fig. 1
figure 1

Geographical collaboration network heat map of publications Jan - Sep 2023

Fig. 2
figure 2

Treemap chart of geographic contributions to publications Jan - Sep 2023. The three most prevalent keywords that appear are patient(16.7%), research(12%), and performance(10.6%). Other topically relevant keywords were medical education(4.5%), ethic(1.8%), and plagiarism(1.3%)

The majority of articles analyzed were published in the Cureus Journal of Medical Science(11.8%) [Fig. 3]. Annals of Biomedical Engineering followed with the second most publications(8.3%). The third most published journal, Aesthetic Surgery Journal, is significantly less prolific contributing to 2.3% of all the documents analyzed, 72.3% fewer articles than the Annals of Biomedical Engineering. Both the Aesthetic Plastic Journal and the Radiology Journal have published 1.4% percent of all the articles retrieved, tying them as the fourth most substantial journal on ChatGPT. All remaining journals contained eight or fewer(≤ 1.0%) articles on the topic of ChatGPT.

Fig. 3
figure 3

Publication bibliographic coupling by journals Jan - Sep 2023

When indexing the articles by month, January 2023 had the least publications(≤ 0.005%) while August 2023, the latest month included in data collection, had the most(29.3%) [Fig. 4]. Of growth, from February to March and from April to May experienced the largest in terms of percentage with 1700% and 134%. The month of August exhibited the highest increase of papers published with 104 more papers than in July, increasing the rate of publication to 83%. June was the only month that did not experience research growth, declining 10% from the previous month.

The Web of Science categories are determined by up to 6 tags associated with any given journal. Web of Science designates its tags using subjects of the journal, author and editorial affiliations, funding, citations, and other elements such as a journal’s bibliographic categorization in other databases and a journal’s sponsors. The category of Medicine General Internal includes the most articles included in the analysis(21.0%), followed by Surgery(15.4%), Radiology(7.9%), and Health Care Science Services (7.8%). All remaining categories contained less than 5% of the journals included in this study.

Fig. 4
figure 4

Web of Science Categories Treemap chart of publications Jan - Sep 2023

The 5 most cited articles were also collected for their merit in understanding academic discussions surrounding ChatGPT. The most cited article (121 citations) was “ChatGPT utility in healthcare education, research, and practice: Systematic review on the promising perspectives and valid concerns,” published in March 2023 [4]. “ChatGPT and other large language models are double-edged swords” and “ChatGPT: the future of discharge summaries” were the second most referenced articles, with 90 citations each [5, 6]. “ChatGPT and the future of medical writing” had 84 citations, and “Nonhuman “authors” and implications for the integrity of scientific publications and medical knowledge” had 77 citations [7, 8](Biswas, 2023; Flanagin et al., 2023).

Discussion

Countries

Analyzing the three leading countries in publications regarding ChatGPT yields results that deviate from the norm of global research efforts. The leading country in publications, the United States (US), is often considered the highest researching entity due to spending more on research than any other country and offers the least notable finding [9]. India historically has low research funding, though it is improving, which draws particular attention to their dominance in ChatGPT research [10, 11]. Elucidating this finding is India’s status as the second most avid country using ChatGPT, accounting for 8.5% of the total traffic [11]. Many concerns with ChatGPT’s use and accuracy in medicine go beyond clinical settings and focus on how public use of the platform could lead to medical misinformation. For this reason, the popularity of the platform publicly and privately throughout the country, being mirrored by the country’s research institutions and funding efforts, is crucial in managing potential medical misuse of ChatGPT without explicit medical supervision.

The People’s Republic of China, which is the third most publishing country on ChatGPT with 9.8%, does not allow public access to the platform [12]. With major universities in the country also passing explicit bans on the platform, findings suggest that ChatGPT is researched abstractly on ethical grounds or with special permissions. Furthermore, these findings suggest that the interest lies not with ChatGPT as a program, but with NLP AI and its applications more generally. Additionally, funding for research in China is set to eclipse the US and tension can be seen with competition for NLP AI research superiority [9, 13, 14]. Further research on the nature of ChatGPT research in China specifically is needed, but findings thus far appear to demonstrate that national bans on the platform do not affect publication outputs.

Journals

Cureus is an open-access, peer-reviewed, general medical journal and currently is the most prolific in its publication of articles surrounding ChatGPT. Cureus was one of the first journals to issue a call for papers specifically using ChatGPT, which likely has contributed to its numbers. While Cureus is not technologically specified - as compared to the second most published journal in ChatGPT, Annals of Biomedical Engineering - its dominance in the space can be attributed to its new age structure and business model. Of note, Cureus was one of the first journals to issue a call for papers specifically using ChatGPT, which likely has contributed to its numbers. Cureus’ success in publishing articles on ChatGPT may also hinge on its ability to peer-review articles and publish submitted work quickly [15]. As such, it participates in “rapid research,” or the practice of publishing articles to appropriately respond to the void of information on what was previously an under-researched topic. What was observed in “rapid research” during the beginning of the COVID-19 pandemic is that while the speed of publication of articles was 11.5 times faster than publications on influenza, the rate of retractions and withdrawals was also significantly higher [16,17,18].

Medical disciplines

The Aesthetic Surgery Journal and Aesthetic Plastic Journal were two of the most prolific in publishing articles on ChatGPT and surgery was the second largest category, raising questions about ChatGPT’s usage in healthcare fields such as plastic surgery. A bibliometric study that addresses a plastic surgeon’s use of ChatGPT specifically yields four key findings: use in research and creation of original work, clinical applications, surgical education, and ethics/commentary on previous studies [19]. These findings are reinforced by our study, particularly in regard to the prevalence of the keywords medical education, research, and ethics. The mirroring of these interests fortifies the claim that ChatGPT has significant merit in medical education, with evidence to support that it is being examined for use in surgery but that ethical considerations remain a concern.

Literature surrounding radiology also demonstrates its use in innovative procedures and potential use for mitigating physician workloads [20]. Our study shows that it was the third most published category, suggesting it is particularly applicable to innovations in radiology, and similar studies support this claim [21]. ChatGPT is able to learn as it is fed more data and additionally excels at image analysis and pattern recognition. With physician burnout plaguing the healthcare industry, ChatGPTs use in automating such tasks serves as a potential solution [22]. AI is still subject to error however, and requires careful review if it is to be used in such a way, as to ensure patient safety, demonstrating a need for large in-depth studies into the platform’s reliability as a means of physician automation [20].

On May 24th, 2023 a bibliometric analysis of ChatGPT in Obstetrics and Gynecology (OBGYN) during a 69-day period found 0 relevant articles on its application [23]. Our study’s findings show a significant increase with 28 articles categorized under obstetric gynecology reflecting ChatGPTs adoption by more disciplines. Additionally, the disciplines of oncology, nursing, and medical informatics are all represented significantly in the top 10 categories of ChatGPT medical research. ChatGPT and NLP AI’s uses are extraordinarily dynamic; as more research is being done on its accuracy generally in medicine, more disciplines have begun incorporating it practically.

A similar study to this was conducted by Barrington et al. in Sep of 2023. The findings are mirrored by our results even with the addition of a wider date range of data collection. This demonstrates a trend in the direction of ChatGPT’s medical research and also solidifies the need for unresolved gaps in research to be addressed, namely into the limitations of ChatGPT ethically and in regards to accuracy and safety [24].

Limitations

Among the limitations of this study is its reliance on the Web of Science as the sole source of data. While Web of Science is expansive, there are discrepancies between it and a database such as PubMed, particularly in newer articles [25]. Additionally, this study only examined the keyword ChatGPT and did not explicitly include other forms of NLP AI limiting its ability to create a general image of this technology in medicine.

As with all medical research, academic interests and concerns are bound to change with the addition of new articles. Especially in the case of ChatGPT given the rapid research that surrounds it and its position as an avant-garde tool, even weeks after the publication of this article can see a reshuffling in research priorities. Despite this fact, the study provided does encompass the largest bibliometric date range on ChatGPT as of this time.

Conclusion

The literature on ChatGPT in medicine is extensive considering how new the platform is. Many medical specialties are exploring applications of the platform, and this study has shown that month over month an increasing number of disciplines are getting involved. Much of this research shares the same limitations, the safety, privacy, and accuracy of using ChatGPT for patient care. This gap in the literature needs further research if proposed applications are to be put into practice. Our analysis also emphasizes much-needed skepticism in reviewing said research, as much of the current studies could be at risk for retraction as more information is found. Concerns around the use of ChatGPT’s use medically in nonclinical settings are also found, a topic that is sorely underrepresented in current findings. The conclusion of this paper necessitates that more research be done into ChatGPT’s reliability for providing appropriate patient care in order to allow for applications in clinical settings, and the implications of ChatGPT’s use in non-medically trained hands.

Data availability

No datasets were generated or analysed during the current study.

References

  1. Chatgpt and the Future of Medical Writing. | Radiology,doi.org/10.1148/radiol.223312. Accessed 1 Jan 2024.

  2. Karthikeyan C. Literature Review on pros and cons of ChatGPT implications in Education. Int J Sci Res. 2023. https://doi.org/10.21275/SR23219122412.

    Article  Google Scholar 

  3. American Medical Association. (Accessed 1. Jan 2024). ChatGPT passed the USMLE. What does it mean for med ed? American Medical Association. https://www.ama-assn.org/practice-management/digital/chatgpt-passed-usmle-what-does-it-mean-med-ed.

  4. Sallam M. ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthc (Basel Switzerland). 2023;11(6):887. https://doi.org/10.3390/healthcare11060887.

    Article  Google Scholar 

  5. Patel SB, Lam K. ChatGPT: the future of discharge summaries? The Lancet. Digit Health. 2023;5(3):e107–8. https://doi.org/10.1016/S2589-7500(23)00021-3.

    Article  CAS  Google Scholar 

  6. Shen Y, Heacock L, Elias J, Hentel KD, Reig B, Shih G, Moy L. ChatGPT and other large language models are double-edged swords. Radiology. 2023;307(2):e230163. https://doi.org/10.1148/radiol.230163.

    Article  PubMed  Google Scholar 

  7. Biswas S. ChatGPT and the future of medical writing. Radiology. 2023;307(2):e223312. https://doi.org/10.1148/radiol.223312.

    Article  PubMed  Google Scholar 

  8. Flanagin A, Bibbins-Domingo K, Berkwits M, Christiansen SL. Nonhuman authors and implications for the integrity of scientific publication and medical knowledge. JAMA: J Am Med Association. 2023;329(8):637. https://doi.org/10.1001/jama.2023.1344.

    Article  Google Scholar 

  9. Amy Burke,  , Okrent, A., & Hale, K. (n.d.). The state of U.s. science and engineering 2022. Nsf.gov. Retrieved December 17, 2023, from https://ncses.nsf.gov/pubs/nsb20221/u-s-and-global-research-and-development.

  10. Van Noorden R. India by the numbers. Nature. 2015;521(7551):142–3. https://doi.org/10.1038/521142a.

    Article  CAS  PubMed  Google Scholar 

  11. Dandona L, Dandona R, Kumar GA, Cowling K, Titus P, Katoch VM, Swaminathan S. Mapping of health research funding in India. Natl Med J India. 2017;30(6):309–16. https://doi.org/10.4103/0970-258X.239069.

    Article  PubMed  Google Scholar 

  12. Hung J, Chen J. The benefits, risks and regulation of using ChatGPT in Chinese academia: a content analysis. Social Sci (Basel Switzerland). 2023;12(7):380. https://doi.org/10.3390/socsci12070380.

    Article  Google Scholar 

  13. Puderbaugh AP, Ellis AP, Payne JW, Scutti S, Conway C. (Jan/Feb 2020). China overtaking US as global research leader. Global Health Matters, 19(1).

  14. Reshetnikova MS. Will China win the AI race? Lecture notes in networks and systems. Springer International Publishing; 2021. pp. 2064–74.

  15. Adler J. A new age of peer reviewed scientific journals. Surg Neurol Int. 2012;3(1):145. https://doi.org/10.4103/2152-7806.103889.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Schonhaut L, Costa-Roldan I, Oppenheimer I, Pizarro V, Han D, Díaz F. Scientific publication speed and retractions of COVID-19 pandemic original articles. Revista Panam De Salud Publica [Pan Am J Public Health]. 2022;461. https://doi.org/10.26633/rpsp.2022.25.

  17. Khan H, Gupta P, Zimba O, Gupta L. Bibliometric and altmetric analysis of retracted articles on COVID-19. J Korean Med Sci. 2022;37(6). https://doi.org/10.3346/jkms.2022.37.e44.

  18. Standish K. Retracted article: COVID-19, suicide, and femicide: Rapid Research using Google search phrases. J Gen Psychol. 2021;148(3):305–26. https://doi.org/10.1080/00221309.2021.1874863.

    Article  PubMed  Google Scholar 

  19. Liu HY, Alessandri-Bonetti M, Arellano JA, Egro FM. Can ChatGPT be the plastic surgeon’s new digital assistant? A bibliometric analysis and scoping review of ChatGPT in plastic surgery literature. Aesthetic Plast Surg. 2023. https://doi.org/10.1007/s00266-023-03709-0.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Srivastav S, Chandrakar R, Gupta S, Babhulkar V, Agrawal S, Jaiswal A, Prasad R, Wanjari MB. ChatGPT in radiology: the advantages and limitations of artificial intelligence for medical imaging diagnosis. Cureus. 2023;15(7). https://doi.org/10.7759/cureus.41435.

  21. Bera K, O’Connor G, Jiang S, Tirumani SH, Ramaiya N. Analysis of ChatGPT publications in radiology: literature so far. Curr Probl Diagn Radiol. 2023. https://doi.org/10.1067/j.cpradiol.2023.10.013.

    Article  PubMed  Google Scholar 

  22. Yates SW. Physician stress and burnout. Am J Med. 2020;133(2):160–4. https://doi.org/10.1016/j.amjmed.2019.08.034.

    Article  PubMed  Google Scholar 

  23. Levin G, Brezinov Y, Meyer R. Exploring the use of ChatGPT in OBGYN: a bibliometric analysis of the first ChatGPT-related publications. Arch Gynecol Obstet. 2023;308(6):1785–9. https://doi.org/10.1007/s00404-023-07081-x.

    Article  PubMed  Google Scholar 

  24. Barrington NM, Gupta N, Musmar B, Doyle D, Panico N, Godbole N, Reardon T, D’Amico RS. A bibliometric analysis of the rise of ChatGPT in medical research. Med Sci (Basel Switzerland). 2023;11(3):61. https://doi.org/10.3390/medsci11030061.

    Article  Google Scholar 

  25. Falagas ME, Pitsouni EI, Malietzis GA, Pappas G. Comparison of PubMed, Scopus, web of Science, and Google Scholar: strengths and weaknesses. FASEB Journal: Official Publication Federation Am Soc Experimental Biology. 2008;22(2):338–42. https://doi.org/10.1096/fj.07-9492lsf.

    Article  CAS  Google Scholar 

Download references

Funding

No funding was received for this study.

Author information

Authors and Affiliations

Authors

Contributions

SG and MG wrote the main manuscript text and prepared figures. LG supervised the project and data analysis. All authors reviewed the manuscript.

Corresponding author

Correspondence to Latha Ganti.

Ethics declarations

Ethical approval

The study was performed in accordance with the ethical standards as laid down in the 1964 Declaration of Helsinki and its later amendments or comparable ethical standards. The University of Central Florida Institutional Board Review determined this study to be exempt.

Consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

Dr. Latha Ganti has an editorial role at Springer Nature.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Gande, S., Gould, M. & Ganti, L. Bibliometric analysis of ChatGPT in medicine. Int J Emerg Med 17, 50 (2024). https://doi.org/10.1186/s12245-024-00624-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12245-024-00624-2