“Those who can imagine anything, can create the impossible.”

-Alan Turing

Imagine living in a world where patients can self-diagnose their medical condition based on self-reported symptoms and/or images; or doctors can manage any medical diseases without considerable prior domain-specific knowledge; or mundane medical administrative work such as drafting hospital discharge letters can be automated.

All above scenarios, once thought to be idealistic aspirations, are now gradually morphing into attainable realities made possible by artificial intelligence (AI), particularly with the recent advent of ChatGPT. In view of the significant hype and controversies, this Editorial aims to demystify the key concepts of ChatGPT and evaluate its roles in medicine, primarily from ophthalmology perspective.

What is ChatGPT?

ChatGPT (OpenAI, San Francisco, USA) is a generative AI, large language model (LLM)-powered chatbot [1]. It has the ability to process and convert textual and now graphic inputs to text-based outputs, engaging in human-like conversations and creating new human-like contents (“generative”). The model is based on Generative Pretrained Transformer (GPT) architecture and was pre-trained in an unsupervised manner using large-scale text corpora (containing billions of words) derived from books, articles, and various internet-based content, followed by minimal supervised fine-tuning of the model for various downstream tasks (known as “few-shot learning”). The transformer architecture enables processing of the entire input and context wholly and “transforms” it to meaningful human-like text based on statistical prediction of words (guided by previous words and entire context). OpenAI has also created a GPT-3 variant, DALL-E, which can convert textual inputs to graphic outputs. In view of its astounding versatility and functionality, ChatGPT has permeated all fields, encompassing business, entertainment, education, science, legal, and healthcare sectors [2,3,4] (https://tech.co/news/10-ways-businesses-using-chatgpt).

ChatGPT has undergone several iterations, refining from GPT-1 in 2018, GPT-2 in 2019 (1.5 billion training parameters) and GPT-3 in 2020 (175 billion training parameters), to the more recent ChatGPT-3.5 in November 2022 and ChatGPT-4 in March 2023 (though the latest architecture, training data and parameters remain undisclosed). ChatGPT has quickly accrued over one billion users and a PubMed search on 14/05/2023 returned >400 ChatGPT-related results, highlighting an immense interest in this technology within the medical field.

Potential promises

ChatGPT has the potential of revolutionising medical field, including ophthalmology, in diverse ways, spanning patients, healthcare professionals/systems, research, and education/training (Fig. 1).

Fig. 1: Potential of ChatGPT in ophthalmology.
figure 1

Potential of ChatGPT in ophthalmology, spanning patients, healthcare professionals/organisations, research, education, and training.

Patients

The plausibility of deploying ChatGPT for medical purposes largely stems from its superior question-answering performance in medical exams where it could attain scores paralleling candidates in the top 10% [1, 5, 6]. However, it is important not to over-extrapolate these findings as question-answering in medical examination does not directly translate to real-world clinical practice where interpersonal skills, clinical reasoning/contextualisation, and holistic patient care are essential. Several studies specifically evaluated ChatGPT’s performance in answering ophthalmology questions and demonstrated fair-to-good accuracy (40–80%) [7,8,9]. This was influenced by question difficulty and domain (e.g. higher accuracy in general medicine than subspecialised domains). Interestingly, the model is receptive to training and refinement [8].

As such, non-medically trained individuals/patients may potentially utilise ChatGPT as a virtual assistant in self-triaging and self-diagnosing his/her ophthalmic conditions, ranging from innocuous to potentially sight-threatening ocular conditions (see examples in Fig. 2A, B). In addition, ChatGPT may be effective in generating patient educational materials, by translating jargon-heavy medical terminology into simplified and/or empathetic language tailored to the level of a layperson (Fig. 2C) [3, 10], or act as a “therapist” for counselling individuals with mental health illness [11]. This will be invaluable as patients with chronic eye diseases or visual impairment are more susceptible to psychological distress/illness [12,13,14]. ChatGPT’s multilingual translation feature can further cater to multi-ethnic patient populations, and potentially be integrated with text-to-speech audio generation (particularly useful for visually impaired patients) and text-to-image or video generation platforms to enhance patient experience.

Fig. 2: Several examples of real-life conversation with ChatGPT in relation to the field of ophthalmology (some parts of the generated responses are shortened).
figure 2

A, B An example of an individual “consulting” ChatGPT for potential subconjunctival haemorrhage and acute-angle closure glaucoma, respectively. C An example of an individual “consulting” ChatGPT to understand about cataract surgery. D An example of an individual/ophthalmologist “consulting” ChatGPT for the management of optic neuritis. Note that for any medical advice provided by ChatGPT, it is always concluded with a statement that “consultation with a healthcare professional is crucial”. E An example of an individual/ophthalmologist using ChatGPT for medical education and training purposes.

Healthcare professionals/systems

ChatGPT may empower physicians with curated, synthesised information and guidelines from otherwise diverse and disparate sources, facilitating the process of formulating differential diagnoses, decision-making, and treatment plans (Fig. 2D) [15]. Administrative tasks may potentially be automated and expedited using ChatGPT, including drafting of discharge summaries and medical reports, and answering administrative enquiries like appointment rescheduling and medication refill requests [16, 17]. ChatGPT can acknowledge and rectify its mistakes by prompting, highlighting the importance of human oversight to maximise the benefits of ChatGPT as an assistive tool [18]. It is capable of processing vast structured and unstructured clinical information stored within the electronic health records (EHR), which are increasingly adopted in ophthalmology [19, 20], (https://www.nodaudit.org.uk/) and generate concise summaries of patient record [21]. This allows the physicians to focus their consultation on clinical management and patient communication.

Research

Similarly, ChatGPT’s performance in summarisation tasks can accelerate time-consuming research process. ChatGPT can efficiently perform literature review, summarise findings, and draft abstracts/manuscripts, with quality of work being indistinguishable from clinicians/scientists [22, 23]. There are ethical debates on whether ChatGPT should be recognised as an author in research articles. Based on the ICJME criteria, many publishers, including Springer Nature, stated that any LLM, including ChatGPT, should not be included as an author in view of the lack of ability to approve the final manuscript or to take accountability of the research [24, 25]. Nevertheless, ChatGPT’s assistance should be acknowledged in publication where appropriate. Further, ChatGPT can also facilitate original research along the model development pipeline, by generating new scientific ideas, cleaning large datasets, debugging codes, assisting patient recruitment, and advancing basic science discovery [18, 26].

Education/training

ChatGPT’s encouraging performance in medical examination question-answering tasks and ability to “interact” with human show potential in enhancing medical education and training [4]. In addition to generating answers, ChatGPT can further justify the responses with supplementary logical explanations of high concordance and insight [27,28,29]. The multilingual function of ChatGPT can be utilised to improve the learning experience of students with language barrier, by using ChatGPT’s response as a benchmark and tool to enhance their language proficiency and writing styles [4]. It can also generate ideas, strategies and exam questions or quizzes to stimulate learning (Fig. 2E).

Potential pitfalls

As a generative AI, ChatGPT generates outputs based on statistical prediction of the text without human-like reasoning. This can potentially result in plausible-sounding but inaccurate responses, known as “hallucination” or “fabrication” [30]. Lack of recency is another limitation as the model was trained with data up to September 2021. ChatGPT would have missed recent advancements in medicine and ophthalmology (which is rapidly evolving) (https://investors.apellis.com/news-releases/news-release-details/fda-approves-syfovretm-pegcetacoplan-injection-first-and-only). The model was trained using various resources, including unverified internet-based contents, which can exacerbate inaccuracies and lead to patient harm if prematurely deployed to clinical practice without rigorous validation.

Current ChatGPT model lacks transparency and explainability as the responses are not substantiated by references (which renders fact-checking impossible) and details on training dataset/system are undisclosed. Bias may be inadvertently introduced through biased training data, which may lead to biased responses. Other ethical concerns are data privacy and security where personal sensitive data may be inadvertently breached, (https://www.chargedretail.co.uk/2023/05/04/chatgpt-data-breach/) calling for stricter regulatory oversight and cybersecurity. Patients or doctors may upload eye/facial photos for ChatGPT’s “medical assistance”, which may risk breaching biometric security and/or identity theft [31]. Also, ChatGPT cannot take accountability nor abide by any ethical/moral boundaries. Therefore, doctors and/or patients should only consider ChatGPT as an assistive tool (at best). Potential wide-ranging roles of ChatGPT in research have also raised questions about research work originality (and plagiarism), crediting, and intellectual properties if such technology is used [18].

ChatGPT represents a significant breakthrough in AI and LLM. However, its readiness to be deployed as an automatic decision-making tool in medical field remains to be elucidated [5, 6, 27]. With further model refinement, ChatGPT may reach sufficient maturity, robustness and transparency to be embraced by relevant end users, including patients, healthcare professionals, healthcare systems, researchers, educationalists, regulatory bodies, and governmental stakeholders. When the time comes, this may well herald the dawn of a new era in medicine.