Introduction

Google’s Gemini AI represents a significant leap in chatbot technology, showcasing advanced capabilities and innovative features. Central to Gemini’s design is its status as a “native multimodal” model, enabling it to process and learn from various data types, including text, audio, and video. Gemini’s technical capabilities is evident in its ability to analyse complex data sets, such as charts and images, which is a substantial advancement over the earlier Bard AI models [1]. This capability is particularly relevant for applications in medicine and ophthalmology, where data often comes in visual formats like medical images/scans. By analysing these images, Gemini could potentially be a useful tool to healthcare professionals in diagnosing and treating a wide range of conditions.

Moreover, Gemini’s potential in medicine extends beyond image analysis. Its advanced language processing abilities enable it to understand and interpret medical literature, patient histories, and research data, providing valuable insights for medical professionals. In ophthalmology, Gemini could assist in diagnosing eye conditions, analysing patient-reported symptoms, and even suggesting treatment plans based on the latest research and clinical guidelines. ChatGPT has previously attempted these tasks, however did not yet perform at suitable levels to be used clinically [2,3,4,5,6,7]. Large language models such as ChatGPT can make errors in understanding the context of information, or provide outdated information, which further complicates the usage of these technologies in a clinical context [8,9,10,11].

We first decided to ask Bard to advise a patient of what to do when they complained of waking up with painful red eyes. Bard’s response was thorough and practical, providing a list of steps the patient could take, such as applying cool compress, using artificial tears, and avoiding eye rubbing, to relieve any on-going inflammation (Fig. 1A). ChatGPT similarly provided very similar, yet a longer more comprehensive list of practical guidance and steps that the patient could take to reduce their discomfort. Bard and ChatGPT’s responses were medically sound and in-line with current clinical guidelines.

Fig. 1: Output responses generated by Bard and ChatGPT.
figure 1

A Output generated by Bard from the prompt “I woke up this morning with painful red eyes, what can I do about it?”. B Output generated by ChatGPT 3.5.

Next, we asked Gemini about how often an individual should have an eye exam. Gemini AI suggested four age-based recommendations for eye exams, noting that individual needs may vary due to factors like eyeglass use, existing eye conditions, or family medical history. Similarly, ChatGPT had categories of ‘Children and teenagers’, ‘Adults’, and ‘Older adults’. Both Gemini AI and ChatGPT highlighted the importance of consulting with an eye specialist.

Next we prompted both of AI chatbots about a patient reporting “flashes of lights” in one eye, and if they should attend the emergency department. Both Bard and ChatGPT correctly recommended to attend the emergency department, particularly if this vision change occurred suddenly. Both chatbots also appropriately stated that this symptom could be a sign of a retinal tear or detachment, requiring urgent evaluation. These AI-generated outputs were both specific, and appropriate.

Finally, we prompted both AI chatbots about what a patient should do if they started seeing floaters or black dots (see Fig. 2). There are several causes of floaters ranging from relatively benign (e.g. age-related) to more serious causes (e.g. retinal detachment). Bard accurately reported a few potential reasons and suggested a formal consultation with an eye care specialist if sudden blindness developed or if the patient started experiencing changes in floater size or light flashes, which correctly addresses potential risk Bard also provided practical tips to reduce discomfort due to floaters. ChatGPT’s response was similar to Bard and also correctly explained causes of floaters and when to seek urgent medical assistance. ChatGPT, unlike Bard, also provided information on floaters treatment. In addition, ChatGPT advised seeing an eye doctor if there were several floaters, light flashes or a seeing a curtain over the vision field (Fig. 3).

Fig. 2: Output responses generated by Bard and ChatGPT.
figure 2

A Output generated by Bard from the prompt “How often should I have an eye exam?” (Left Panel). B Output generated by ChatGPT 3.5. (Right Panel).

Fig. 3: Output responses generated by Bard and ChatGPT.
figure 3

A Output generated from the prompt “I have noticed seeing some Floaters or Black Dots in my eyes, what should I do?” (Left Panel). B Generated by ChatGPT 3.5. (Right Panel).

Finally, we wanted to test the image analysis capabilities of Gemini AI against GPT-4.

Gemini AI unfortunately could not process the file despite attempting a variety of prompts. On the other hand, GPT-4 correctly identified the image of a human eye and that the picture was taken using an operating microscope. However, GPT-4 failed to correctly describe the red coloration as hyphaema (Fig. 4).

Fig. 4: Output responses generated by Bard and ChatGPT.
figure 4

A Response by Gemini AI (Top Panel). B Response by GPT-4 (Bottom Panel). Reprinted without changes from “Cheers not tears: champagne corks and eye injury, https://doi.org/10.1136/bmj.p2520.

Conclusion

Overall, the new Gemini AI model represents a notable improvement in text-based output than predecessor models. The comparative analysis between Gemini AI and ChatGPT/GPT-4 reveals distinct attributes and capabilities of these advanced AI models. Gemini AI shows promise with unique strengths in areas such as language understanding. It emerges as a strong competitor to ChatGPT, suggesting a dynamic and evolving landscape in AI language models. Both models exhibit exceptional capabilities but differ in various aspects of language processing and response generation. The analysis underlines the fact that each AI model, including ChatGPT, GPT-4, Bard, and Gemini AI, possesses unique strengths and weaknesses, making them suitable for different applications and use cases. It is important to note that further advancements are necessary prior to being the use of AI chatbots in clinical settings [12, 13].