To assess the role of artificial intelligence (AI)-based automated software for detection of diabetic retinopathy (DR) and sight-threatening DR (STDR) by fundus photography taken using a smartphone-based device and validate it against ophthalmologist’s grading.
Three hundred and one patients with type 2 diabetes underwent retinal photography with Remidio ‘Fundus on phone’ (FOP), a smartphone-based device, at a tertiary care diabetes centre in India. Grading of DR was performed by the ophthalmologists using International Clinical DR (ICDR) classification scale. STDR was defined by the presence of severe non-proliferative DR, proliferative DR or diabetic macular oedema (DME). The retinal photographs were graded using a validated AI DR screening software (EyeArtTM) designed to identify DR, referable DR (moderate non-proliferative DR or worse and/or DME) or STDR. The sensitivity and specificity of automated grading were assessed and validated against the ophthalmologists’ grading.
Retinal images of 296 patients were graded. DR was detected by the ophthalmologists in 191 (64.5%) and by the AI software in 203 (68.6%) patients while STDR was detected in 112 (37.8%) and 146 (49.3%) patients, respectively. The AI software showed 95.8% (95% CI 92.9–98.7) sensitivity and 80.2% (95% CI 72.6–87.8) specificity for detecting any DR and 99.1% (95% CI 95.1–99.9) sensitivity and 80.4% (95% CI 73.9–85.9) specificity in detecting STDR with a kappa agreement of k = 0.78 (p < 0.001) and k = 0.75 (p < 0.001), respectively.
Automated AI analysis of FOP smartphone retinal imaging has very high sensitivity for detecting DR and STDR and thus can be an initial tool for mass retinal screening in people with diabetes.
All individuals with diabetes, irrespective of the type of diabetes, require regular and repetitive annual retinal screening for early detection and timely treatment of diabetic retinopathy (DR), particularly sight-threatening diabetic retinopathy (STDR) [1, 2]. Screening for retinopathy is conventionally done through fundus examination by ophthalmologists or retinal colour photography using conventional mydriatic or non-mydriatic fundus cameras by optometrists or trained eye technicians . Smartphone-based retinal imaging has emerged as one of the recent cost-effective ways of screening for retinopathy in the community [3,4,5]. However, till date irrespective of the type of fundus camera used, the retinal images had to be graded for the presence and severity of DR by ophthalmologists (retinal specialists) or trained graders [3,4,5].
Given the alarming increase in the number of people with diabetes and shortage of trained retinal specialists and graders of retinal photographs, an automated approach involving a computer-based analysis of the fundus images would reduce the burden of the health systems in screening for DR [6,7,8]. There is hence an increasing interest in the recent past in the development of automated analysis software using computer machine learning/artificial intelligence (AI)/deep neuronal learning for analysis of retinal images in people with diabetes [7, 9]. AI is simulation of human intelligence by a software/machine. It is a specialised field which is based on teaching the machine to recognise specific patterns. It has been used for different kinds of technical tasks including accurate classification of high-resolution images. AI for detection and classification of DR happens by providing thousands of retinal images of varying grades of DR to the system for learning. The machine after being exposed to a lot of annotated retinal images learns to grade DR by itself. Some studies done in the recent past have shown that AI could be used to grade retinal images taken using the conventional fundus cameras and determine which patients with DR need referral to the ophthalmologist [7,8,9]. Many of the AI algorithms have been trained and developed using retinal images from high-quality conventional fundus cameras.
To our knowledge, assessment of the use of AI along with smartphone-based fundus photography has not been done so far. The aim of this study was to evaluate the usefulness of an automated AI-based interpretation of smartphone-based fundus photography system for screening at a physician clinic (without the algorithm being trained specifically on retinal images from smartphone fundus photography). This paper studies the accuracy of the automated DR software in DR detection and screening for STDR using a previously validated smartphone-based retinal imaging system.
Three hundred and one patients with type 2 diabetes, aged 18 years and above, undergoing treatment for diabetes at a tertiary care diabetes hospital in Chennai (formerly Madras) in southern India, with varying duration of diabetes underwent retinal colour photography (fundus photography) at the eye department as a part of regular retinal screening for DR. A pilot study was conducted with retinal images of 50 patients to assess the sensitivity and specificity of automated DR detection using the EyeArtTM software. The sample size for this study was calculated based on the results of the pilot phase. A written informed consent was obtained from all participants and the study was approved by the Ethical Committee of the Madras Diabetes Research Foundation.
After preliminary eye examination, the pupils were dilated with tropicamide eye drops after ruling out any history of allergy to the dilating eye drops. The retinal photographs were acquired using Remidio Fundus on Phone (FOP), a smartphone-based imaging device (Remidio Innovative Solutions Pvt. Ltd, Bangalore, India). It is a portable fundus camera that consists of an annular illumination design that eliminates corneal reflection, which mates with any commercially available smartphone to acquire retinal photographs . The FOP provides a 45° field of view for fundus imaging and can be used in a handheld mode or be fit onto any standard slitlamp as shown in Fig. 1. Four fields were captured in each eye on the FOP camera: macula centred, disc centred, superior-temporal and inferior-temporal quadrants of the retina. Retinal photographs were coded using an identification number and assessed in a masked manner for the presence and severity of DR in order to minimise any possible bias. The photographs were graded by two ophthalmologists (retina specialists) who were masked to the patient’s identity and clinical diagnosis. The kappa agreement between the grading of the two ophthalmologists was k = 0.89. Any disagreement in the retinopathy grading between the two graders was adjudicated by a third retina specialist whose DR grading was taken as final. The grading of retinopathy was done based on the International Clinical Diabetic Retinopathy (ICDR) severity scale . The ICDR severity scales provides a classification of five stages of DR as follows: (1) no apparent retinopathy—no abnormalities; (2) mild non-proliferative DR (NPDR)—microaneurysms only; (3) moderate NPDR—more than just microaneurysms, but less than severe NPDR; (4) severe NPDR—one or more of the following: (i) more than 20 intra-retinal haemorrhages in each of four quadrants, (ii) definite venous beading in two or more quadrants, (iii) prominent intra-retinal microvascular abnormality in one or more quadrants; (5) proliferative DR (PDR)—retinal neovascularisation with or without vitreous/preretinal haemorrhage . Sight-threatening DR (STDR) was defined by the presence of severe NPDR, PDR and/or diabetic macular oedema (DME)/clinically significant macular oedema (CSME) . Photographs were graded and assigned a retinopathy level and the final diagnosis for each patient was determined from the level of DR of the worse eye using ICDR severity scale. Figure 1 shows some Remidio FOP retinal images of varying severity of DR.
The retinal images after masking the identity of the patient and the diagnosis were shared with EyeNuk Inc. (Los Angeles, CA, USA) for the automated analysis with the EyeArtTM software (EyeNuk Inc., Los Angeles, CA). EyeArtTM is a computerised, cloud-based software that can automatically analyse retinal images and provide DR severity and screening recommendation by automatically detecting the presence, size, position and number of retinal lesions. In addition to DR severity level on the ICDR severity scale, presence/absence of surrogate markers for CSME which is defined by the presence of hard exudates within one disc diameter of the centre of the macula are also reported by the algorithm. EyeArt’s core analysis engine contains DR analysis algorithms including those for image enhancement, interest region identification, descriptor computation, in conjunction with an ensemble of deep artificial neural networks for DR classification, and detection of CSME surrogate markers . The EyeArt AI algorithm has been clinically AI trained and validated using retinal images of 78,685 patients taken using conventional desktop mydriatic fundus cameras-obtained from the EyePACS database . Referable diabetic eye disease (RDR) by EyeArt is defined as the presence of (i) moderate NPDR or higher and/or (ii) CSME [7, 9]. A “no refer” recommendation was provided for patients with no apparent signs of DR or signs of mild DR without CSME .
A total of 2408 Remidio FOP retinal images of 301 patients were shared through the cloud with EyeNuk for this study. EyeArtTM software (version v2.1.0) which was used in this study provided a patient wise diagnosis of DR.
All statistical analyses were performed using SAS statistical package version 9.2. The sensitivity and specificity of EyeArt algorithm for detecting DR and diagnosing referable/sight-threatening DR were calculated by generating 2 × 2 tables taking the ophthalmologist grading as the reference standard. Positive predictive value (PPV) was defined as the probability of the presence of disease given a positive test result and negative predictive value (NPV) was defined as the probability of the absence of disease given a negative test result. 95% confidence intervals (CIs) were used for sensitivity, specificity, PPV and NPV. The degree of agreement between automated analysis and manual (ophthalmologist) grading was quantified and assessed using the kappa (ĸ) statistics. For all statistical tests, p-value <0.05 was considered significant.
DR grading was done by the ophthalmologists for Remidio FOP retinal images of 301 (602 eyes) type 2 diabetes patients and a total of 2408 FOP retinal images were shared with EyeNuk Inc. The EyeArtTM AI algorithm provided diagnosis for retinal images of 296 patients. Retinal photographs of five patients were deemed to be of inadequate quality for automated analysis when the media was unclear due to cataract or asteroid hyalosis or vitreous haemorrhage. Thus the statistical analysis and the comparisons were done for 296 patients. DR was detected by the ophthalmologist in 191 (64.5%) patients by grading the smartphone-based retinal images while the EyeArt software detected DR in 203 (68.3%) of the 296 patients. The varying grades of DR in the two modes of retinopathy grading based on ICDR are depicted in Fig. 2. Based on manual grading by the ophthalmologist 142 (48%) patients had RDR (moderate NPDR and above), while based on the EyeArt grading, 189 (63.9%) patients had RDR. STDR was detected in 112 (37.8%) by the ophthalmologist and in 146 (49.3%) patients by EyeArt algorithm.
Figure 3 shows the Venn diagram of the STDR identified by the ophthalmologist vs. the software and the overlap and matched diagnosis of STDR observed in 110 patients.
The sensitivity and specificity for detecting any DR, DME, PDR, STDR, RDR, by EyeArt software against the ophthalmologist grading as the reference standard and the degree of agreement between two is shown in Table 1. The automated software correctly identified 95.8% of patients with retinopathy and 80.2% of patients without retinopathy. EyeArt showed 95.8% (95% CI 92.9–98.7) sensitivity and 80.2% (95% CI 72.6–87.8) specificity for detecting DR and 99.1% (95% CI 95.1–99.9) sensitivity and 80.4% (95% CI 73.9–85.9) specificity in diagnosing STDR. The degree of agreement between automated and ophthalmologist grading for any DR was 0.78 (p < 0.001) and for STDR was 0.75 (p < 0.001) using the kappa (ĸ) statistics. The sensitivity for referable DR (RDR) was 99.3% and the specificity was 68.8% and the ĸ agreement between the human vs. machine grading for RDR was 0.67 (p < 0.001). As the sensitivity values were high, the PPV and the NPV were also calculated as shown in Table 1. The PPV for detection of DR, STDR and RDR was 89.7% (95% CI 85.5–93.8), 75.3% (95% CI 68.4–82.3) and 74.6% (95% CI 68.4–80.8), respectively.
This study has assessed the role of automated AI algorithm in detection of DR using a low-cost smartphone-based imaging device. To our knowledge, this is the first time that an automated AI DR screening software has been tested for its accuracy for smartphone-based retinal imaging.
Regular retinal screening for all people with diabetes is still an unmet need in most countries especially in poor developing countries. Retinal photography with grading and interpretation by ophthalmologists/retina specialists and trained graders is widely accepted for screening for DR [4, 12, 13]. However, availability of retina specialists/trained retinal graders is a major limitation in most countries, including India. Even when available, there could be a time delay in graders submitting their DR grading and advice due to their busy schedule. This leads, not only to delayed interpretation, but also loss to follow-up, miscommunication, and delay in proceeding for management of STDR . Deep learning/AI for detection of DR happens by machine learning by providing thousands of retinal images of varying grades of DR to the system for learning [7, 8]. The need and efforts for a comprehensive automated method of DR screening have made good progress using image classification, pattern recognition and machine learning techniques [15, 16].
We report here, a high sensitivity for detection of DR, STDR and RDR (above 95% for DR, 99% for STDR and RDR) using the EyeArt software when used on retinal images taken with FOP. This is similar to the high sensitivity in the Google AI algorithm which showed a high sensitivity and specificity for RDR when used on conventional retinal photography (sensitivity of 97.5% and specificity of 93.4% in the EYEPACS-1 and 96.1% sensitivity and 93.9% specificity for Messidor-2 set) . One of the main reasons for the lower specificity for RDR in our study was because of higher estimation of Moderate NPDR with the images assessed by the algorithm. Some non-DR retinal lesions like drusen, RPE atrophic patch, a retinal telangiectatic vessel at macula, RPE hypertrophy, tessellated fundus and retinal vein occlusion were the causes of false positives. However, these non-DR retinal lesions also need advice from the ophthalmologist/retina specialist and hence they were not truly false positives. The specific impact of formal training on retinal images of varying DR severity from Remidio FOP/smartphone-based fundus cameras on the algorithm could probably lead to higher specificity for RDR detection in the future.
A recent study done by EyeNuk with retinal images taken with traditional desktop fundus cameras showed that EyeArt’s sensitivity for DR screening sensitivity was 91.7% (95% CI: 91.3–92.1%) and specificity was 91.5% (95% CI: 91.2–91.7%) . They also showed that the sensitivity for detecting treatable diabetic eye disease (STDR) was 98.5%, i.e. of the 470 eyes with treatable DED as per reference standard, 465 eyes were correctly provided “refer” recommendations by the algorithm . This is similar to the sensitivity for detecting DR and STDR in our study using smartphone-based retinal images.
In a very recent major study publication on validation of deep learning (AI) by Ting et al.  done in Singapore with multiple retinal images taken with conventional fundus cameras from multi-ethnic cohorts of people with diabetes, their algorithm showed a high sensitivity and specificity for identifying DR and other eye diseases like age-related macular degeneration. The sensitivity and specificity for RDR was 90.5% (95% CI 87.3–93.0%) and 91.6% (95% CI 91.0–92.2%) respectively and for STDR the sensitivity was 100% (95% CI 94.1–100.0%) and the specificity was 91.1% (95% CI 90.7–91.4%) in their study . In our study done using FOP smartphone-based retinal images, we have reported a similar high sensitivity for RDR and STDR detection.
The IRIS (intelligent retinal imaging system), an automated teleretinal DR screening programme, compared non-mydriatic retinal images with a standard data set images from Early Treatment Diabetic Retinopathy Study (ETDRS) and gave suggestions for referral. Any patient with severe NPDR or more advanced disease was considered suitable for the referral . IRIS screening programme had a good sensitivity and a low false-negative rate . The sensitivity of the IRIS algorithm in detecting STDR compared with the reading centre interpretation was 66.4% (95% CI 62.8–69.9) with a false-negative rate of 2% and the specificity was 72.8% (95% CI 72.0–73.5). In our study the sensitivity for STDR was 99.1% and the specificity was 80.4%. The lower sensitivity with IRIS was possibly because the system used non-mydriatic retinal images.
It is of interest that, in this study, despite no formal machine learning/training of the algorithm on the images from Remidio FOP, the EyeArt solution was able to grade images for STDR with a high sensitivity of 99.1% and specificity of 80.4%. The results reported in this paper are in agreement with our earlier study that reported a high level of agreement when comparing the ophthalmologist grading of retinal images from the FOP and the Zeiss FF450 conventional mydriatic fundus camera for retinopathy screening . The EyeArt software is already trained on retinal images from conventional fundus cameras like the Zeiss FF450 . It is thus reassuring that the AI algorithm also works well on the FOP smartphone-based imaging system.
Use of AI to analyse retinal images is appealing as it fits in with the current trend of tele-ophthalmology and telemedicine [13, 19]. Automated DR grading softwares have potential benefits of efficiency, reproducibility and early detection of DR happening at the physician level and thus would be useful in reducing the burden to the health systems in screening of the increasing number of people with diabetes [16, 20]. Only those who have sight-threatening DR and referable DR would need to meet the ophthalmologist/ retina specialist. Urgent referral of patients with sight-threatening DR to the retina specialist for further evaluation and treatment is essential, especially since DR affects people with diabetes during their prime productive years of life .
Strengths of the study
To the best of our knowledge, this is the first study that has looked at the role and accuracy of automated AI-based DR analysis in smartphone-based retinal imaging. In this study, as the EyeArtTM software had already been validated  and the Remidio FOP imaging system also had been validated for DR screening , quality automated DR detection in this study was possible without formal training with large number of FOP retinal images for deep learning. Automated softwares using the AI like the EyeArt along with telemedicine can enable faster real-time screening of large number of people with diabetes. Smartphone retinal colour photography combined with an automated detection system can ideally result in models with potential for cost-effective routine clinical use by the primary care physicians. Admittedly, further work is needed before recommending its regular use in eye care.
Limitations of the study
The sample size is relatively small when compared to other studies that have recently assessed role of AI in DR. However, as this study has not used any images from Remidio FOP to train EyeArt, large datasets were not necessary. Although the AI algorithm could replicate manual grading by ophthalmologists, it could not overcome physical limitations, such as inability to acquire photographs in some patients due to poor mydriasis, poor image quality due to media opacities like cataract . The conclusions from this study cannot necessarily be extended to all smartphone-based imaging devices, unless they have been independently validated for performance in DR screening.
Conclusions of the study
In summary, an AI-based grading algorithm in combination with validated smartphone-based imaging of diabetic patients can be used to reliably and accurately screen patients for sight-threatening DR who could then be referred to the retina specialist for further evaluation and treatment. As patients will anyway be referred to the ophthalmologist, false-positive cases can be excluded by them and those who need treatment can be given the appropriate therapy. The automated analysis algorithm installed inside a low-cost sleek fundus camera can also be very useful for mass scale DR screening programmes particularly in remote areas of poorly developed countries where trained personnel may not be available.
What was known before
Retinal images had to be graded for the presence and severity of DR by ophthalmologists (retinal specialists) or trained graders.
Studies done in the recent past have shown that Artificial intelligence (AI) could be used to grade retinal images taken using the conventional fundus cameras and determine which patients with DR need referral to the ophthalmologist.
What this study adds
Assessment of the use of Artificial intelligence (AI) along with smartphone-based fundus photography for Diabetic retinopathy (DR) detection and classification has not been done so far.
Automated AI analysis of smartphone retinal imaging has very high sensitivity for detecting DR and sight-threatening DR and thus can be an initial tool for mass retinal screening in people with diabetes.
Fong DS, Aiello LP, Ferris FL, Klein R. Diabetic retinopathy. Diabetes Care. 2004;27:2540–3.
Namperumalswamy P, Nirmalan PK, Ramaswamy KM. Developing a screening program to detect sight threatening retinopathy in south India. Diabetes Care. 2003;26:1831–5.
Ryan ME, Rajalakshmi R, Prathiba V, Anjana RM, Ranjani H, VenkatNarayan KM, et al. Comparison Among Methods of Retinopathy Assessment (CAMRA) study: smartphone, nonmydriatic, and mydriatic photography. Ophthalmology. 2015;122:2038–43.
Rajalakshmi R, Arulmalar S, Usha M, Prathiba V, Kareemuddin KS, Anjana RM, et al. Validation of smartphone based retinal photography for diabetic retinopathy screening. PLoS ONE. 2015;10:e0138285.
Russo A, Morescalchi F, Costagliola C, Delcassi L, Semeraro F. Comparison of smartphone ophthalmoscopy with slit-lamp biomicroscopy for grading diabetic retinopathy. Am J Ophthalmol. 2015;159:360–4.e1.
Abràmoff MD, Niemeijer M, Suttorp-Schulten MS, Viergever MA, Russell SR, van Ginneken B. Evaluation of a system for automatic detection of diabetic retinopathy from color fundus photographs in a large population of patients with diabetes. Diabetes Care. 2008;31:193–8.
Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316:2402–10.
Walton OB, Garoon RB, Weng CY, Gross J, Young AK, Camero KA, et al. Evaluation of automated teleretinal screening program for diabetic retinopathy. JAMA Ophthalmol. 2016;134:204–9.
Bhaskaranand M, Cuadros J, Ramachandra C, Bhat S, Nittala MG, Sadda SR, et al. Automated diabetic retinopathy screening and monitoring using retinal fundus image analysis. J Diabetes Sci Technol. 2016;10:254–61.
Wilkinson CP, Ferris FL III, Klein RE, Lee PP, Agardh CD, Davis M, Global Diabetic Retinopathy Project Group, et al. Proposed International Clinical Diabetic Retinopathy and diabetic macular edema disease severity scales. Ophthalmology. 2003;110:1677–82.
Solanki K, Bhaskaranand M, Ramachandra C, Bhat S. Comprehensive clinical validation study of a fully-automated diabetic retinopathy screening system using color fundus images against 7-field ETDRS stereoscopic reference standard. EURETINA 2016 Abstract
Bhargava M, Cheung CY, Sabanayagam C, Kawasaki R, Harper CA, Lamoureux EL, et al. Accuracy of diabetic retinopathy screening by trained non-physician graders using non-mydriatic fundus camera. Singapore Med J. 2012;53:715–9.
Raman R, Padmaja RK, Sharma T. The sensitivity and specificity of non-mydriatic digital stereoscopic retinal imaging in detecting diabetic retinopathy. Diabetes Care. 2007;30:1.
Rachapelle S, Legood R, Alavi Y, Lindfield R, Sharma T, Kuper H, et al. The cost-utility of telemedicine to screen for diabetic retinopathy in India. Ophthalmology. 2013;120:566–73.
Abramoff M, Niemeijer M, Russell SR. Automated detection of diabetic retinopathy: barriers to translation into clinical practice. Expert Rev Med Dev. 2010;7:287–96.
Kapetanakis VV, Rudnicka AR, Liew G, Owen CG, Lee A, Louw V, et al. A study of whether automated diabetic retinopathy image assessment could replace manual grading steps in the English National Screening Programme. J Med Screen. 2015;22:112–8.
Solanki K, Bhaskaranand M, Bhat S, Ramachandra C, Cuadros J, Nittala MG, et al. Automated diabetic retinopathy screening: large-scale study on consecutive patient visits in a primary care setting. EASD 2016
Ting DSW, Cheung CY, Lim G, Tan GSW, Quang ND, Gan A, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA. 2017;318:2211–23.
Mohan V, Prathiba V, Pradeepa R. Tele-diabetology to screen for diabetes and associated complications in rural India: the Chunampet Rural Diabetes Prevention Project Model. J Diabetes Sci Technol. 2014;8:256–61.
Larsen N, Godt J, Grunkin M, Lund-Andersen H, Larsen M. Automated detection of diabetic retinopathy in a fundus photographic screening population. Invest Ophthalmol Vis Sci. 2003;44:767–71.
Rajalakshmi R, Amutha A, Ranjani H, Ali MK, Unnikrishnan R, Anjana RM, et al. Prevalence and risk factors for diabetic retinopathy in Asian Indians with young on set type1and type 2 diabetes. J Diabetes Complications. 2014;28:291–7.
Scanlon PH, Foy C, Malhotra R, Aldington SJ. The influence of age, duration of diabetes, cataract, and pupil size on image quality in digital photographic retinal screening. Diabetes Care. 2005;28:2448–53.
We acknowledge the help of all the optometrists at DMDSC for performing the digital retinal colour photography for all the patients with Remidio Fundus on phone camera and the ophthalmologists at DMDSC, Gopalapuram for the ICDR grading. We acknowledge Dr. Kaushal Solanki, Founder and CEO, EyeNuk Inc., Los Angeles, California and his team for doing the automated analysis and for providing the technical information about Eye Art software.
Conceived and designed the experiment: RR and VM. Checked the grading: RR. Analysed the data: RS. Wrote the manuscript: RR. Helped revising the manuscript for important intellectual content: RMA and VM. Read and approved the final manuscript: RR, RS, RMA and VM.
About this article
Scientific Reports (2019)
Ophthalmology and Therapy (2018)
Der Diabetologe (2018)