Abstract
To describe a database of longitudinally graded telemedicine retinal images to be used as a comparator for future studies assessing grader recall bias and ability to detect typical progression (e.g. International Classification of Retinopathy of Prematurity (ICROP) stages) as well as incremental changes in retinopathy of prematurity (ROP). Cohort comprised of retinal images from 84 eyes of 42 patients who were sequentially screened for ROP over 6 consecutive weeks in a telemedicine program and then followed to vascular maturation or treatment, and then disease stabilization. De-identified retinal images across the 6 weekly exams (2520 total images) were graded by an ROP expert based on whether ROP had improved, worsened, or stayed the same compared to the prior week’s images, corresponding to an overall clinical “gestalt” score. Subsequently, we examined which parameters might have influenced the examiner’s ability to detect longitudinal change; images were graded by the same ROP expert by image view (central, inferior, nasal, superior, temporal) and by retinal components (vascular tortuosity, vascular dilation, stage, hemorrhage, vessel growth), again determining if each particular retinal component or ROP in each image view had improved, worsened, or stayed the same compared to the prior week’s images. Agreement between gestalt scores and view, component, and component by view scores was assessed using percent agreement, absolute agreement, and Cohen’s weighted kappa statistic to determine if any of the hypothesized image features correlated with the ability to predict ROP disease trajectory in patients. The central view showed substantial agreement with gestalt scores (κ = 0.63), with moderate agreement in the remaining views. Of retinal components, vascular tortuosity showed the most overall agreement with gestalt (κ = 0.42–0.61), with only slight to fair agreement for all other components. This is a well-defined ROP database graded by one expert in a real-world setting in a masked fashion that correlated with the actual (remote in time) exams and known outcomes. This provides a foundation for subsequent study of telemedicine’s ability to longitudinally assess ROP disease trajectory, as well as for potential artificial intelligence approaches to retinal image grading, in order to expand patient access to timely, accurate ROP screening.
Similar content being viewed by others
Introduction
Retinopathy of prematurity (ROP) is a retinal vascular disease that affects premature and low-birth-weight infants. The underlying pathophysiology of the disease involves abnormal angiogenesis of the retina that can lead to irreversible vision loss if not detected and properly treated in time1,2,3,4. In the United States, the incidence of ROP from 1997 to 2005 was 0.17%, with an incidence of 15.58% for premature infants with an initial hospital stay greater than 28 days5. Additionally, it is estimated that ROP occurs in 68% of infants with a birth weight of less than 1251 g6. If accurate and timely screening for ROP followed by appropriate management does not occur, permanent visual impairment can ensue. Despite the existence of effective treatment methods, ROP remains a leading cause of childhood blindness worldwide2,4,7,8. As advances in medicine lead to continued improvement in neonatal survival, the population at risk for ROP increases, raising the need for effective screening methods capable of meeting expanding screening demands8,9.
Although the current gold standard for diagnosis of ROP is bedside binocular indirect ophthalmoscopy, the increasing number of at-risk infants requiring screening and a lack of sufficient ROP experts, both nationally and around the globe, has created a significant barrier to timely screening and treatment of ROP10,11,12. Telemedicine assessment of retinal photographs is an emerging method for ROP screening that has been shown to be diagnostically accurate, safe, and reliable, and has the potential to expand access to expert ROP screening9,11,13,14,15,16,17,18,19. Telemedicine assessment of retinal images removes some of the challenges of bedside exam, such as time constraints and specialist availability, allowing for a more detailed and extended review of retinal features. Furthermore, telemedicine assessment of ROP has the key benefit of allowing for longitudinal comparison of retinal images from serial exams to provide a more accurate assessment of ROP disease trajectory and treatment needs.
Prior studies of image-based ROP analysis have used traditional methods of ROP assessment based on the International Classification of ROP (zone, stage, plus, and extent)11,19,20,21,22,23,24,25,26, using static measurements without direct assessment of disease progression over time, which is necessary given the dynamic nature of ROP and a key factor in the determination of the timing of screening and treatment intervention2,27,28. The human brain is well-trained at pattern recognition, which plays a key role in telemedicine assessment since evaluators have the ability to compare multiple images side-by-side to detect changes in disease trajectory, as opposed to relying upon memory of a single image or findings of prior exams29,30. We aim to evaluate this concept in the context of telemedicine and the longitudinal clinical assessment of ROP disease progression.
In this study, an ROP expert graded retinal images from serial weekly patient exams, assessing ROP trajectory compared to progression from the prior week’s exam in order to longitudinally evaluate ROP progression through time. We then hypothesized which image features and retinal components may have influenced the examiner’s ability to detect longitudinal change. Using a unique, rigorous image grading system, we compared individual image views and retinal components to overall clinical assessment in order to better understand how ROP is assessed and how specific image or retinal factors correlate with overall longitudinal ROP assessment. This methods study describes a grading system that allows for assessment of ROP disease changes on a granular level, which will serve as a baseline for a subsequent prospective study to evaluate telemedicine as compared to bedside indirect ophthalmoscopy with respect to pattern recognition and the ability to longitudinally assess ROP disease trajectory. This study provides an important foundation for an enhanced understanding of telemedicine capabilities and clinical applications for ROP screening and a basis for artificial intelligence.
Methods
Institutional Review Board (IRB 8752) approval was obtained from the Stanford University School of Medicine and the described research adhered to the tenets of the Declaration of Helsinki. Informed consent was obtained from the legal guardians of all participants for the capture, use, and publication of their retinal images.
Retinal photos were obtained through the Stanford University Network for Diagnosis of Retinopathy of Prematurity (SUNDROP) database, a database of retinal images from infants at eleven participating sites who were enrolled in the SUNDROP initiative from December 1st, 2005 to November 30th, 2019. Color fundus photographs were taken by trained nursing staff using RetCam I/II/III (Natus Medical Incorporated, Pleasanton, CA) with a 130° lens. All patients’ eyes were dilated using a combination of 0.2% cyclopentolate/1% phenylephrine or 2.5% phenylephrine/1% tropicamide, administered twice 5 min apart, with patients dilated 30–60 min prior to the imaging session. Per imaging protocol, five images were taken with different orientations respective to the optic disc: central, inferior, nasal, superior, and temporal (Fig. 1).
Retinal images from 84 eyes of 42 patients were selected from the SUNDROP database for analysis in this study. Sample size was determined by calculations to assure adequate power for a subsequent prospective analysis of bedside binocular ophthalmoscopy versus telemedicine for longitudinal ROP assessment using these same images and grading system. All patients were sequentially screened for six consecutive weeks in the SUNDROP telemedicine program and then followed all the way to vascular maturation or treatment and then disease stabilization. Our cohort included a heterogeneous mix of patients who required treatment, patients who developed disease that then spontaneously regressed, and patients who never developed disease. Patients who received treatment for ROP and had at least 6 weeks of ROP screening images prior to their retinopathy treatment were initially selected (n = 13 patients), with the remaining 29 patients randomly selected. All patients selected had to have retinal images of each eye from at least six weekly ROP examinations.
Image grading
A Stanford University ROP expert (D.M.M.) evaluated all images in this study. All images were de-identified with the expert evaluator blinded to infant demographic information, such as gestational age and birth weight, and treatment information. Although all patients had known anatomic outcomes in this retrospective study, all anatomic and clinical outcomes were masked to the examiner. Images were evaluated for ROP disease progression in a given week compared to the immediate prior week’s images, to assess whether ROP was getting better, worse, or staying the same. Since each patient had six weeks of exam images, this corresponded to five sets of evaluations of ROP progression (Week 1 to Week 2, Week 2 to Week 3, Week 3 to Week 4, Week 4 to Week 5, Week 5 to Week 6). In total, 2100 images were graded, plus an additional 420 images serving as the baseline Week 1 images for each eye of each patient.
Gestalt score
For a given week, each set of retinal images was graded based on ROP progression compared to the prior week. An overall “gestalt” score was given each week, per eye, based on overall assessment of ROP disease trajectory, with a score of + 1 given if ROP appeared to be worsening, − 1 if ROP had improved, or 0 if ROP stayed the same compared to the immediate prior week. The gestalt score was repeated four times by the same grader for each eye of each patient at the six weekly visits, with identical results with all four gradings.
View score
After gestalt scoring was complete for all eyes across all weeks, the examiner assessed the patient cohort again, scoring each eye with respect to a set of image and retinal features hypothesized to have potentially impacted the examiner’s initial gestalt assessment of ROP disease trajectory. With respect to image-based features, a “view score” was used wherein each eye was graded for ROP progression in each of the 5 image views relative to the optic nerve (central, inferior, nasal, superior, temporal) using the same grading scale of + 1 if ROP was worse, − 1 if ROP was better, and 0 if ROP stayed the same compared to the prior week.
Component score and component-by-view score
Images were also graded based on the week-to-week progression of individual retinal components. The components evaluated were the following: vascular tortuosity (VT), vascular dilation (VD), stage (ST), hemorrhage/other abnormality (H), and growth of vessels (G), the latter defined as immature retinal vascularization growing in a centrifugal pattern emanating from the optic nerve.
These five retinal components were graded using the same − 1 to + 1 scale, where − 1 meant that the specific component had improved, 0 meant that the component had stayed the same, and + 1 meant that the component had worsened compared to the prior week. These five retinal components were graded as an overall "component score" (overall VT, VD, ST, H, and G), as well as within each of the five individual retinal views ("component-by-view score"): central, inferior, nasal, superior, and temporal vascular tortuosity (CVT, IVT, NVT, SVT, TVT); central, inferior, nasal, superior, and temporal vascular dilation (CVD, IVD, NVD, SVD, TVD); central, inferior, nasal, superior, and temporal stage (CST, IST, NST, SST, TST); central, inferior, nasal, superior, and temporal hemorrhage (CH, IH, NH, SH, TH); and central, inferior, nasal, superior, and temporal growth of vessels (CG, IG, NG, SG, TG).
Statistical analysis
All statistical analysis was conducted using R version 3.6.3 (R Foundation for Statistical Computing, Vienna, Austria). Analysis of grading variable agreement was modeled after prior studies of image-based ROP grading19,21,27,31,32. Image scores were analyzed by “individual” week-to-week assessments of ROP progression (scale of − 1 to + 1) and as “longitudinal” total scores across all 5 grading sessions, using the sum of each week-to-week assessment (scale − 5 to + 5) to assess overall disease trajectory. The primary outcome was agreement between gestalt scores and view scores, gestalt scores and overall component scores, and gestalt scores and component by view scores. Agreement was assessed by percent agreement (calculated as the number of times in which the two variable scores being compared were equal divided by the total number of image scores), absolute agreement, and Cohen’s weighted kappa statistic (used as a conservative estimate since it factors in the agreement that occurred by chance). As a secondary outcome measure, the Mann Whitney U test (Wilcoxon rank sum test) was used to compare average longitudinal gestalt scores in patients who eventually received treatment (treatment-warranted-ROP) and those who did not receive treatment33,34,35. Cohen’s weighted kappa statistic was interpreted using a commonly accepted scale, where 0.00–0.20 = slight, 0.21–0.40 = fair, 0.41–0.60 = moderate, 0.61–0.80 = substantial, and 0.81–1.00 = almost perfect agreement36,37. P values less than 0.05 were considered statistically significant.
Results
The median weekly gestalt score for all images was + 1 (range − 1 to + 1), with a median longitudinal gestalt score, corresponding to the sum across all weeks for each eye, of + 4 (range − 1 to + 5).
Agreement of gestalt and view scores
When assessed as individual week-to-week comparison scores, the percent agreement between scores of each view and gestalt scores ranged from 73.2 to 83.2%, absolute agreement ranged from 0.40 to 0.60, and Cohen’s weighted kappa ranged from 0.45 to 0.63, corresponding to moderate to substantial agreement (Table 1A). For all three measures of agreement, the order of views remained unchanged, with the central view showing the most agreement with gestalt scores (κ = 0.63), followed by the nasal view (κ = 0.55), temporal view (κ = 0.50), inferior view (κ = 0.46), and lastly the superior view (κ = 0.45).
When assessed longitudinally as the sum of weekly scores for each eye, the percent agreement ranged from 39.3 to 83.2%, absolute agreement ranged from 0.63 to 0.78, and Cohen’s weighted kappa ranged from 0.49 to 0.63 (Table 1B). The order of views in terms of most agreement to least agreement remained relatively unchanged, with the central view showing substantial agreement and the remaining four views showing moderate agreement. Furthermore, the central view had a much higher percent agreement (83.2%) with the longitudinal gestalt score than the remaining four views (39.3–45.2%).
Agreement of gestalt and overall component scores
The five retinal components compared to the week-to-week gestalt scores demonstrated a wide range of percent agreement (26.0–77.6%), absolute agreement (0.08–0.38), and Cohen’s weighted kappa (0.008–0.47) (Table 2A). Four of the five components (vascular dilation, stage, hemorrhage, and growth of vessels) showed only slight agreement with the gestalt score (κ 0.008–0.16); however, vascular tortuosity showed moderate agreement (κ = 0.47). Vascular tortuosity and vascular dilation had the highest percent agreement, 77.6% and 62.6% respectively.
Similar results were shown when longitudinal scores were compared, with vascular tortuosity demonstrating moderate agreement with gestalt scores (κ = 0.51) and all other components demonstrating only slight agreement (κ = 0.008–0.16) (Table 2B).
Agreement of gestalt and component-by-view scores
When retinal components were assessed within each individual image view, Cohen’s weighted kappa showed only slight to fair agreement within all five views for vascular dilation (κ = 0.12–0.21), stage (κ = 0.003–0.32), hemorrhage (κ = − 0.004 to 0.02), and growth of vessels (κ = 0.006–0.02) (Table 3A). Vascular tortuosity demonstrated moderate agreement with gestalt scores within all five views, with the central view showing the highest agreement (κ = 0.57), followed by the temporal view (κ = 0.48), nasal view (κ = 0.44), inferior view (κ = 0.43), and lastly the superior view (κ = 0.42). Percent agreement was highest in all five views for vascular tortuosity (74.9–81.6%), followed by all five views of vascular dilation (59.7–65.1%).
Longitudinal evaluation similarly showed the most agreement for vascular tortuosity within all five views compared to the other components, with substantial agreement demonstrated for vascular tortuosity in the central view (κ = 0.61) and moderate agreement demonstrated in the other four views (κ = 0.47–0.51) (Table 3B). All other components by view demonstrated only slight to fair agreement (κ = 0.001–0.25). All components by view showed similar absolute agreement and Cohen’s weighted kappa results, with the exception of vascular dilation in the superior and temporal views, which showed relatively low kappa statistics (SVD κ = 0.13, TVD κ = 0.16), but had relatively high absolute agreement (SVD 0.66, TVD 0.66).
Comparison of treated and non-treated patients
Of the 42 patients included in the study, 13 patients (31.0%) eventually received treatment and 29 patients (69.0%) did not. Median longitudinal gestalt score among treated patients was + 5 (range + 4 to + 5), whereas median longitudinal gestalt score among non-treated patients was + 3 (range − 1 to + 5). When compared using the Mann Whitney U test (Wilcoxon rank sum test), treated patients had a higher average longitudinal gestalt score than non-treated patients (p < 0.0001).
Discussion
As a whole, our cohort showed a general trend of consistent worsening of ROP from week-to-week, as evidenced by the median gestalt scores both weekly (+ 1) and longitudinal across all weeks (+ 4). Patients who ultimately received treatment had significantly higher gestalt scores, suggesting worse ROP than those who did not receive treatment and serving as a clinical validation of our gestalt score.
Although Cohen’s weighted kappa statistics were similar for individual week-to-week assessment and longitudinal assessment, absolute agreement generally increased, especially for the view scores and for vascular tortuosity, both overall and within each image view. This suggests a potential stronger correlation between individual image features and overall clinical ROP assessment when disease trajectory is assessed over a longer period of time, with general disease trends accurately captured over time despite potential mismatches in scores for an individual week.
The central image view showed the strongest agreement with the gestalt score, with substantial agreement demonstrated both for individual week-to-week scores and longitudinal assessment. After the central view, the temporal and nasal views had the next highest agreement, which is consistent with the fact that ROP tends to develop in the horizontal retinal (nasal, central, and temporal regions), often sparing the superior and inferior regions38. Given that no individual view showed near-perfect agreement with the gestalt score, our results suggest that analysis of all five views together, rather than analysis of only a single image, provided enhanced information that contributed to the overall clinical assessment of ROP progression by our grader. This is consistent with prior studies that have shown progressively increasing sensitivity to detect referral-warranted ROP when a single image view versus three views versus five views were assessed22.
Many prior studies of image-based grading of ROP have focused on diagnosis of “plus” disease, which is determined by the presence of abnormal arterial tortuosity and venous dilation to a greater degree than a standardized image, and is often the predominant factor that determines whether treatment is warranted20,27,28,31,39,40. However, the specific degree of vascular tortuosity and dilation indicative of a diagnosis of plus disease is difficult to quantify and varies greatly among practitioners27. While vascular tortuosity had by far the strongest agreement with clinical gestalt of any of the individual retinal components assessed in this study, with moderate agreement shown both overall and when analyzed within each view, vascular dilation only showed slight to fair agreement by Cohen’s weighted kappa and absolute agreement. Additionally, stage, which assesses retina and blood vessel changes at the border (“demarcation line”) of the vascularized and avascular retina, is often included as one of the key assessments of ROP11,19,20,21,26. However, our results showed only slight to fair agreement of stage with the gestalt scores, suggesting that other factors had a stronger influence on the overall clinical assessment of ROP progression by the grader in this study.
Limitations of this study include variations in image quality, typically due to varying degrees of patient cooperation as is often the case with assessment of actual patients. Poorer image quality in a specific view could be compensated for in the gestalt score by better quality images in the remaining views but could have disproportionately affected the analysis of image view agreement with gestalt scores. Additionally, by simply assessing ROP disease trajectory as better, worse, or the same compared to the prior week, we were unable to evaluate the scale of change, since a score of + 1 could mean a slight worsening from the week prior or a significant worsening. However, by keeping our grading scale more general, we were able to highlight overall trends in disease progression and reduce the variation that could stem from an attempt to quantify the amount of change observed. Lastly, our study used a single ROP expert to grade the images, and therefore the score agreement results may not be applicable to other graders, as prior studies have shown the potential for systematic biases that could lead an individual grader to over or undercall ROP disease or to place a greater respective focus on particular retinal features27.
In conclusion, this study describes a rigorous grading approach to image-based ROP assessment used to enhance our understanding of the individual factors that may influence overall assessment of ROP progression. We found that the central image view showed the most agreement with the gestalt score, followed by the nasal and temporal views. Vascular tortuosity showed by far the strongest agreement with gestalt, with the other components of vascular dilation, stage, hemorrhage, and vessel growth showing only slight to fair agreement. Ultimately, no single component or view had near-perfect agreement with the gestalt score, suggesting that an amalgamation of factors likely underlied the expert grader's gestalt assessment of overall ROP progression. These data provide the foundation for subsequent evaluation of telemedicine compared to bedside indirect ophthalmoscopy in the longitudinal assessment of ROP disease trajectory and can serve as a basis for an artificial intelligence approach to grading retinal images in order to expand patient access to timely and accurate ROP screening.
References
Flynn, J. T. Retinopathy of prematurity. Pediatr. Clin. N. Am. 34, 1487–1516 (1987).
Blencowe, H., Lawn, J., Vazquez, T., Fielder, A. & Gilbert, C. Preterm-associated visual impairment and estimates of retinopathy of prematurity at regional and global levels for 2010. Pediatr. Res. 74, 35–49 (2013).
Connor, A. O. & Fielder, A. R. Long term ophthalmic sequelae of prematurity. Early Hum. Dev. 84, 101–106 (2008).
Hartnett, M. & Penn, J. Mechanisms and management of retinopathy of prematurity. N. Engl. J. Med. 367, 2515–2526 (2012).
Lad, E., Hernandez-Boussard, T., Morton, J. & Moshfeghi, D. Incidence of retinopathy of prematurity in the United States: 1997 through 2005. Am. J. Ophthalmol. 148, 451–458 (2009).
Good, W. et al. The incidence and course of retinopathy of prematurity: Findings from the early treatment for retinopathy of prematurity study. Pediatrics 116, 15–23 (2005).
Steinkuller, P. et al. Childhood blindness. J. AAPOS 3, 26–32 (1999).
Valentine, P. H., Jackson, J. C., Kalina, R. E. & Woodrum, D. E. Increased survival of low birth weight infants: Impact on the incidence of retinopathy of prematurity. Pediatrics 84, 442–445 (1989).
Murakami, Y. et al. Stanford University Network for Diagnosis of Retinopathy of Prematurity (SUNDROP): 12-month experience with telemedicine screening. Br. J. Ophthalmol. 92, 1456–1460 (2008).
Coyner, A. S. et al. Automated fundus image quality assessment in retinopathy of prematurity using deep convolutional neural networks. Ophthalmol. Retin. 3, 444–450 (2019).
Chiang, M. F. et al. Telemedical retinopathy of prematurity diagnosis. Arch. Ophthalmol. 125, 1531–1538 (2007).
Richter, G. M., Williams, S. L., Starren, J., Flynn, J. T. & Chiang, M. F. Telemedicine for retinopathy of prematurity diagnosis: Evaluation and challenges. Surv. Ophthalmol. 54, 671–685 (2009).
Silva, R. et al. Stanford University network for diagnosis of retinopathy of prematurity (SUNDROP): 18-month experience with telemedicine screening. Graefes Arch. Clin. Exp. Ophthalmol. 247, 129–136 (2009).
Murakami, Y. et al. Stanford University network for diagnosis of retinopathy of prematurity (SUNDROP): 24-month experience with telemedicine screening. Acta Ophthalmol. 88, 317–322 (2010).
Silva, R., Murakami, Y., Lad, E. & Moshfeghi, D. Stanford University network for diagnosis of retinopathy of prematurity (SUNDROP): 36-month experience with telemedicine screening. Ophthalmic Surg. Lasers Imaging Retin. 42, 12–19 (2011).
Wang, S. et al. SUNDROP: Six years of screening for retinopathy of prematurity with telemedicine. Can. J. Ophthalmol. 50, 101–106 (2015).
Fijalkowski, N. et al. Stanford University Network for Diagnosis of Retinopathy of Prematurity (SUNDROP): Five years of screening with telemedicine. Ophthalmic Surg. Lasers Imaging Retin. 45, 106–113 (2014).
Fijalkowski, N. et al. Stanford University Network for Diagnosis of Retinopathy of Prematurity (SUNDROP): Four years of screening with telemedicine. Curr. Eye Res. 38, 283–291 (2013).
Biten, H. et al. Diagnostic accuracy of ophthalmoscopy vs telemedicine in examinations for retinopathy of prematurity. JAMA Ophthalmol. 136, 498–504 (2018).
International Committee for the Classification of Retinopathy of Prematurity. The international classification of retinopathy of prematurity revisited. Arch. Ophthalmol 123, 991–999 (2005).
Daniel, E. et al. Validated system for centralized grading of retinopathy of prematurity: Telemedicine approaches to evaluating acute-phase retinopathy of prematurity (e-ROP) study. JAMA Ophthalmol 133, 675–682 (2015).
Rogers, D. L. et al. Comparison of strategies for grading retinal images of premature infants for referral warranted retinopathy of prematurity. J. AAPOS 21, 141–145 (2017).
Prakalapakorn, S. G., Wallace, D. K., Dolland, R. S. & Freedman, S. F. Evaluation of the accuracy of grading indirect ophthalmoscopy video images for retinopathy of prematurity screening. J. Pediatr. Ophthalmol. Strabismus 52, 85–92 (2015).
Quinn, G. E. et al. Analysis of discrepancy between diagnostic clinical examination findings and corresponding evaluation of digital images in the telemedicine approaches to evaluating acute-phase retinopathy of prematurity study. JAMA Ophthalmol. 134, 1263–1270 (2016).
Williams, S. L. et al. Telemedical diagnosis of retinopathy of prematurity: Accuracy of expert versus non-expert graders. Br. J. Ophthalmol. 94, 351–356 (2010).
Molinari, A., Weaver, D. & Jalali, S. Classifying retinopathy of prematurity. Community Eye Heal. 30, 55–56 (2017).
Kalpathy-Cramer, J. et al. Plus disease in retinopathy of prematurity: Improving diagnosis by ranking disease severity and using quantitative image analysis. Opthalmology 123, 2345–2351 (2016).
Brown, J. M. et al. Automated diagnosis of plus disease in retinopathy of prematurity using deep convolutional neural networks. JAMA 97239, 803–810 (2020).
Mattson, M. Superior pattern processing is the essence of the evolved human brain. Front. Neurosci. 8, 265 (2014).
Unzicker, A., Jütner, M. & Rentschler, I. Similarity-based models of human visual recognition. Vision Res. 38, 2289–2305 (1998).
Kim, S. J. et al. Accuracy and reliability of eye-based vs quadrant-based diagnosis of plus disease in retinopathy of prematurity. JAMA Ophthalmol. 136, 648–655 (2018).
Ludwig, C. A., Greven, M. A. & Moshfeghi, D. M. Predictors of treatment-warranted retinopathy of prematurity in the SUNDROP cohort: Influence of photographic features. Graefes Arch. Clin. Exp. Ophthalmol. 255, 1935–1946 (2017).
Leon, A. C. 3.12 descriptive and inferential statistics. In Comprehensive Clinical Psychology 243–285 (1998).
Hoffman, J. I. E. Analysis of variance. I. One-way. In Basic Biostatistics for Medical and Biomedical Practitioners 391–417 (2019).
McDonald, J. Kruskal–Wallis test. In Handbook of Biological Statistics 157–164 (2014).
Landis, J. R. & Koch, G. G. The measurement of observer agreement for categorical data. Biometrics 33, 159–174 (1977).
Hripcsak, G. & Heitjan, D. Measuring agreement in medical informatics reliability studies. J. Biomed. Inf. 35, 99–110 (2002).
Fielder, A. R., Shaw, D. E., Robinson, J. & Ng, Y. K. Natural history of retinopathy of prematurity: A prospective study. Eye 6, 233–242 (1992).
Prakalapakorn, S. G. et al. Real-world simulation of an alternative retinopathy of prematurity screening system in Thailand: A pilot study. J. Pediatr. Ophthalmol. Strabismus 55, 245–253 (2018).
Raufi, N. N., Morris, C. K., Freedman, S. F., Wallace, D. K. & Prakalapakorn, S. G. Facilitated versus self-guided training of non-ophthalmologists for grading pre-plus and plus disease using fundus images for retinopathy of prematurity screening. J. Pediatr. Ophthalmol. Strabismus 53, 179–185 (2016).
Funding
This work was supported by the Heed Fellowship awarded through the Society of Heed Fellows to Cassie A. Ludwig, MD, MS.
Author information
Authors and Affiliations
Contributions
T.R.R. designed the study, conducted data collection/analysis, and wrote the main manuscript text. M.H.J. conducted data collection and contributed to the writing/editing of the manuscript. D.V. contributed to data analysis and the writing/editing of the manuscript. C.A.L. contributed to data analysis and the writing/editing of the manuscript. A.A. contributed to the writing/editing of the manuscript. M.V.P. contributed to the writing/editing of the manuscript. N.F.C. contributed to the writing/editing of the manuscript. J.K. contributed to study design, data analysis, and the writing/editing of the manuscript. D.M.M. designed the study, conducted data collection/analysis, and contributed to the writing/editing of the manuscript. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
T.R. Rosenblatt, None; M.H. Ji, None; D. Vail, None; C.A. Ludwig, None; A. Al-Moujahed, None; M. Veerappan Pasricha, None; N.F. Callaway, None; J. Kumm, Pr3vent INC (F,CEO); D.M. Moshfeghi, 1-800Contacts (BOD,E), Akceso Advisors AG (C), Akebia (SAB), Alcon (C), Aldeyra Therapeutics (PI), Alexion (C), Allegro (SAB), Apellis (PI), Bayer Pharma AG (SC,C), CMEOutfitters.com (C), Congruence Medical Solutions (C), dSentz, Inc. (F,E,BOD), Grand Legend Technology, LTD (E), Iconic Therapeutics (SC), Irenix (SAB), Linc (F,E,BOD), M3 Global Research (C), Northwell Health (C), Novartis Pharmaceuticals (DMC, C), Ocular Surgery News (C), Pr3vent INC (F,E,BOD), Praxis UNS, Inc. (C), Prime Medical Education (C), Promisight, Inc (F, E, BOD), Pykus (E,SAB), Regeneron (SC,PI), Retina Technologies LLC (SAB,C,E), Retina Today/Pentavision (C), Shapiro Law Group (C), SLACK, Inc (C), Versl, Inc. (F E), Vindico (C), Visunex (E,SAB).
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rosenblatt, T.R., Ji, M.H., Vail, D. et al. Key factors in a rigorous longitudinal image-based assessment of retinopathy of prematurity. Sci Rep 11, 5369 (2021). https://doi.org/10.1038/s41598-021-84723-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-021-84723-7
This article is cited by
-
Telemedicine retinopathy of prematurity severity score (TeleROP-SS) versus modified activity score (mROP-ActS) retrospective comparison in SUNDROP cohort
Scientific Reports (2023)
-
Effective field of view of wide-field fundus photography in the Stanford University Network for Diagnosis of Retinopathy of Prematurity (SUNDROP)
Scientific Reports (2022)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.