Introduction

During the last decade, new devices for the diagnosis of glaucoma have been developed. Examples include the frequency-doubling perimeter (FDT; Carl Zeiss Meditec Inc., Dublin, CA, USA) and the nerve fibre analyser (GDx; Laser Diagnostic Technologies, San Diego, CA, USA). Many cross-sectional studies have shown a reasonable screening performance of these two devices. Whether they are suitable for the follow-up of glaucoma patients or whether they are able to predict glaucomatous visual field loss in glaucoma suspect patients is less clear; longitudinal studies with FDT and GDx are relatively rare.1, 2, 3, 4

In 2000–2001, we included a cohort of consecutive patients attending our glaucoma service, both patients with glaucomatous visual field loss and glaucoma suspect patients, in a prospective follow-up, the Groningen longitudinal glaucoma study.5 The aims of this study were to investigate (1) whether baseline FDT/GDx test results were able to predict glaucomatous visual field loss as assessed by standard automated perimetry (SAP) in glaucoma suspect patients and (2) whether FDT/GDx could be used to monitor progression in patients with glaucomatous visual field loss already present at baseline. In this study, we addressed the first aim. This issue has been addressed before for FDT and GDx separately,1, 2, 3, 4 but, to the best of our knowledge, this is the first study investigating the combined use of FDT and GDx for this purpose.

Material and methods

The data set and other details have been described elsewhere.5 In short, all patients who visited our glaucoma service between 1 July 2000 and 30 June 2001 and who provided informed consent underwent frequency-doubling perimetry (FDT; Carl Zeiss Meditec, Dublin, California, USA), a GDx measurement (Laser Diagnostic Technologies, San Diego, California, USA), conventional perimetry (see below), and tonometry (Goldmann applanation tonometry; Haag Streit AG, Bern, Switzerland) as a baseline measurement for a prospective follow-up. The study protocol was approved by the internal review board of the department of Medical Technology Assessment of the University of Groningen. Patients received written information at home at least 2 weeks before their visit. The aim of the study was explained; participation was voluntary and participation could be stopped in any stage of the study. Additional information—if requested—was provided by a technician who was not involved in routine care and who also performed the FDT and GDx measurements (everything else was part of routine care), if the patient provided informed consent. Of the 875 (predominantly white) patients included in the study, 423 patients had normal baseline fields in both eyes. Most of these patients suffered from ocular hypertension (271; defined as >20 mmHg on at least two separate visits), had a positive family history of glaucoma (104; father, mother, brother, or sister) or both (61). These patients were followed irrespective of the aspect of their optic disc. Of the remaining 109 patients, 43 were followed because of a suspected optic disc as an isolated finding (defined as a vertical cup–disc ratio >0.6). Most of the remaining 66 patients had a positive family history not strictly according to our definition (grandmother, grandfather, uncle, or aunt, or a less clear history concerning father, mother, brother, or sister). All these patients, who were followed by an ophthalmologist because discharge was deemed unwise given a presumed increased risk of glaucoma, and who had normal baseline fields in both eyes, were eligible for this study. Within the framework of this study, they were named glaucoma suspect patients.

Patients were followed for a period of four years with conventional perimetry performed yearly. In the case of suspected conversion, the test was repeated to confirm or falsify the finding. Especially in the last follow-up year, the repeat test was performed shortly; all abnormal tests had to be confirmed or falsified at the end of the follow-up. Apparently normal but formally unreliable5 tests were not repeated in our clinical setting. For this study, patients had to be tested with SAP (see below), both at baseline and during follow-up. Goldmann perimetry (Haag Streit AG, Bern, Switzerland) was applied in some patients (only Goldmann perimetry was available in our department until 1997 and some patients were still followed with this technique during (the beginning of) our study), but these patients were not included in the current analysis.

Glaucoma suspects who developed a reproducible visual field defect (see below) on SAP in at least one eye during follow-up were classified as converters. Patients with normal, reliable5 SAP test results in both eyes in the last follow-up year were classified as non-converters. All other patients were excluded.

Perimetry

We used the Humphrey Field analyser (HFA) 30-2 Sita Fast (Carl Zeiss Meditec Inc.). An abnormal field was defined as either

  1. 1

    Glaucoma hemifield test (GHT) outside normal limits,

  2. 2

    Pattern SD (P<0.05), or

  3. 3

    Three adjacent non-edge points P<0.05 in the pattern deviation probability plot of which at least one point P<0.01 and all points on the same side of the horizontal meridian (LTG-P criterion).6

For conversion, two consecutive, reliable5 tests had to be classified as abnormal7, 8 (according to any of the criteria as listed above), and any reliable test performed after these two tests had to be abnormal as well. Defects had to be in the same hemifield and at least partially overlapping. Moreover, defects had to be compatible with glaucoma and without any other explanation.

Frequency-doubling perimetry

Testing was performed with the FDT version 2.60, using the C-20 full threshold mode. An FDT measurement was considered positive if there was an abnormal test in at least one eye. An abnormal test was defined as >1 depressed test points P<0.01 in the total deviation probability plot.5 No attention was paid to the reliability indices.9

Nerve fibre analyser

Testing was performed with the GDx version 2.010. For this study, we confined the analyses to GDx parameter ‘the number’. A GDx measurement was considered positive if there was an abnormal test in at least one eye. An abnormal test was defined as the number >29.5 Six images were acquired for each eye. Another six images were acquired if the first series did not yield an image with a good image quality (well-centred optic nerve head, an image in focus, equal illumination in all quadrants, and no motion artefacts). A mean image was created if there were at least two images with a good image quality.

Statistical analysis

Proportions were compared using χ2-statistics. The t-test was applied for continuous variables. Main outcome measure was the relative risk (RR) for the development of glaucomatous visual field loss of suspects with an abnormal baseline FDT/GDx test result compared to suspects with a normal baseline test result. In addition, Cox proportional hazards analyses were performed to account for follow-up duration. In converters, follow-up duration was taken until the first abnormal field that eventually resulted in the conversion classification; in non-converters, follow-up duration was taken until the end of the follow-up. A P-value of 0.05 or less was considered statistically significant.

Results

A total of 174 suspects were included in this study. Of these suspects, 134 suffered from ocular hypertension, 49 had a positive family history of glaucoma, 53 had a vertical cup–disc ratio >0.6, and 163 had at least one of these three characteristics. The remaining 11 suspects had at least two less strict characteristics (an IOP at least twice 20 mmHg, a less strict positive family history (see Materials and methods), or a vertical cup–disc ratio equal to 0.6). Table 1 shows the characteristics of the included suspects. Of the excluded suspects, 115 did not visit our department during the last year of the follow-up, 72 had a visit without an HFA test (Goldmann test or no test at all), and 62 had ambiguous test results (mainly normal test results that had to be classified as unreliable but were—in our clinical setting—not repeated within the same year because of a low suspicion of abnormality).

Table 1 Characteristics of the included glaucoma suspect patients (n=174): age at baseline, peak intraocular pressure (the highest IOP ever measured), intraocular pressure at baseline, sex, family history of glaucoma, myopia, IOP-lowering treatment at any time during follow-up, and vertical cup–disc ratio at baseline

Of the included suspects, 26 of 174 were classified as converters. This corresponds to a conversion rate of 3.7% per year (95% confidence interval (CI): 2.4–5.1% per year). In the subgroup of patients with ocular hypertension without a positive family history of glaucoma, this rate was 3.8% per year; in the subgroup of patients with a positive family history of glaucoma without ocular hypertension, this rate was 2.2% per year (P=0.58). Of the remaining 148 non-converters, 77 had a normal HFA during all follow-up visits, whereas 71 had one or more abnormal fields during follow-up (but their last field was—by definition—normal). Table 2 presents the characteristics of the converters and non-converters. On average, there was no significant difference between the two groups, except for the vertical cup–disc ratio at baseline.

Table 2 Characteristics of converters (n=26) and non-converters (n=148)

Table 3 shows the baseline FDT and GDx test results of converters and non-converters. The RR for conversion of an abnormal baseline FDT was 1.8 (CI: 0.9–3.7; P=0.10) and of an abnormal baseline GDx 2.7 (CI: 1.2–6.3; P=0.01). Positive predictive value was 0.22 for both FDT (11/(11+39)) and GDx (19/(19+68)), to be compared to the prior probability of conversion, being 0.15 (26/174; CI: 0.10–0.20). Negative predictive value was 0.88 for FDT (109/(15+109)) and 0.92 for GDx (80/(7+80)), to be compared to the prior probability of no conversion, being 0.85 (148/174; CI: 0.80–0.90). Similar results were found using Cox proportional hazards analysis, taking follow-up duration into account (FDT: hazard ratio 1.9 (CI: 0.9–4.1; P=0.10); GDx: hazard ratio 2.8 (CI: 1.2–6.7; P=0.02)).

Table 3 Baseline FDT and GDx test results for converters and non-converters

Table 4 shows the combined use of FDT and GDx. As mentioned above, 15% (26 of 174) of the suspects converted during the follow-up. Of those with a normal baseline GDx, only 8% (7 of 87) converted. In these patients, the baseline FDT apparently did not yield any additional information. Of those with an abnormal baseline GDx, 22% (19 of 87) converted. In these patients, the baseline FDT seemed to be more informative: of the patients with an abnormal baseline FDT, 30% converted, as compared to 17% of the patients with a normal baseline FDT. The small numbers of converters in the various rows of Table 4, however, do not warrant firm conclusions. In line with these findings, Cox proportional hazards analysis applied to FDT and GDx simultaneously revealed an at most modest additional yield of FDT in the case of a combined use of both techniques (FDT: hazard ratio: 1.6 (CI: 0.7–3.5; P=0.26); GDx: hazard ratio: 2.6 (CI: 1.1–6.2; P=0.04)).

Table 4 Combined use of FDT and GDx

None of the converters developed serious glaucomatous damage during the 4 years of follow-up. At the end of the follow-up, the mean deviation (MD) of the HFA of the worse eye of the converters ranged from −7.6 to −0.8 dB, with an average value of −3.6 dB.

Discussion

This study describes the outcome of a large cohort of glaucoma suspect patients who were followed prospectively in an observational study, the Groningen longitudinal glaucoma study. We found an increased risk of developing glaucomatous visual field loss in suspects with an abnormal baseline GDx.

Landers et al1 followed 62 suspects prospectively for three years. At the end of the follow-up, five patients had converted to abnormal SAP (conversion rate: 2.7% per year), all of whom had an abnormal baseline FDT. Medeiros et al2 followed 105 suspects for 3.4 years. They found conversion to abnormal SAP in 17 patients (4.7% per year) and found an RR for conversion to glaucoma of an abnormal baseline FDT of 3.2 (CI: 1.2–8.3). Mohammadi et al3 followed 160 suspects for 3.8 years. They found conversion to abnormal SAP in 16 patients (2.6% per year) and found several GDx parameters to be significant predictors of conversion. Kamantigue et al4 followed 63 eyes of 50 suspects for 62 months. They found conversion to abnormal SAP in 14% of the eyes (2.8% per year) and found an abnormal baseline FDT to be a significant predictor. Katz et al7 and Johnson et al8 found conversion rates of 2.8 and 4.4% per year, respectively. Our conversion rate (3.7% per year) and RR findings (1.8 for FDT; 2.7 for GDx) are in line with those earlier studies. As compared to the ocular hypertension treatment study, those conversion rates are somewhat high.10 A possible explanation for this difference could be that patients in the ocular hypertension treatment study were entered into the study in an—on average—earlier stage of the disease.11

An RR should always be interpreted together with a prior probability. In the presence of a low prior probability of conversion, the clinical significance of a statistically significant RR of 2.7, as found for GDx, is limited. In our study, the prior probability of conversion was 0.15 (26/174) after a 4-year follow-up. With an abnormal GDx, the probability of conversion increased from 0.15 to 0.22 (positive predictive value); with a normal GDx, the probability of no conversion increased from 0.85 (1−0.15) to 0.92 (negative predictive value).

There is an ongoing discussion as to whether FDT and GDx are able to detect glaucoma earlier than SAP. This study, and also the earlier studies,1, 2, 3, 4 showed prospectively that in some cases FDT and GDx are able to detect abnormalities earlier than SAP. However, these studies are not suitable to settle this discussion, as the cases that were included were required to have normal SAP test results at baseline, yielding a selection bias towards negative SAP findings. Several cross-sectional studies have revealed suboptimal FDT and GDx sensitivity for early glaucoma, that is, cases with abnormal SAP, but normal FDT/GDx are not uncommon.5, 12, 13 Apparently, some cases are detected by FDT/GDx before SAP, whereas other cases are detected by SAP first. Longitudinal studies starting with normal baseline test results for all tests should settle this discussion.14 Likewise, the fact that abnormal GDx findings (87/174) were more common at baseline than abnormal FDT findings (50/174; using cutoff points that yielded similar specificities with both techniques5) might suggest that structural changes precede functional changes. The fact that the suspects in our study were classified by requiring a normal functional test (HFA), however, might have induced a bias towards normal functional test findings.

For the cutoff points as used in this study, FDT and GDx have similar sensitivities and specificities (0.90 and 0.81 for FDT; 0.90 and 0.78 for GDx).5 At this cutoff point, FDT predicted only 11 of the 26 converters, whereas GDx predicted 19 (Table 3). To increase the number of converters predicted by FDT, the cutoff point could be lowered to >0. With this cutoff point, 70 of the 174 participants would have had an abnormal baseline test result and 16 of the 26 converters would have been predicted by FDT (RR: 2.4 (1.1–5.0)). At this cutoff point, the sensitivity and specificity of FDT were 0.94 and 0.70.5 For GDx, these sensitivity and specificity values can be reached with a cutoff point >25. With this cutoff point, 101 of the 174 participants would have had an abnormal baseline GDx and 21 of the 26 converters would have been predicted by GDx (RR: 3.0 (1.2–7.6)).

The vast majority of the converters were within a range often denoted as early glaucoma (HFA MD: −6 dB or more) at the end of the follow-up; none of the converters had developed severe glaucomatous damage in either eye. This finding reflects the slow nature of glaucoma, and together with the low conversion rate, it is clear that visits of glaucoma suspects intended to screen for glaucoma may generally be widely spaced in time. Some delay in detecting an incident glaucoma case will not immediately cause visual impairment, and frequent visits will induce too many false-positive findings because of the low prior probability of glaucoma (prior probability=conversion rate × intertest interval).

The Groningen longitudinal glaucoma study is a purely observational study, that is, all the data were collected as part of routine clinical care. IOP and HFA data were available for the clinician and thus could, for example, have influenced the decision to start or stop IOP-lowering medication. An exception was made for the FDT and GDx measurements. These were performed by a separate technician and were not available for the clinician. The aim of—and the philosophy behind—the study was that it should yield information concerning the additional value of FDT and GDx measurements in a clinical setting.

All analyses in this study were performed on a ‘by patient’ basis. A glaucoma patient is a patient with glaucoma in at least one eye, a suspect is a patient with normal fields in both eyes. Likewise, an FDT/GDx test result is normal if normal in both eyes, and abnormal if abnormal in at least one eye. We did not require that SAP defects had to develop in the same area or eye as predicted by baseline FDT/GDx abnormalities. After all, if a test correctly identifies a patient as having glaucoma, or as being at risk of getting glaucoma, by showing an abnormality anywhere, it is a good test. For confirmation of SAP conversion itself, however, we required—as mentioned in the Materials and methods section—topographical consistency.

Unlike SAP abnormalities, baseline FDT/GDx abnormalities were not confirmed or falsified. In this study, baseline FDT/GDx were considered determinants and SAP conversion was the dependent variable. Some misclassification of determinants results in a conservative risk estimate. Hence, our conclusion that GDx can predict SAP conversion will not be affected by the fact that we only had a single baseline test. In FDT, on the other hand, misclassification at baseline could have masked a significant result. A learning effect was studied separately, but no learning effect was found.15 In line with this finding, of the 50 patients with an abnormal FDT at baseline, only seven had a normal test result during the first follow-up visit. Patients with an apparently normal but formally unreliable HFA test result in the last follow-up year were excluded. This might be a possible source of bias.

It is an intrinsic problem of longitudinal studies that devices (or preferred treatment modalities) may change during the study. A 24-2 version of FDT (Matrix) has been developed with 54 test locations instead of 17, and a version of GDx with a variable corneal compensator (GDx-VCC) has been launched. Neither the Matrix nor the GDx-VCC was available at the time we started our study. Shortly after the introduction of the SITA strategies, SITA fast was adopted in our department. Change of strategy during a longitudinal study was deemed unwise. Two studies reported a slightly higher sensitivity for SITA standard in comparison with SITA fast;16, 17 one study reported a higher sensitivity for SITA fast.18 The highest reported sensitivity difference was 7% (based on a group of only 26 patients).17 Hence, we would have detected at most two more converters by using SITA standard instead of SITA fast.

In conclusion, in a clinical setting, abnormal GDx findings in glaucoma suspect patients are related to an increased risk of developing glaucomatous visual field loss. GDx may, therefore, be helpful for the clinician to optimise the care for this group of patients. Especially in the case of normal findings, the next screening visit may be postponed, and in the case of borderline IOP readings, the start of IOP-lowering treatment may be delayed. For FDT, the findings were less pronounced.