Introduction

The technology of visual field testing has continuously evolved since the “island of vision surrounded by a sea of blindness” was first described in An Introduction to Clinical Perimetry by Scottish ophthalmologist Harry Traquair in the 1930s [1]. Visual field testing can be performed using a variety of methods. These range from basic screening tools such as the manual confrontation technique [2], to more advanced, computer based methods. The ideal visual field test maximizes the following attributes: ease of administration, sensitivity (how well it can detect a defect), specificity (how well it can identify a normal visual field), and patient tolerability.

Boundaries of the normal visual field are approximately 60° superior, 60° nasal, 70° inferior and 100° temporal to fixation. At approximately 10°–15° nasal to the fovea, nerve fiber bundles collect at the optic nerve head. This area is devoid of photoreceptors and creates the physiologic blind spot [3]. Defects related to loss of retinal nerve fiber bundles are the most common visual field changes caused by glaucoma. Currently, these defects are most frequently assessed using standard automated perimetry, described in the first section of this review.

Individual retinal nerve fiber bundles include three general subsets of retinal ganglion cells (RGCs): magnocellular parasol cells, parvocellular midget cells, and koniocellular bistratified cells [4]. It is well established that RGCs from each of these subsets are damaged during glaucomatous neuropathy [3, 4]. Some studies suggest that individual subsets may be affected earlier or with greater severity than others depending on the patient [3, 5]. These observations form the basis for function-specific, or selective perimetry: a perimetric method designed to capture subjective responses from specific sets of RGCs. The hypotheses behind this form of perimetry remain controversial in that currently there is no consistent evidence that RGCs are damaged selectively [6, 7]. Selective perimetry may instead be demonstrating visual field loss caused by RGCs with decreased amounts of overlap and redundancy within the visual pathway, such that a visual field defect cannot be masked [8•]. Selective perimetry is discussed in greater detail within “Standard Automated Perimetry” section of this review.

Testing methods that rely on subjective patient responses can result in the acquisition of unreliable outcome measurements that negatively affect test accuracy and precision. Objective measurements of visual function circumvent these problems. “Strategies to Minimize Test Time: FASTPAC and SITA” section of this review discusses glaucoma-relevant objective visual function tests.

Standard Automated Perimetry

Automated perimetry is currently the mainstay for evaluating functional damage related to glaucoma [4, 9, 10••]. Standard automated perimetry (SAP) using an achromatic (white-on-white) stimulus is the most commonly employed functional test and is the archetype to which newer technologies are often compared. The Humphrey Field Analyzer (HFA; Carl Zeiss Meditec, Dublin, CA, USA) is the most commonly used perimeter in the United States. However, several other manufacturers provide SAP systems worldwide.

Using computer-controlled projection, SAP presents a static white light stimulus of constant size against a white background and places the stimulus at various points in the visual field. The standard stimulus size is termed “size III” and has a diameter of 4 mm2; the size can be adjusted based on the patient’s visual acuity [3, 4]. The area of vision analyzed may be the central 10°, 24°, or 30°. In the case of the Humphrey 24-2 algorithm, 54 individual points are tested [3]. The patient’s threshold (the dimmest light intensity that has a 50 % probability of being seen) is estimated at each point. To do this, the HFA presents a series of pre-calculated, incremental increases or decreases in static light stimulus intensity from subthreshold and suprathreshold levels, respectively [3, 4]. Units of luminance, called apostilbs, represent the absolute intensity of the presented light. Measured light sensitivity is provided in a logarithmic unit, otherwise known as a decibel (dB) [3]. The threshold is provided in decibels (dB ranging from 0 to 50 dB, where 0 dB is the brightest and 50 dB is the most attenuated target the perimeter is able to project [11]. It is important to note that a decibel is a relative unit. It represents the same amount of change in intensity across perimeters; however, the absolute light intensity (measured in apostilbs), of a stimulus may not be the same between machines [3].

Upon test completion, a normative database compares the patient’s sensitivities to those obtained from age-matched normal individuals. The HFA then provides the clinician with a printout (Fig. 1) that includes multiple scales used to interpret the visual field. At the top of the printout, patient identification and testing parameters are provided. Three test reliability indices are reported on the upper left of the printout. They include: (1) the number of times the patient responds when the stimulus is presented within his/her blind spot (fixation losses), (2) the number of times a patient responds without a stimulus (false positive rate), and (3) the number of times the patient fails to respond to a stimulus brighter than the previously estimated threshold at that test location (false negative rate). High values in each of these parameters indicate a visual field with low reliability and one that should be interpreted with caution [3, 11, 12•].

Fig. 1
figure 1

Humphrey visual field 24-2 right eye

The HFA then provides a numerical map of the patient’s threshold values adjacent to a corresponding gray scale. These plots are not adjusted for age or diffuse loss and so are not reliable indicators of visual field defects [12•]. No diagnosis should be made based solely on the gray scale [11]. It may be helpful, however, in highlighting areas that need further review and may function as a gross map by which to educate the patient about his or her test results.

Within the HFA-provided print-out, the total deviation plot is located below the gray scale. It provides the clinician with a representation of age-corrected deviation from normal. In this analysis, each location tested is graded as normal or abnormal via a p value [4, 11, 12•]. The total deviation plot does not correct for generalized depression of sensitivity that can be caused by cataract or pupil size. To do this, the pattern deviation plot is derived from the total deviation plot. It is located to the right of the total deviation plot. The pattern deviation plot represents localized change and provides p-values in addition to age-corrected normal values. Because it is a sensitive tool for identifying localized defects, the pattern deviation plot is a very useful plot for glaucoma detection [4].

Automated analyses also generate global indices that provide the ophthalmologist with mathematical evaluations of the visual field. These indices are found on the bottom right of the printout. The mean deviation (MD) index reflects diffuse change and is a weighted average of all the points in the total deviation plot [3]. The more negative the MD value, the brighter the average stimulus needed for patient response [12•]. The pattern standard deviation (PSD) index reflects variability across the visual field [13]. A higher PSD value reflects the presence of a more focal visual field defect [12•].

The glaucoma hemifield test (GHT) is another HFA-provided automated evaluation. GHT analyzes differences in threshold sensitivity between the upper and lower hemifields as a method to separate normal from glaucomatous visual fields. The test is based on observations that glaucomatous visual field defects typically show asymmetry with respect to the horizontal midline. The test algorithm compares five corresponding sector pairs based on the normal anatomy of the retinal nerve fiber layer [14], and analyzes the difference between the upper and lower hemifields. When used in correlation with the pattern deviation plot, the GHT has been shown to have high sensitivity and specificity for changes associated with early glaucoma [15].

Several confounding variables affect SAP results. The technician must instruct patients on how to perform the exam, properly correct any relevant refractive error, and hold a supervisory role throughout the test, redirecting patients if necessary. A recent study showed that independent of glaucoma severity stage, an inexperienced perimetrist could negatively influence the test result, possibly lowering the mean deviation [16•].

Accurate results are also patient-dependent. Alertness, cooperation and fixation all contribute to an effective test. Such factors are negatively influenced by lengthy test times and inconsistent stimulus visualization ability, leading to unreliable results. The Ocular Hypertension Study revealed that three consecutive and reliable visual field test results have greater specificity and stability than either one or two consecutive tests to confirm a visual field abnormality [17••]. In addition, other patient-related physiologic factors contribute to decreased test reliability. Increasing age is associated with reduced retinal threshold sensitivity [18], and the normally aging lens may cause enough light scatter to influence threshold measurements [19]. These confounding variables, however, are largely corrected by the HFA’s normative database and via the total and pattern deviation plots.

Even performed reliably, SAP has inherent critical limitations. In many eyes, the temporal relationship between structural and functional defects can be delayed. [20, 21]. Therefore, SAP does not provide ideal sensitivity for detection of early glaucomatous changes because a certain number of RGCs must be lost (25–50 %) before a visual field defect is detected [22, 23, 24].

Strategies to Minimize Test Time: FASTPAC and SITA

Full threshold testing can be tiresome for the patient, leading to exacerbation of patient-dependant limitations including alertness and cooperation. To minimize test time, two alternative threshold strategies have been created: FASTPAC and the Swedish interactive threshold algorithm (SITA). In contrast to full threshold SAP, FASTPAC changes threshold stimulus intensity in larger increments and crosses the threshold only once [3]. Test time is significantly reduced, but this method can deliver short-term fluctuations and has been reported to underestimate defect severity [25, 26]. The FASTPAC strategy has become largely obsolete since the application of SITA [27].

The SITA strategy has two commercially available versions: SITA standard and SITA fast. When measured against full threshold SAP, SITA standard takes about half the time to complete, and SITA fast is even shorter in duration [3]. To accomplish this, the SITA strategy not only utilizes available knowledge from normal and glaucomatous visual fields to estimate the threshold value [27, 28], the algorithm is constantly being updated throughout the test based on patient response [3]. Through this approach, SITA standard is able to decrease the actual number of stimuli presented to the patient compared to full threshold SAP, making test time approximately 7 min per eye [4]. SITA fast allows for even shorter examinations by using larger jumps in stimulus intensity between test spots, specifically in 4 dB steps that reverse once as opposed to the standard 4–2 dB increments, that cross threshold twice [4]. SITA fast may be the test of choice in patients who are unable to tolerate a longer exam.

In a study by Budenz et al., when evaluating the detection of mild glaucomatous field loss compared to full threshold SAP, SITA standard had a sensitivity of 92 % and SITA fast had a sensitivity of 85 %. Both SITA standard and SITA fast showed 100 % sensitivity for severe visual field loss [27]. In addition, studies have found less test/retest variability with SITA standard when compared to full threshold SAP [28]. Variability is comparable with SITA fast, but only when based on threshold sensitivities of more than 25 dB [29]. Reliable sensitivity and specificity, along with decreased variability, have lead to widespread use of SITA algorithms in the clinical setting.

Selective Perimetry: SWAP, FDT, and HPRP

A significant number of RGCs must be depleted before a visual field defect is measureable using SAP [30]. Selective perimetry was developed to identify patients with visual field loss not detected by SAP. This visual field testing method stimulates responses from specific subsets of ganglion cells, those derived from magnocellular, parvocellular, or koniocellular pathways. Compared to SAP, selective perimetry methods target only a small fraction of the visual system. Because fewer cells are tested, there is reduced overlap between RGC receptive fields. Selective perimetry’s ability to examine patient responses that correlate to a less redundant aspect of visual function may account for earlier detection of glaucomatous vision defects with this technology as compared to SAP [8•]. The following section will describe the current methods available to selectively measure a visual field based on isolated cell types.

Short Wavelength Automated Perimetry

Short wavelength automated perimetry (SWAP) tests the function of an RGC subgroup in the koniocellular pathway: the blue-yellow small bistratified ganglion cells [31]. This is done using a large blue stimulus (size V, 64 mm2) against a yellow background. It is available on HFA II (Carl Zeiss Ophthalmic Systems, Humphrey Division, Dublin, CA, USA) and on the Octopus 300 and 900 perimeters. Both full threshold and SITA strategies can be used, and aside from the larger stimulus size, all features of testing are similar to SAP.

SWAP has shown improved sensitivity for identification of early glaucomatous vision loss compared with SAP [32]. In fact, SWAP may detect functional deficits up to 5 years earlier than SAP in some patients [32, 33]. In addition, in many cases, defects detected by SWAP correlate with early structural glaucomatous damage [34, 35]. Moreover, some longitudinal studies suggest that SWAP may provide a better predictor of photographically-determined structural change than SAP [36, 37•, 38].

A disadvantage of the SWAP technique is longer test time; this can decrease both patient satisfaction and clinical practice efficiency. Integration of the SITA threshold strategy has reduced test duration [39]. However, SWAP is also influenced by media opacities and shows significant intratest and intertest variability compared with SAP. This may lead to false positive test results in many glaucoma suspects [4042].

Frequency Doubling Technology

Another available form of selective perimetry is frequency-doubling technology (FDT). FDT stimulates large-diameter ganglion cells in the magnocellular pathway, specifically a subset of parasol cells called M y ganglion cells [43]. The test employs the “frequency doubling illusion” as described by Kelly in 1966. The “illusion” consists of alternating dark and light bars shown at high temporal frequency and low spatial frequency. When viewed, the bars appear to double in number [44]. Specifically, the FDT stimulus is a vertical sine wave grating with spatial frequency of 0.25 c/deg displayed with counterphase flickering at a temporal frequency of 25 Hz. The C-20 presentation pattern tests the central 20° of the visual field and presents a 10° square target at 16 locations (four per quadrant). A small central target (5°) is also projected on the macular region. For the N-30 presentation, two additional points (total of 19 points) are presented within the nasal visual field.

Both screening and full threshold algorithms are available for FDT, and the test time is short: approximately 1 and 5 min, respectively [10••]. The Humphrey FDT Perimeter (Carl Zeiss Meditec; Dublin, CA, USA) is easy to transport and relatively inexpensive. In addition, the test results are less affected by media opacities than SWAP [45]. Studies have shown FDT to have a high sensitivity and specificity for differentiating non-glaucoma individuals from those with glaucoma [46]. In fact, evidence also suggests that some false positive results may actually represent detection of glaucomatous damage earlier than SAP [47••]. However, in general, when used in the clinic setting, the FDT screening protocol offers levels of specificity and sensitivity similar to SAP. FDT has also been evaluated as a population-based screening tool [4850]. At the population level, it does not appear to have sensitivity and specificity measures needed to detect early glaucoma. However, it may be useful for population-based testing when the goal is detection of advanced disease.

The Matrix perimeter (Carl Zeiss Meditec; Dublin, CA. Welch Allyn) is the most recent FDT testing system. It provides video eye monitoring, along with improved spatial resolution of visual field defects. This is done by using a greater number of testing locations and smaller, 4° square stimuli, spaced in a 6° grid. The testing pattern is similar to the 24-2 protocol used in the HFA. In addition, the Matrix perimeter offers programs containing 2° wide stimuli for testing the central visual field (10-2), improved threshold calculations, and video eye monitoring [51, 52]. Like FDT, Matrix FDT is able to offer a sensitivity and specificity similar to SAP [53]; it does, however, have longer test times than standard FDT (6 min per eye).

High-Pass Resolution Perimetry

High-pass resolution perimetry (commercially available as the Ophthimus High Pass Resolution Perimeter, Hightech Vision, Goteburg, Sweden) allows another selective evaluation of visual field function that tests the parvocellular system of RGCs. Here, ring shaped stimuli of different sizes but with fixed contrast are presented at 50 different locations in the central 30° of the visual field [54, 55]. Some evidence suggests that HPRP may be able to diagnose early glaucomatous damage [56, 57]. For example, Iester et al. compared SAP (HFA 30-2), FDT (C-20), and HPRP parameters in 52 eyes determined to have glaucoma by disc examination, visual field, or both and in 30 eyes with IOP of more than 21 mmHg but no other abnormalities. Within the entire group, a significant correlation was found between HFA indices and those of HPRP or FDT. However, FDT detected 29 visual fields as abnormal despite normal HFA and HPRP detected 26 fields as abnormal despite normal HFA. The authors concluded that FDT and HPRP may be useful to identify early glaucomatous disease [57].

Other studies have not found that HPRP is superior to other perimetry tests. Sample et al. compared HPRP with SWAP, FDT N-30, and SAP. At a specificity of 90 % for photographic progression of the optic disc, the sensitivity of the best SWAP parameter was 45 %, that for FDT was 68 %, and that for HPRP was 52 % [58].

Electrophysiological studies provide a novel avenue for diagnosis of glaucomatous nerve damage. These techniques circumvent the inherent subjectivity that can be problematic to the visual function tests described above. Many of the following methods are well studied in retinal disease and hold promise for analysis of optic nerve disorders including glaucoma.

Electroretinography

Electroretinography (ERG) provides a record of the bioelectrical changes that occur across the retina. This activity is recorded using an electrode placed on the corneal or bulbar conjunctival surface. The classic ERG displays a negative a-wave (representing the photoreceptors) and a positive b-wave (representing the bipolar cell layer). These recorded electrical changes occur in response to a visual stimulus such as a flash of light. Comparison is then made to characteristic normal responses [59, 60]. Evaluation of electrophysiological techniques using experimental glaucoma models suggests that they may be helpful in detection of early glaucomatous changes [61].

The prototypal ERG provides the clinician with a summated, pan-retinal response, making localization of retinal pathology difficult. However, in the early 1990s, a multifocal stimulus was first used to evaluate topographic retinal responses [62]. The multifocal ERG (mfERG) has since come into widespread clinical use. It offers insight into many vision-threatening diseases that affect the outer retina including retinitis pigmentosa, macular degeneration, and retinal detachment.

With respect to glaucoma diagnosis, no simple correlation exists between mfERG responses and the retinal dysfunction observed in glaucomatous visual field defects. This is because the contribution of retinal ganglion cell activity to the mfERG response is indirect and complicated to measure [63]. However, several different mfERG stimulus and analysis paradigms have been evaluated for use in glaucoma testing [64]. In fact, some recent studies show good differentiation between normal and glaucomatous patients [65•]. In addition, recent work using a laser-induced primate model of glaucoma showed significant correlations between focal measurements of global flash mfERG, regional retinal nerve fiber layer thickness, and local visual field sensitivity [66].

Despite promising initial studies, early detection of glaucoma using mfERG has not yet been clinically established [10••, 67]. In fact, some authors have suggested that mfERG changes do not show a clinically useful correlation with glaucomatous damage [68]. However, mfERG technology continues to evolve and there is hope that it will become an important tool for diagnosis and management of glaucoma [6870].

The pattern electroretinogram (PERG) can detect RGC dysfunction in glaucoma and glaucoma suspects [71••, 72]. PERG presents a reversing black and white gradient, or checkerboard stimulus, to the central retina and provides a direct measure of RGC function [71••, 72, 73]. Preliminary data suggest that PERG abnormalities are measureable in early glaucoma and show moderate correlation with SAP values and with structural optic nerve assessments [7476]. In addition, recent studies indicate that reversal of PERG abnormalities occurs following medical and surgical intervention for glaucoma. Therefore, PERG may offer a useful measure both of RGC dysfunction in early-stage disease (preperimetric glaucoma) and of an individual patient’s response to treatment [72, 77]. Of note, PERG is negatively correlated with intraocular pressure, age, media opacity, and visual acuity [71••]. Therefore, to be clinically useful, results must be compared to a normative database and adjustments made for these factors. [71••, 78] Also, PERG requires a more skilled examiner than that for automated visual field tests [79]. A simplified PERG known as PERGLA (Pattern Electroretinogram for Glaucoma Detection; Glaid, Lace Elettronica, Pisa, Italy) rectifies some of the shortcomings of standard PERG. This updated paradigm is fast, completely automated, and uses skin rather than corneal electrodes. PERGLA has an improved signal-to-noise ratio compared with standard PERG, and is less affected by eye movements [80]. PERGLA amplitude is also affected by age, IOP, cataract, and visual acuity [80]. In addition, amplitudes show overlap between normal and glaucomatous individuals. [8183]

The photopic negative response (PhNR) has also been evaluated as a method of glaucoma diagnosis. This response originates primarily from activity of RGCs and consists of a slow negative potential that follows the b-wave on ERG [84, 85]. Studies have found reduction in PhNR amplitude to be a sensitive indicator of glaucoma [84] and the focal PhNR to be a valid method to detect functional loss in early and intermediate glaucoma [86]. The PhNR amplitude also correlates with morphology and structural changes seen in eyes with open-angle glaucoma [87]. Advantages of the PhNR over the PERG include less media opacity effect and no refractive correction requirement [84]. However, media opacities do still have some effect on the PhNR; dense cataract or vitreous hemorrhage make focal PhNR testing impossible. Also, during testing, the patient must gaze at each fixation point for 1 min with minimal eye movement. Thus, this test has some of the patient reliability issues faced with subjective perimetry [88].

Visual Evoked Potentials

Visual evoked potentials (VEP) are electrical signals produced by the visual cortex in response to a light stimulus. These responses have been traditionally used to evaluate macular function, but recent work has also focused on optic nerve disease [61]. Specifically, the multifocal VEP (mfVEP) may be useful in glaucoma detection [10••]. Like the mfERG, it provides topographic mapping of retinal responses. Multiple VEPs can be recorded rapidly and simultaneously [40]. Hood et al. [89] have found the mfVEP to be a promising alternative for patients who repeatedly record unreliable visual fields by standard perimetry. Moreover, using interocular analysis, mfVEP can detect mild, unilateral defects with high sensitivity. However, this test can miss abnormalities when damage exists in the same location within both eyes [89, 90]. Multifocal VEP and SAP have been shown to correspond in up to 81 % of early glaucomatous defects [91]. In addition, in cases of uncertain or subjective field defects, studies have found mfVEP to be a valuable adjunct to SAP in finalizing the diagnosis of glaucoma [92].

Conclusion

There is no assessment of visual function that approaches the “ideal test.” However, in comparison to full threshold SAP, algorithms such as SITA have improved many test performance measures.

Currently, the most commonly used selective perimetry modalities are FDT and SWAP. For routine clinical use, the FDT screening protocol offers levels of specificity and sensitivity comparable to SAP. In addition, it may hold promise as a tool to diagnose glaucoma in the community-based setting. SWAP is likely capable of detecting glaucoma earlier than SAP but may indicate false positive results more frequently than SAP.

Objective methods of evaluating visual function offer potential advantages over traditional subjective visual field tests. However, further studies are needed in order to determine the role of these technologies in glaucoma diagnosis. In addition, equipment availability, provider expertise, and patient acceptance issues currently limit adoption of an objective method for routine clinical use. Characteristics and limitations differ among the modalities used for functional testing in glaucoma. Understanding how these methods vary will assist the physician in making a diagnosis of disease.