Obstacles on the ground or discontinuities in the ground plane, such as steps, pose hazards for people with low vision. Visual impairment is a risk factor for both falls and fractures in the elderly population.1,2 Recognizing ground-plane irregularities, such as steps and ramps, is an important component of the visual accessibility of public spaces for people with impaired vision. Visual accessibility is the use of vision to travel efficiently and safely through an environment, to perceive the spatial layout of the environment, and to update one's location in the environment. A long-term goal of our research on visual accessibility is to provide a principled basis for guiding the design of safe environments for the mobility of people with low vision.
In our previous study, we investigated the impact on the detection of steps and ramps of environmental factors, such as target-background contrast and lighting arrangements, and also viewing conditions such as distance to target.3 Adults with normal vision, wearing acuity-reducing goggles, were tested in a windowless classroom using methods like those described in this article. Among the results of the previous study, a step-up was more visible with the acuity-reducing goggles than a step-down. The effects of target-background contrast were greater than the effects of lighting arrangement. The empirical results were interpreted in the context of a probabilistic model of target cue detection. Ongoing research in our lab is extending the findings to subjects with low vision.
In the present study, we address the influence of two additional visual factors expected to enhance the recognition of ramps and steps under low-resolution viewing–surface texture and self-locomotion. Here, low-resolution viewing refers to viewing through acuity-reducing goggles (refer to METHODS for details).
Texture is thought to provide information about the distance and orientation of surfaces.4 Although computer graphics specialists have contributed much to our understanding of the image properties of real materials, with the goal of simulating such materials in virtual displays, there is a dearth of research on visual texture perception in real environments.5 However, some evidence demonstrates the impact of surface texture on real-world perception of target distance and surface slant.6–8 These studies leave open the question of whether surface texture would help or hinder the recognition of steps, ramps, or other ground-plane irregularities.
In the real world, ground-plane surface textures often result from distributions of small elements, often with low-contrast features, (e.g., grass, carpet weave, or a gravel path). With normal vision, such texture elements might be an effective source of information, but with low vision, they would often be invisible and unlikely to convey useful information. We hypothesized that surfaces with large high-contrast texture elements would enhance the identification of steps and ramps with low-resolution viewing. If so, appropriately designed visual texture patterns on walking surfaces might facilitate safe mobility for people with reduced acuity or contrast sensitivity. In our first experiment (experiment 1), subjects viewed steps and ramps on a surface with a coarse black-and-white checkerboard pattern. Because the angular size of texture elements depends on viewing distance, we expected any benefits from texture to depend on acuity and viewing distance.
Motion perception tends to be resistant to blur and contrast reduction,9–11 prompting our interest in motion cues for visual accessibility. Although steps and ramps are usually static in the real world (escalators being an interesting exception12), self-locomotion (walking) provides a common source of retinal-image motion. It is ecologically relevant to study detection and recognition of obstacles during walking because people are often mobile when making such judgments. It is possible that self-motion might enhance the visibility of low-contrast contours or yield information from motion parallax or other change-of-view cues to improve detection and recognition of ramps and steps.
In brief, we hypothesized that both surface texture and self-locomotion would enhance the recognition of ramps and steps with low-resolution vision.
EXPERIMENT 1 METHODS
Stimuli and Procedure
A large, windowless, 33.25 × 18.58-ft (10.13 by 5.66 m) classroom in the basement of the psychology building on the campus of the University of Minnesota was used as the test space for all experiments. A schematic drawing is shown in Fig. 1.
A uniform gray sidewalk (4 ft wide by 24.5 ft long, 1.3 m by 7.5 m) was constructed using hardboard deck portable stage risers (Fig. 1). This sidewalk was elevated 16 in (0.4 m) above the floor. Five possible targets were shown at a fixed location on the sidewalk's south end: a single step-up or -down (7-in height), a ramp-up or -down (7-in change of height over 8 ft), or flat (Fig. 2). A 4 × 8-ft (1.2 m by 4.3 m) by 2-in-thick rectangular panel of expanded polystyrene, painted uniform gray, formed the target. Sidewalk and target were also painted gray. Using motorized scissor jacks, the target panel was adjusted by raising or lowering one or both ends of the panel above or below the sidewalk.
The visual background for the targets was formed by the classroom floor, far wall, and right-hand wall (Fig. 3). The walls were paneled with rectangular sections of expanded polystyrene, and the section of floor on the left of the target was covered with a wooden panel (painted to match the background).
Overhead lighting was produced by four rows of three end-to-end 2 × 4-ft luminaries (recessed acrylic prismatic 4 lamp SP41 fluorescent). This lighting produced a luminance of approximately 77 cd/m2 on white squares and 5 cd/m2 on black squares in the texture pattern used in experiment 1. The overhead illumination was representative of typical ambient room lighting. For more information about the test space and apparatus, please refer to the study by Legge et al.3
Effective acuity through the subject's dominant eye (determined using an aiming task) was reduced using either one or two Bangerter occlusion foils,13 mounted in a goggle frame. The foils were attached to one (mild blur) or both (severe blur) sides of a clear acrylic lens mounted in a welding goggle frame. Examples of the effect of blur on target visibility are shown in Fig. 4. To reduce glare from illumination by the overhead fluorescent lights through the goggles, a cylindrical, black, acrylic viewing tube was attached to the goggles in front of the dominant eye. The tube reduced the field of view from approximately 48 to 33° (refer to the study by Legge et al.3 for details). Effective acuity through the blur foils was determined by measuring each subject's acuity while wearing the blur goggles using the Lighthouse Distance Acuity chart. The mean acuity with mild blur was 20/152 (logMAR = 0.88) and with severe blur was 20/884 (logMAR = 1.65). Contrast sensitivity was also estimated psychophysically (Pelli-Robson chart3) as 0.8 (mild blur) and 0.6 (severe blur). Luminance was attenuated by approximately a factor of two through the blur foils.
We compared ramps-and-steps recognition performance for a textured surface with similar recognition data for a uniform gray surface reported by Legge et al.3 The texture was a checkerboard pattern composed of large high-contrast squares. The goal was to determine whether visible texture would enhance the recognition of ramps and steps.
Twenty-four normally sighted young adults, aged 18 to 24 (mean = 21.3) years, with mean acuity of 20/16 (logMAR = −0.097) and mean contrast sensitivity of 1.73, participated. Each subject completed the experiment in one session lasting from 2 to 3 hours. The experimenter obtained informed consent in accordance with procedures approved by the University of Minnesota's institutional review board.
Targets and Texture Pattern
The sidewalk and target were covered by linoleum flooring with a continuous texture of alternating black and white high-contrast (Michelson contrast: 0.87) squares (1 ft per side). An additional 4-ft-long by 7-in-wide narrow rectangle of black and white flooring was used to cover the riser for the step-up target.
Before each testing condition, subjects were shown the targets with normal viewing (no blur). This was done to equate their previous knowledge of the targets.
During each trial, the seated subject reported which of the five targets was shown (5-alternative forced choice), with a viewing time of 4 s. Subjects viewed the targets through the blurring goggles from three distances of 5, 10, and 20 ft (1.52, 3.05, and 6.10 m), respectively. To mask auditory cues associated with changing the target configuration, subjects wore noise-reducing earmuffs and listened to auditory white noise. Subjects were instructed to turn their head to face the right-hand wall between trials, preventing them from viewing the target adjustment.
Each subject completed 60 trials, two for each of the five targets, for two blur levels (mild and severe) and three viewing distances (5, 10, or 20 ft). The appearances of the targets with approximations of the mild and severe are shown in Fig. 4. Viewing distances are shown in Fig. 1. Half of the subjects (n = 12) viewed the targets against a gray background, and the other half viewed them against a black background (painted with Valspar interior satin dark kettle black acrylic latex). Because the contrast of the gray targets against the black background was higher (Michelson contrast: 0.82) than that against the gray background (0.25), it was hypothesized that subjects would perform with higher accuracy in the former condition. Within each background group (black or gray), trials were blocked by blur level and viewing distance.
EXPERIMENT 1 RESULTS
In the presentation of results for experiment 1, data for the uniform surface conditions come from the study by Legge et al.,3 in which performance with the same black and gray backgrounds was compared. The size of the subject groups and testing conditions in the study by Legge et al. were identical to the current study, apart from the difference in surface pattern (texture vs. uniform). Accuracy for target identification (% correct) is reported. Chance accuracy was 20% because there were five targets.
To achieve normality of the group data, accuracy data were arcsine transformed before statistical testing. We conducted a repeated-measures analysis of variance on the transformed accuracy data, with three between-subjects factors—sidewalk type (texture or uniform), blur (mild or severe), and background color (black or gray)—and one within-subjects factor—viewing distance (5, 10, and 20 ft). T-tests, using a Bonferroni correction for multiple comparisons, were used in post hoc testing.
We hypothesized that texture would facilitate recognition of ramps and steps. However, contrary to expectation, the uniform sidewalk groups performed slightly better overall (71% correct) than the texture sidewalk groups (67% correct; p = 0.029). The difference was greater with severe blur: performance for the uniform and texture groups was 58% and 47% correct, respectively (p = 0.001). With mild blur, there was no significant difference between textured and uniform groups. The three-way interaction between sidewalk type, background color, and blur showed that textured sidewalk was only significantly worse than uniform sidewalk with severe blur and the black background (F (1, 88) = 11.34, p = 0.001).
Fig. 5 shows that performance was much better with mild blur (85% correct) than with severe blur (52% correct; F (1, 88) = 242.83, p < 0.0001). The high-contrast black background (88% correct) yielded better performance than the low-contrast gray background (78%) for the uniform groups (p = 0.01).3 Unexpectedly, there was no difference in performance between the black and gray backgrounds for the texture groups.
Performance varied with viewing distance (5, 10, or 20 ft) from the target (F (2, 88) = 19.72, p < 0.001). With texture and mild blur, performance was better at 10 and 20 ft than at 5 ft, but with gray and mild blur, performance was best at 10 ft (all p values <0.03). For severe blur, performance was best at the shortest distance of 5 ft for both groups and declined at longer distances, presumably due to acuity limitations (p < 0.001).
Fig. 6 shows that performance depended on the target (step-up, step-down, ramp-up, ramp-down, or flat). Both texture and uniform groups performed better on step-up than on step-down, ramp-down, or flat (all p values <0.01). Similarly, both texture and gray groups performed better on ramp-up than on step-down (all p values <0.05). Overall, target accuracies for the texture study (experiment 1) roughly correspond to those found with overhead lighting in the study by Legge et al. when the corresponding conditions are compared (e.g., black background and mild blur at 10 ft).3
EXPERIMENT 2: LOCOMOTION
In this experiment, we compared recognition performance for a stationary condition and a walking condition. In the walking condition, subjects approached the targets along the sidewalk, stopping at the designated viewing distances (5 ft or 10 ft) to make their recognition decisions. The goal of this experiment was to determine whether locomotion facilitated the recognition of ramps and steps.
Eighteen normally sighted young adults aged 18 to 36 (mean = 23.4) years, with mean acuity of 20/18 (logMAR = −0.046) and mean contrast sensitivity of 1.72, participated. Each subject completed the experiment in one session lasting from 2 to 3 hours. The experimenter obtained informed consent in accordance with procedures approved by the University of Minnesota's institutional review board.
Procedure: Walking vs. Stationary
This experiment was like the texture experiment, with the following exceptions. Subjects made target recognition judgments from distances of 5 ft or 10 ft (but not 20 ft) either after stationary viewing or after walking 10 ft along the sidewalk from a greater distance to the same viewing locations. Weight-bearing railings were added to both sides of the sidewalk to enhance safety, and subjects were asked to keep one hand on the left railing at all times during the experiment. Tactile markers were used to indicate the 5-ft, 10-ft, 15-ft, and 20-ft distances.
In the stationary trials, subjects stood at the designated viewing distance and made their recognition decisions. In the moving trials, subjects walked 10 ft toward the target, stopped at the designated viewing distance, and made a recognition decision.
Testing was conducted with the same two surface patterns on the sidewalk (uniform and textured) used in experiment 1. In both cases, the background was black (refer to Fig. 4 for examples). Because our previous results3 showed that the largest differences in performance among conditions were present with severe blur, subjects were tested with only severe blur.
Each subject completed 80 trials, two for each of the five targets, for two viewing distances (5 or 10 ft), two surface types (uniform vs. textured), and two movement conditions (walking vs. stationary). Within each surface type, trials were blocked by viewing distance and movement.
To determine whether walking trials were longer than stationary trials, we recorded three types of time measurements each for 10 trials (n = 4). Stationary trial time was the time taken in stationary trials from the experimenter's verbal signal for subjects to look at the targets until the subject made a recognition response. Average time for a stationary trial, 3.6 s, was close to the 4-s time limit in our earlier study.3 For the moving trials, total moving trial time was the time between the onset of walking and the subject's verbal recognition response. Walking time was the time from the onset of walking until the subject stopped, not including any additional time before a verbal recognition response. Average total moving trial time was 8.2 s, and average walking time was 5.8 s.
For the walking trials, subjects were asked to wait until they arrived at the designated viewing location to give their response. They usually responded very soon after reaching this point (mean of 2.4 s), so most of the trial time was taken up with walking, rather than standing after the walk. Also, as subjects responded whenever they were ready during stationary trials, it is unlikely that requiring them to wait the average duration of a moving trial (∼8 s) before giving their response would affect performance.
To achieve normality of the group data, accuracy data were arcsine transformed before statistical testing. We conducted a repeated-measures analysis of variance on the transformed accuracy data, with three within-subjects factors—locomotion (walking or stationary), sidewalk type (texture or uniform), and viewing distance (5 ft or 10 ft). The analysis revealed significant main effects of locomotion (F (1, 18) = 51.90, p < 0.0001), sidewalk type (F (1, 18) = 7.20, p < 0.01), and viewing distance (F (1, 18) = 9.96, p < 0.006).
Subjects performed much better in the moving condition (74%) than in the stationary condition (52%). A comparison of moving and stationary performance with the textured sidewalk and the uniform sidewalk is provided in Fig. 7. As in experiment 1, subjects performed better with the uniform sidewalk (67.50% correct) than with the textured sidewalk (58%; p < 0.01). Subjects performed better at the 5-ft distance (66%) than at the 10-ft distance (59%; p < 0.01).
The aim of these experiments was to determine whether surface texture and locomotion toward the target would enhance recognition of steps and ramps. Contrary to expectation, surface texture detracted from performance. As hypothesized, locomotion toward a step or ramp improved recognition compared with stationary observations. The locomotion experiment also replicated the findings of the texture experiment, showing that for severe blur, recognition of steps and ramps was poorer with the texture pattern.
Why did the surface texture interfere with recognition performance? Texture contours may mask some of the critical features for target recognition. In our previous study, we identified a set of cues useful for distinguishing among the five targets.3Fig. 8 illustrates these cues.
Our previous study3 indicated that the transition contrast cue, diagnostic of step-up, was primarily responsible for good recognition of this target. But this transition contrast cue appears to be less salient in the presence of the high-contrast texture edges (Fig. 9). The transition contrast cue is much more obvious with the uniform sidewalk than with the checked sidewalk, particularly for severe blur.
The contour clutter produced by the horizontal and vertical contrast features of the texture pattern may reduce the salience or mask the subtle contours for the L-junction cue, which is diagnostic of step-down (Fig. 10). The even more subtle bend-in and bend-out cues, for ramp-down and ramp-up, respectively, are also obscured with the textured sidewalk under conditions of severe blur. However, our results with a checkerboard pattern with strong orthogonal straight contours may represent a worst-case scenario. Other coarse patterns, such as a diamond texture (contours rotated 45° with respect to the step riser) or wide black-and-white stripes, might possibly produce better performance than the checkerboard texture. For example, stripes parallel to the bounding contour of the sidewalk would replicate the cues along this contour at the step or ramp transition, possibly increasing target visibility.
Why does locomotion enhance recognition of steps and ramps? Retinal-image motion could help in several ways. Motion parallax is known to improve depth discrimination in low vision and for normal vision under conditions of blur or low contrast.11 In the present case, motion parallax might have facilitated detection of depth differences between the boundary contours for the sidewalk and target panel. But even in the stationary viewing condition, subjects were allowed to move their heads, so the parallax cue was potentially available in this condition as well.
A second possible cue is accretion and deletion of surface features as the viewpoint changes between a nearer surface and a more distant partially overlapping surface.14 In our case, as the subject moved along the sidewalk toward the target, a little more of the lower surface came into view in the step-down condition. This might explain the greater benefit of locomotion we observed for recognition of the step-down. More generally, locomotion produces observations from a range of slightly different viewpoints. Perhaps, the integration of the multiple views provides information not available from a single viewpoint.
A third possibility is that locomotion produces greater retinal-image motion of informative image contours, enhancing their visibility. This might be especially significant for our severe blur condition because it is well known that contrast sensitivity for patterns composed of low spatial frequencies is enhanced by abrupt temporal onsets or offsets.15
The foregoing possibilities refer to enhancement of the visibility of cues by retinal-image motion. There is also evidence that the value of retinal-image motion in conveying three-dimensional information about objects and surfaces is enhanced for active compared with passive observers, even when the visual input is identical.16 These authors propose that extravisual movement-based information is incorporated into judgments of three-dimensional structure.
Although we expect the qualitative features of our results to generalize to people with reduced acuity associated with low vision, several caveats are in order.
First, the Bangerter blur foils reduce acuity and contrast sensitivity for normally sighted subjects, but are not necessarily representative of any particular form of low vision. For example, the contrast sensitivities through the blur foils associated with the two levels of acuity reduction (refer to METHODS) are not well matched to the measured correlations of acuity and contrast sensitivity in low-vision subjects.17 Our subjects differed from a typical group of low-vision subjects in other ways. Our subjects were young, whereas most subjects with low vision are older. Our subjects had to deal with reduced acuity and contrast sensitivity, whereas many people with low vision also experience visual-field loss. Our subjects undoubtedly had much less experience functioning with low-resolution vision than a typical group of people with low vision.
Second, we studied monocular viewing to simplify the optical arrangements for our subjects, and to simplify potential extension of the findings to low vision. Many people with mild or severe low vision have unequal vision status (acuities and other visual characteristics) of the two eyes, with performance determined primarily by the better eye.18 In principle, stereopsis could be a useful binocular cue in recognizing ground-plane irregularities. However, stereoacuity declines at low spatial frequencies and for unequal contrasts in the two eyes,19 which may imply weak or absent stereopsis in many cases of low vision.
Third, our subjects knew that one of the five targets was present in each trial, and where to look for it, but low-vision pedestrians navigating unfamiliar locations in the real world do not always know when and where obstacles will appear in their path. Such uncertainties pose challenges for mobility not present in our study.
We conclude that, contrary to our first hypothesis, a coarse texture pattern on the ground plane can hinder the visibility of ramps and steps under low-resolution viewing conditions. It is likely that contours associated with the texture pattern itself interfere with the visibility of pertinent cues for the ramps and steps. Consistent with our second hypothesis, we conclude that locomotion toward ramps and steps does enhance their visibility. If our results generalize to people with low vision, our findings may prove helpful in designing spaces to enhance visual accessibility. The findings may also be helpful for rehabilitation specialists who can inform their low-vision clients about the potential interfering effects of surface patterns or the advantages of locomotion in the visual exploration of their surroundings.
Tiana M. Bochsler
Minnesota Laboratory for Low-Vision Research
Department of Psychology
University of Minnesota
75 East River Rd
Minneapolis, MN 55455
We thank Muzi Chen for help testing our subjects. This work was supported by National Eye Institute, National Institutes of Health Grant EY017835 to GL.