Looking leads gaze to objects; seeing recognizes them. Visual crowding makes seeing difficult or impossible before looking brings objects to the fovea. Looking before seeing can be guided by saliency mechanisms in the primary visual cortex (V1). We have proposed that looking and seeing are mainly supported by peripheral and central vision, respectively. This proposal is tested in an observer with central vision loss due to macular degeneration, using a visual search task that can be accomplished solely through looking, but is actually impeded through seeing. The search target is an uniquely oriented, salient, bar among identically shaped bars. Each bar, including the target, is part of an “X” shape. The target’s X is identical to, although rotated from, the other X’s in the image, which normally causes confusion. However, this observer exhibits no such confusion, presumably because she cannot see the X’s shape, but can look towards the target. This result demonstrates a critical dichotomy between central and peripheral vision.
Vision involves looking and seeing. Looking shifts gaze and attention to objects; seeing identifies those objects. Typically, looking and seeing are seamlessly coordinated. However, Figure 1 illustrates an example in which they battle. This shows stimulus condition in a previous study (Zhaoping & Guyader, 2007), along with the characteristic confusion exhibited by normal observers as they searched for an uniquely oriented bar. Crowding in this cluttered image makes seeing difficult before looking. Nevertheless, guided by saliency, gaze reached the target within one second in 50% of the trials in untrained observers (Zhaoping & Guyader, 2007; Zhaoping, 2024). (The target’s unique orientation makes its location salient by V1 mechanisms (Li, 2002).) After seeing the “X” containing the target bar at fovea, gaze abandoned the target in confusion to continue searching elsewhere. This confusion arises from rotational invariance in shape recognition, since this X is not distinctive from the background X’s.
Looking and seeing in a trial of condition by Zhaoping & Guyader (2007). On top of a search image (spanning in visual angle), in black and white, is superposed a gaze trajectory (together with explanations) in red, magenta, and cyan, from the start, to the later moments, of a search. The gaze started at the image center when the image appeared. The target, a bar uniquely tilted counterclockwise from vertical, was in visual angle from the initial fixation. The first saccade (red) led the gaze to the target—looking. About 0.5 seconds later, the gaze departed from the target (in magenta and cyan). Visual crowding makes the target bar and its associated X illegible before the looking. Seeing, recognizing this as just another X, occurred after the gaze reached the target, causing confusion and gaze departure.
Figure 2 shows example search images adapted for the current study for condition and the control conditions , , and . In each, the target is an oblique bar tilted either uniquely clockwise or uniquely counterclockwise from vertical. Cardinal, horizontal or vertical, bars only appear in conditions and , each intersects an oblique bar to make an X. In , all the X’s have the same shape, causing the confusion once the target X is seen. This confusion is absent in since the target’s X is distinctively thinner. Zhaoping & Guyader (2007) showed that conditions and were equivalent in the time looking needed to gaze at the target for the first time. However, in condition , the gaze seldom abandoned the target. Hence, a longer response time (RT) to report the target in condition than reflects the confusion triggered by seeing the X’s shape.
Example search images in four different conditions: , , , and . Observers were asked to find a uniquely oriented target bar in each image on a touch-screen display, and touch the target as quickly as possible. In images, all bars are clockwise or counterclockwise from vertical; the target is uniquely oriented. Modifying images gives images, when a horizontal or vertical bar intersects each original bar to make an X. Modifying images gives images, when the target bar’s orientation tilts just from the intersecting horizontal/vertical bar. Removing all the horizontal/vertical bars from images gives images.
Central vision is essential for seeing the target in clutter. Hence, our 86-year-old observer, with her central vision loss due to age-related macular degeneration (MD), may be free from the confusion. She gave up reading ten years ago, but can manage routine housework. She and four age-matched control observers (8288 years old) performed the search on a touch-screen display spanning cm, containing 9 rows 12 columns of search items. Each observer sat at a self-determined distance (about 50 cm) from the display to perform the task comfortably, making each stimulus bar about in length.
The four conditions were randomly interleaved, with 15–20 trials per condition in each experimental session. A search image appeared after observers’ button press, and disappeared once the screen was touched (to report the target) or when 60 seconds had elapsed since its onset. The MD individual complained of not seeing the search items well. Days later, she participated in a second session in which each bar was enlarged by 50% (in length and width) in a sparser (6 rows 8 columns) search array.
A test trial was deemed successful if the observer touched, within 60 seconds, a screen location no more than 12.5% of the screen’s width/height away from the target’s center horizontally/vertically. RT (, , , and ) is the average time needed to touch the screen in the successful trials (Figure 3A). The time-out rate, nonzero only in condition , is the fraction of the trials without a screen touch within 60 seconds (Figure 3B). Performance accuracy is the fraction (, , , and ) of the non-time-out trials that were successful (Figure 3C). Each control observer manifested the X confusion, with significantly ( values from statistical tests) and, except for one control observer, had either substantial time-out rate () for condition or had significantly ().
Response time (RTs), time-out rates, and accuracies in the search task.
By contrast, the MD observer exhibited no X confusion in her first session. She had no time-out trials. Her seconds was statistically indifferent from her (), (), and (). After each RT is normalized (divided) by the observer’s , her was significantly shorter () than the average of the control observers. Her and were statistically equivalent (), and significantly better () than , the chance level, although worse () than and of each control observer. Manifestly, she searched mainly by looking (orienting and touching) without seeing.
In her second session, using enlarged search items in a sparser array, she could see much better, such that her became statistically equivalent () to the in three out of four control observers, without changing her significantly (). Consequently, she manifested the X confusion, with (), (), a 30% time-out rate for condition , and a significantly () increased (normalized) (statistically equivalent () to the average of the control observers).
In natural behavior, differential functional specializations by central and peripheral vision are obscured, especially with trans-saccadic integration of recognition (Stewart et al., 2020). Our MD individual’s search behavior provides a clearer demonstration that looking and seeing are mainly functions of peripheral and central vision, respectively, as proposed by the Central-peripheral Dichotomy theory (Zhaoping, 2019).
Footnotes
Acknowledgements
I thank Peter Dayan, Charles Leek, and an anonymous reviewer for very helpful comments, and the observers for their participation in the experiments.
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
This work is supported in part by funding from the Max-Planck-Society and the University of Tübingen.
ORCID iD
Li Zhaoping
How to cite this article
Zhaoping, L. (2024). Looking with or without seeing in an individual with age-related macular degeneration impairing central vision. i-Perception, 15(4), 1–5.
References
1.
LiZ. (2002). A saliency map in primary visual cortex. Trends in Cognitive Sciences, 6, 9–16.
2.
StewartE. E.ValsecchiM.SchützA. C. (2020). A review of interactions between peripheral and foveal vision. Journal of Vision, 20, 1–35.
3.
ZhaopingL. (2019). A new framework for understanding vision from the perspective of the primary visual cortex. Current Opinion in Neurobiology, 58, 1–10.
4.
ZhaopingL. (2024). Peripheral vision is mainly for looking rather than seeing. Neuroscience Research, 201, 18–26.
5.
ZhaopingL.GuyaderN. (2007). Interference with bottom-up feature detection by higher-level object recognition. Current Biology, 17, 26–31.