Try it: Tiny eye movements affect how we see contrast

(Credit: Getty Images)

Contrast sensitivity depends in part on small eye movements that we’re not even aware of making, report researchers.

It is often difficult for a driver to see a person walking on the side of the road at night—especially if the person is wearing dark colors. One of the factors causing this difficulty is a decrease in contrast, making it hard to segment an object, such as a person, from its background.

Researchers previously believed that the optics of the eye and processing in the brain mainly dictated contrast sensitivity function—the minimum level of black and white that a person needs to detect a pattern. The new research shows that eye movement is another factor at play.

“Historically these movements have been pretty much ignored,” says Michele Rucci, a professor of brain and cognitive sciences at the University of Rochester. “But what seems to be happening is that they are contributing to vision in a number of different ways, including our contrast sensitivity function.”

Pelli-Robson test gif
Test your contrast sensitivity: The Pelli-Robson test is one type of test that measures contrast sensitivity. The test consists of a chart with six letters per line, and contrast varying from high at the top left to low at the bottom right. To test your contrast sensitivity function, read the letters starting with the highest contrast in the top left, until you are unable to read two or three letters in a single group. Each group has three letters of the same contrast level. (Credit: U. Rochester)

Critical movements

When we fix our eyes on a single point, the world may appear stable, but at the microscopic level, our eyes are constantly jittering. These small eye movements, once thought to be inconsequential, are critical to the visual system in helping us reconstruct a scene, Rucci says.

“Some scientists believed that because they are so small, the eye movements might not have much impact, but compared to the size of the photoreceptors on the retina, they are huge, and they are changing the input on the retina,” she says.

Think of a scene or object like a computer image made up of different pixels, or points. Each point is a different color, intensity, luminance, and so on. Our eyes take in signals from each of the points and project the signals onto photoreceptors on the retina: the arrangement of these points makes a spatial pattern that we perceive as a scene or object. But, if a spatial pattern is projected as a stationary image, it will fade from view once the retina’s photoreceptors become desensitized to the signal—like a student who becomes bored in class if the teacher repeats the same information over and over again.

Researchers have long known that the tiny eye movements—always jittering and taking in different points—continually change the signal to the retina and refresh the image so it does not fade. The new research suggests, however, that these movements do more than prevent fading; they are one of the very mechanisms by which the visual system functions, Rucci says.

“The way the visual system encodes information is based on these temporal changes. Eye movements transform a spatial pattern into temporal changes on the retina.”

The system is similar to that involved in the sense of touch: to glean information about the surface of a solid object, we don’t simply place our fingertips on the surface, but also move them along the object. We perceive the object based on the interaction between a sensory process (the tactile receptors in our fingers) and a motor process (the way we move our fingertips).

“Since our eyes are never at rest even when we fixate a point in the visual scene, a similar mechanism holds for vision,” says coauthor Antonino Casile, a researcher at the Istituto Italiano di Tecnologia (Italian Institute of Technology). “Contrast sensitivity results from the interaction of two processes: a sensory process—the response properties of neurons in the early visual system—and a motor process.”

Black and white stripes

In order to measure contrast sensitivity and whether or not eye movements play a role, the researchers showed human participants gratings with black and white stripes. The researchers gradually varied the width of the stripes, making them “thinner and thinner, until the participants eventually said they no longer saw separate bars,” Rucci says.

The width of the bars is known as the spatial frequency. For each spatial frequency, researchers measured the minimum level of black and white that participants needed to be able to see a contrast, while, at the same time, carefully measuring their eye movements.

To measure contrast sensitivity, the researchers showed human participants gratings with black and white stripes of different widths (known as spatial frequency, image A above). For each frequency, they determined the minimum amount of contrast (separation from black and white, image B above) that would enable the subjects to still see the grating pattern. The resulting function of spatial frequency is known as contrast sensitivity function. You can see your own contrast sensitivity function in image C (below), where frequency varies on the horizontal axis and contrast varies on the vertical axis. (Credit: U. Rochester)

The researchers then simulated this task in a computer model of the retina to see if the responses of neurons in the retina matched the human subjects’ contrast sensitivity.

“We found that they are only compatible when we include the motion of the eye movements,” Rucci says. “When we don’t include this movement factor in the computer model, the simulated neurons don’t give the same responses that the subjects do.”

Knowing that eye movements affect contrast sensitivity, researchers can input this factor into models of human vision, providing more accuracy in understanding exactly how the visual system processes information—and what can go wrong when the visual system fails.

The research also highlights that movement and motor behavior may be more fundamental to vision than previously thought, Rucci says.

“Vision isn’t just taking an image and processing it via neurons. The visual system uses an active scheme to extract and encode information. We see because our eyes are always moving, even if we don’t know it.”

The study appears in the journal eLife.

Source: University of Rochester