New research points to source of peripheral vision problems by Thea Singer January 7, 2016 Share Facebook LinkedIn Twitter September 3, 2014 – Peter Bex is a Professor of Psychology with expertise in basic and clinical vision science. You’re hustling across Huntington Avenue, eyes on the Marino Center, as the “walk” sign ticks down seconds: Seven, six, five…. Suddenly bike brakes screech to your right. Yikes! Why didn’t you see that cyclist coming with your peripheral vision? Researchers in the lab of Northeastern psychology professor Peter J. Bex may now have the answer—one that may lead to relief for those with age-related macular degeneration, or AMD, an eye disorder that destroys central vision: the sharp, straight-ahead vision that enables us to read, drive, and decipher faces. AMD could affect close to 3 million people by 2020, according to the Centers for Disease Control. William Harrison, postdoctoral fellow in psychology Photo by James Retell The problem with peripheral vision—which people with AMD particularly rely on as their central vision fails—is that it’s notoriously poor because it’s subject to “crowding,” or interference from surrounding “visual clutter,” says William J. Harrison, co-author of the study and a former Northeastern postdoctoral fellow. “You know something’s there, but you can’t identify what it is.” Bex and Harrison’s breakthrough paper, published last month in the journal Current Biology, uses computational modeling of subjects’ perceptions of images to reveal why that visual befuddlement occurs, including the brain mechanism driving it. That knowledge could pave the way for treatments to circumvent crowding. An evolutionary compromise Our vision operates on a gradient: It has high resolution in the center and progressively coarser resolution in the periphery. The diminishment is an evolutionary necessity. “If we were to have the same resolution that we have in the center of our vision across our whole visual field, we’d need a brain and an optic nerve that were at least 10 times larger than they currently are,” says Bex, who specializes in basic and clinical visual science. It’s that compromise that makes us vulnerable to crowding. “It’s not the physical properties of the eye—its shape, the number of photo receptors—that determine what we see,” says Harrison, who’s currently a postdoctoral fellow in psychology at the University of Cambridge. “It’s the wiring in the brain.” Crowding, simply put, exemplifies the limited bandwidth of our visual processing system. A new paradigm Previous attempts to understand crowding spring from that insight. One hypothesis holds that those limited resources lead our brains to “average” the symbols in our peripheral vision, in effect adding the distractions to the target image and dividing the whole by the parts to produce an unidentifiable object. Another hypothesis suggests that each part is represented accurately in the brain’s vision center, in the occipital lobe. But other regions of the brain—such as the frontal lobe, which is largely responsible for visual attention—don’t have the power to select the target image from among the others. Knowing how crowding works, when it happens, and when it doesn’t, means we can start to modify the way information is presented to reduce the crowding. —Peter Bex, Northeastern University professor Further blurring the picture, the prevailing hypotheses that explain crowding conflict with one another, requiring qualifiers for special cases. But now Bex and Harrison, using an innovative experimental design, have reconciled those differences in a single computational model based on how cells in the brain’s visual center represent what we see. “Our model integrates a combination of hypotheses and enables us to find evidence for any account of crowding,” says Harrison. Decoding the images Most vision tests for crowding ask people to decipher a single letter of the alphabet, say an “A,” surrounded by lines at various angles and distances, while focusing on a dot with their central vision. Hence they elicit either a right or a wrong answer (“Er, maybe it’s an “N”?). A demonstration of visual crowding: Fixate on the green spot. Without moving your eyes, see if you can identify the letter on the left and on the right side of the display. Most people will easily be able to identify the letter on the left, whereas the same letter is almost impossible to identify on the right because of crowding—the brain’s propensity to combine nearby visual information into a single object. Image by William Harrison In Bex and Harrison’s study, however, participants viewed the image of a broken ring, similar to the letter “C,” and noted where in the image the opening appeared. The measure produced “an index of continuous perceptions,” says Bex, thus permitting the researchers not only to know if crowding was occurring but also to create a computer simulation of how the cells in the brain actually decoded the image. “Knowing how crowding works, when it happens, and when it doesn’t, means we can start to modify the way information is presented to reduce the crowding,” says Bex. “And reducing it is what we need to do to help people with AMD.”