Blog
Brain mechanism can automatically links objects in our minds research news by Yale University

Brain mechanism can automatically links objects in our minds

When people see a toothbrush, a car, a tree - any individual object - their brain automatically associates it with other things it naturally occurs with, allowing humans to build context for their surroundings and set expectations for the world.

By using machine-learning and brain imaging, researchers measured the extent of the "co-occurrence" phenomenon and identified the brain region involved. The findings appear in Nature Communications.

"When we see a refrigerator, we think we're just looking at a refrigerator, but in our mind, we're also calling up all the other things in a kitchen that we associate with a refrigerator," said corresponding author Mick Bonner, a Johns Hopkins University cognitive scientist. "This is the first time anyone has quantified this and identified the brain region where it happens."

In a two-part study, Bonner and co-author, Russell Epstein, a psychology professor at the University of Pennsylvania, used a database with thousands of scenic photos with every object labeled. There were pictures of household scenes, city life, nature - and the pictures had labels for every mug, car, tree, etc. To quantify object co-occurrences, or how often certain objects appeared with others, they created a statistical model and algorithm that demonstrated the likelihood of seeing a pen if you saw a keyboard or seeing a boat if you saw a dishwasher.

With these contextual associations quantified, the researchers next attempted to map the brain region that handles the links.

While subjects were having their brain activity monitored with functional magnetic resonance imaging, or fMRI, the team showed them pictures of individual objects and looked for evidence of a region whose responses tracked this co-occurrence information. The spot they identified was a region in the visual cortex commonly associated with the processing of spatial scenes.

"When you look at a plane, this region signals sky and clouds and all the other things," Bonner said. "This region of the brain long thought to process the spatial environment is also coding information about what things go together in the world."

Researchers have long known that people are slower to recognize objects out of context. The team believes this is the first large-scale experiment to quantify the associations between objects in the visual environment as well as the first insight into how this visual context is represented in the brain.

"We show in a fine-grained way that the brain actually seems to represent this rich statistical information," Bonner said.

 

Source: John Hopkins University
ImageCredits: David Cassolato