The entire process that occurs when your brain makes quick decisions based on color or motion may take place in an area just behind your forehead.
In this brain region, known as the prefrontal cortex, researchers have found that color and motion signals converge in a specific circuit of neurons.
In a study published in the journal Nature, they hypothesize that these neurons act together to make two snap judgments: Is color or motion the most relevant sensory input in the current context, and what action to take as a result.
Until now, neuroscientists have believed that decisions of this sort involved two steps: One group of neurons performed a gating function to ascertain whether motion or color was most relevant to the situation, and a second group of neurons considered only the sensory input relevant to making a decision under the circumstances.
But in a study that combined brain recordings from trained monkeys, and a sophisticated computer model based on that biological data, Stanford University neuroscientist William Newsome and three co-authors discovered that the entire decision-making process may occur in a localized region of the prefrontal cortex.
“We were quite surprised,” says Newsome, a professor of neurobiology at Stanford School of Medicine.
He and Valerio Mante, a former Stanford neurobiologist now at the University of Zurich and the Swiss Federal Institute of Technology, began the experiment expecting to find that the irrelevant signal, whether color or motion, would be gated out of the circuit long before the decision-making neurons went into action.
“What we saw instead was this complicated mix of signals that we could measure, but whose meaning and underlying mechanism we couldn’t understand,” Newsome says. “These signals held information about the color and motion of the stimulus, which stimulus dimension was most relevant, and the decision that the monkeys made.
“But the signals were profoundly mixed up at the single neuron level. We decided there was a lot more we needed to learn about these neurons, and that the key to unlocking the secret might lie in a population-level analysis of the circuit activity.”
Software model of neurons
To solve this brain puzzle, the neurobiologists began a cross-disciplinary collaboration with Krishna Shenoy, professor of electrical engineering at Stanford, and David Sussillo, a postdoctoral scholar in Shenoy’s lab.
Sussillo created a software model to simulate how these neurons worked. The idea was to build a model sophisticated enough to mimic the decision-making process, but easier to study than taking repeated electrical readings from a brain.
The general model architecture they used is called a recurrent neural network: a set of software modules designed to accept inputs and perform tasks similar to how biological neurons operate. The scientists designed this artificial neural network using computational techniques that enabled the software model to make itself more proficient at decision-making over time.
“We challenged the artificial system to solve a problem analogous to the one given to the monkeys,” Sussillo explains. “But we didn’t tell the neural network how to solve the problem.”
As a result, once the artificial network learned to solve the task, the scientists could study the model to develop inferences about how the biological neurons might be working.
The entire process was grounded in the biological experiments.
The monkeys decide
The neuroscientists trained two macaque monkeys to view a random-dot visual display that had two different features—motion and color. For any given presentation, the dots could move to the right or left, and the color could be red or green.
The monkeys were taught to use sideways glances to answer two different questions depending on the currently instructed “rule” or context. Were there more red or green dots (ignore the motion)? Or, were the dots moving to the left or right (ignore the color)?
Eye-tracking instruments recorded the glances, or saccades, that the monkeys used to register their responses. Their answers were correlated with recordings of neuronal activity taken directly from an area in the prefrontal cortex known to control saccadic eye movements.
The neuroscientists collected 1,402 such experimental measurements. Each time the monkeys were asked one or the other question.
The idea was to obtain brain recordings at the moment when the monkeys saw a visual cue that established the context (either the red/green or left/right question), and what decision the animal made regarding color or direction of motion.
It was the puzzling mish-mash of signals in the brain recordings from these experiments that prompted the scientists to build the recurrent neural network as a way to rerun the experiment, in a simulated way, time and time again.
‘Multitasking like crazy’
As the four researchers became confident that their software simulations accurately mirrored the actual biological behavior, they studied the model to learn exactly how it solved the task. This allowed them to form a hypothesis about what was occurring in that patch of neurons in the prefrontal cortex where perception and decision occurred.
“The idea is really very simple,” Sussillo explains. Their hypothesis revolves around two mathematical concepts: a line attractor and a selection vector.
The entire group of neurons being studied received sensory data about both the color and the motion of the dots.
The line attractor is a mathematical representation for the amount of information that this group of neurons was getting about either of the relevant inputs, color or motion.
The selection vector represented how the model responded when the experimenters flashed one of the two questions: Red or green, left or right?
What the model showed was that when the question pertained to color, the selection vector directed the artificial neurons to accept color information while ignoring the irrelevant motion information. Color data became the line attractor. After a split second, these neurons registered a decision, choosing the red or green answer based on the data they were supplied.
If question was about motion, the selection vector directed motion information to the line attractor and the artificial neurons chose left or right.
“The amazing part is that a single neuronal circuit is doing all of this,” Sussillo says. “If our model is correct, then almost all neurons in this biological circuit appear to be contributing to almost all parts of the information selection and decision-making mechanism.”
Newsome put it like this: “We think that all of these neurons are interested in everything that’s going on, but they’re interested to different degrees. They’re multitasking like crazy.”
The Howard Hughes Medical Institute, the Air Force Research Laboratory, a Pioneer Award from the National Institutes of Health, and the Defense Advanced Research Projects Agency supported the work.
Source: Stanford University