Humans can do something robots can't (yet). We can look at a target, close our eyes, and reach for the target and grasp it without opening our eyes.
It sounds simple, doesn't it. But it's not! Let's look at an example.
First of all, vision and reaching occur in different coordinate systems. Vision is Cartesian (after a few calculations) - every object we see has x, y, and z coordinates. But reaching (and all other forms of somatic motor activity) are vectorized, the movement is independent of its starting coordinates.
It's just like computer graphics, where bitmaps and jpg's are Cartesian, whereas svg's and PostScript files are vectorized. So computer graphics is a good way to get a clue about how to translate from one representation to the other.
Let's consider our 5 senses. Only one of them is truly three dimensional, that's hearing. Vision has no "rear", it only happens in front of us. And even though our bodies have sensory receptors in back, the extent of the map is strictly limited by body space (we can't feel beyond the boundaries of our skin).
Our simple example that we will consider, comes from an area of the human brain called "superior colliculus". It's in the midbrain, approximately at the same level as the cerebellum. It supports a reflex, specifically an eye movement reflex - if someone sticks you with a pin, your eyes are directed to look at the spot where it happened. This is a reflex, if you remove the cerebrum entirely it still persists.
The issue and the question is, there is no VISUAL stimulus in this scenario. So how do the eyes know where to move?
Obviously, there must be some alignment of the body map with the visual field. The location "right index finger" is translated by the superior colliculus, to "20 degrees down and 5 degrees to the right", which causes your eye to move to that location.
How do these maps align? What is the algorithm by which alignment occurs?
First of all, superior colliculus (SC) gets input from the retina and the visual cortex. This is a topographic mapping. If I flash a light in front of you, your eyes will move to exactly the coordinates of the flashing. This part is fairly intuitive and easy to understand - wherever the light appears, is the target for your saccade.
And we kinda-sorta understand how the visual map is registered with the eye movement map. But the SC is a layered structure, and in addition to topographic input from the retina and visual cortex, there is also a layer that receives input from the somatic sensory system. It is this latter layer we are concerned with.
Now it turns out, that in genetically engineered mice that have a duplicated mapping from the retina to the SC, the axons from V1 split when they grow, to form two entire mappings of the visual field. This occurs very early in development, as soon as the mice can see.
pubmed.ncbi.nlm.nih.gov
This indicates the alignment is "activity dependent", in other words the active cortical axons are trying to find the active retinal signal - and if the retinal signal exists in two places, the axons will split and connect with both.
However the body layer is different. There is no coactivation in the body layer. Instead what happens is there's a chemical gradient that lines up the maps. The body map and visual map both generate gradients of ephrin-A, and the axons align themselves along the gradients.
www.jneurosci.org
The stimulus seeking in the photic reflex is a primitive form of vectorization. It says "move to" this location independently of the starting coordinates of the eyes. In a way, it is much like the reaching for an occluded stimulus, which tells the body to "reach here" independently of posture or current body location.
What is particularly fascinating about the alignment of the body map to the visual map, is it occurs from the somatic sensory cortex, not the motor cortex.
We know what the sensory map looks like, it has no relationship whatsoever to the visual field. This is what it looks like:
It sounds simple, doesn't it. But it's not! Let's look at an example.
First of all, vision and reaching occur in different coordinate systems. Vision is Cartesian (after a few calculations) - every object we see has x, y, and z coordinates. But reaching (and all other forms of somatic motor activity) are vectorized, the movement is independent of its starting coordinates.
It's just like computer graphics, where bitmaps and jpg's are Cartesian, whereas svg's and PostScript files are vectorized. So computer graphics is a good way to get a clue about how to translate from one representation to the other.
Let's consider our 5 senses. Only one of them is truly three dimensional, that's hearing. Vision has no "rear", it only happens in front of us. And even though our bodies have sensory receptors in back, the extent of the map is strictly limited by body space (we can't feel beyond the boundaries of our skin).
Our simple example that we will consider, comes from an area of the human brain called "superior colliculus". It's in the midbrain, approximately at the same level as the cerebellum. It supports a reflex, specifically an eye movement reflex - if someone sticks you with a pin, your eyes are directed to look at the spot where it happened. This is a reflex, if you remove the cerebrum entirely it still persists.
The issue and the question is, there is no VISUAL stimulus in this scenario. So how do the eyes know where to move?
Obviously, there must be some alignment of the body map with the visual field. The location "right index finger" is translated by the superior colliculus, to "20 degrees down and 5 degrees to the right", which causes your eye to move to that location.
How do these maps align? What is the algorithm by which alignment occurs?
First of all, superior colliculus (SC) gets input from the retina and the visual cortex. This is a topographic mapping. If I flash a light in front of you, your eyes will move to exactly the coordinates of the flashing. This part is fairly intuitive and easy to understand - wherever the light appears, is the target for your saccade.
And we kinda-sorta understand how the visual map is registered with the eye movement map. But the SC is a layered structure, and in addition to topographic input from the retina and visual cortex, there is also a layer that receives input from the somatic sensory system. It is this latter layer we are concerned with.
Now it turns out, that in genetically engineered mice that have a duplicated mapping from the retina to the SC, the axons from V1 split when they grow, to form two entire mappings of the visual field. This occurs very early in development, as soon as the mice can see.
Retinal input instructs alignment of visual topographic maps - PubMed
Sensory information is represented in the brain in the form of topographic maps, in which neighboring neurons respond to adjacent external stimuli. In the visual system, the superior colliculus receives topographic projections from the retina and primary visual cortex (V1) that are aligned...
This indicates the alignment is "activity dependent", in other words the active cortical axons are trying to find the active retinal signal - and if the retinal signal exists in two places, the axons will split and connect with both.
However the body layer is different. There is no coactivation in the body layer. Instead what happens is there's a chemical gradient that lines up the maps. The body map and visual map both generate gradients of ephrin-A, and the axons align themselves along the gradients.
Alignment of Multimodal Sensory Input in the Superior Colliculus through a Gradient-Matching Mechanism
The superior colliculus (SC) is a midbrain structure that integrates visual, somatosensory, and auditory inputs to direct head and eye movements. Each of these modalities is topographically mapped and aligned with the others to ensure precise behavioral responses to multimodal stimuli. While it...
The stimulus seeking in the photic reflex is a primitive form of vectorization. It says "move to" this location independently of the starting coordinates of the eyes. In a way, it is much like the reaching for an occluded stimulus, which tells the body to "reach here" independently of posture or current body location.
What is particularly fascinating about the alignment of the body map to the visual map, is it occurs from the somatic sensory cortex, not the motor cortex.
We know what the sensory map looks like, it has no relationship whatsoever to the visual field. This is what it looks like: