Object Processing Visual agnosia is marked by elementary visual

advertisement
Object Processing
Visual agnosia is marked by elementary visual capacity (acuity, visual field)
intactness, but severe perceptual deficits.
VA can be broken into apperceptive (early stage: elementary shape perception) and
associative (shape intact, but can’t associate percept with meaning / name /
pantomime use. must recognize by sound, touch, taste, questions).
Apperceptive patients can’t tell objects apart or shapes. Can’t draw contours. can’t
copy shapes. Also cannot disentangle contours for overlapping objects.
Associative are still good at copying drawings, but not good at drawing from
emmory. printing flaws of the same object cause these patients to think they are
different objects. Totally summed up by people not knowing what is on their plate
until they taste it.
The Kluver-Bucy syndrome is associated with hyper-orality where animals have a
visual agnosia and need to interpret objects with mouth.
Lesions to Inferotemporal cortex causes an inability to distinguish different objects,
a lack of retinament in previously acquired visual discriminations.
IT usually spans across viewing conditions (view-independent-invariance), so if you
lose a concept of an object, you lose all viewpoints of it. This is also true for changes
in illumination.
IT lesioned monkeys can’t transfer visual discirimination from half visual space to
the other.
Dorsal stream for “where”. ventral stream for “what”. IT lesions affect
discrimination but not visuo-spatial tasks like reaching or visospatial judement.
Parietal lesions do the opposite. Ventral stream responds to features. Dorsal
respond to spatial aspects like direction and speed.
As we move from V1 to higher order regions, the receptive fields get larger, like in
TE. There is no retinotopy in TE.
When subjects attend to particular aspects of visual stimuli, PET scans show
preferential increases in blood flow suring this selective attention (dorsal for
motion) ventral for shape, color. This is evidence of top-down modulation. There is
more evidence where TMS over V5 induces moving phosphenes and over V1 creates
static.
Neurons in macaque area V4 aquire directional tuning after adaptation to motion
stimuli. V4-V5 interactions represent modulation between areas at similar level of
visual hierarchy. They also represent cases of cross-talk between ventral and dorsal
streams. V4 neirions are not normally motion selective but can become so due to an
adaptation in V5 neurions that changes V4 neuronal properties.
Connections of these streams with hippocampus (via EC) and amygdala make things
more valent and memorable.
As you move further along the ventral stream, there is an increase in invariance.
There is also a grandmother cell hypothesis where every object is coded in a single
cell in IT, but the population code hypothesis seems more fit where each specific
object activates patterns in multiple cells.
Sparse coding shows that a few cells might code for very selective stimuli (Jennifer
Anniston).
Cells in TE that respond to common visual features are grouped in columns running
perpendicular to cortical surface. Columnar organization is reminiscent of V1.
Temporal neurons show selective responses to biological stimuli (faces – FFA).
some respond to gaze direction, hand-object interactions, biological motion
(walking), and point light displays (lights at joints like making CGI).
Viewpoint dependence in object recognition: object centered is that recognition is
by components and objects can be defined by a small set of primitive shapes
(geons). viewer-centered has a key factor where its objects need to be seen as they
originally were presented. A small set of prototypical views would suffice due to
mental rotation.
Object recognition and experience can be seen under repetition suppression and
mnemonic enhancement. More neurons will be recruited for a trained object.
repetition suppression can happen under anesthesia and unattended stimuli.
Cutting the line between TE and hippocampus reduces pair coding.
TE neurons extract invariant aspects of objects, preferentially respond to
biologically relevant stimuli, and they learn (increased frequency of selective cells
after learning and pair coding.)
Dorsal and ventral double dissociations extend to imagery as well.
Lateral occipital complex is concerned with object processing (actual vs. scrambled
objects shows preferred activation to non-scrambled) . Object sizes don’t matter. All
that matters is object like. LOC is intermediate between V1-5 and TE/IT.
There is also an extra-striate body area that responds to body parts, PPA, STS-FA is
also responsive to the changeable aspect of faces (smiling). FFA more lateral than
PPA.
N200 wave forms show shape selectivity that is strongly induced by faces but not by
changes in low level visual features.
Prosopagnosia – can’t recognize faces. can classify objects or faces from other
objects, but not between faces. Could be due to lack of expertise.
Haxby studied showed that we can classify faces even when the FFA is removed,
which is a great case for distributed coding.
Representational Dissimilarity Matrices that show animate and inanimate
categories in IT as having the most dissimilarity in humans and monkeys alike. Also
there is an unsupervised multidimensional scaling approach that groups neuronal
responses to objects into bodies, faces, natural, and artificial objects. This
dissociation does not extend to V1 and the dissimilarity remains even if highest
BOLD responses are removed….more evidence for distributed coding.
pictures of tools will activate not only visual areas, but also motor areas necessary
for grasping in parietal and premotor. About 50% of grasping motor neurons in
ventral premotor cortex also have visual properties. They fire at the site of
graspable objects. They are called grasping visuo-motor neurons or canonical
neurons.
AIP (anetior intraparietal ariea) in the intraparietal sulcus shows motor and visual
selectivitiy in a mirror neuron fashion.
Space Processing
PPC known for integrating perception and action. Broken up into SPL and IPL.
SPL lesions cause optic ataxia. IPL causes hemispatial neglect.
Simultagnosia is the inability to attend to multiple objects simultaneously.
Ocular apraxia is the inability to move eyes to new target locations.
Optic ataxia is basically the inability to use visual information to guide reaching
movements.
In hemispatial neglect (since its usually seen with lesions to the right IPL) the
patients neglect the left side and “squeeze” all parts of a scene into the right side or
ignore the left all together. Bisecting a line shows a serious rightward bias. In a
whole page of lines, with instructions to cross the lines, they will only cross out one
half.
Its thought that right hemisphere attends to both hemispheres, wheras the left
attends only to contralateral space. Imaging supports this.
Hemispatial neglect can be object-centered where they can attend to both sides of
space, but the frame of reference is the object. They will draw only the right side of
each object if two are presented in both visual fields.
If a monkey makes an identical saccade, but it is either to the right or left of an
object, then neuronal firings will be diminished when its to the right of the bar.
Particularly in the supplementary eye field.
in real life scenarios, this is witnessed as well (Duomo study). If patients stand and
describe a scene, they will neglect the left. If they are asked to imagine standing the
opposite way, they will report details about things they omitted before. This is
evident of implicit processing without conscious awareness.
Anosognosia neglect part of their body and can’t map their condition to their own
body.
with body schema problems, patients will deny that their contralateral limb is their
own. object sin hands are reported to be held by someone else until moved to
ipsilateral hand.
Patients shown red or green lights to L or R visual field and asked to name color.
They respond to right visual field, but not left. however, if response keys are on the
left side, patients can’t even detect the stimuli.
In the line cancellation task, if you provide a mirror so that patients can see both
sides of the paper, neglect patients cancel the left side of the page (perceptual
neglect). However, some patients, with motor neglect, will cancel the right side of
the page.
There is near and far with spatial neglect. Some patients can’t bisect lines in front of
them but can use a laser to bisect a line far away. Vice versa exist as well. This makes
a case for extrapersonal and peripersonal space.
Extrapersonal might be related to FEF lesions, since that region is associated with
focusing the eyes outside of peripersonal space. Ventral premotor lesions might lead
to peripersonal neglect because these concern arm and mouth movements. Thus LIP
and FEF are for extrapersonal and VIP and F4 are for peripersonal.
Thus, space maps are supported by fronto-parietal networks and damage to the SLF
(connects posterior parietal and frontal love) is related to patients with neglect.
Arcuate fasciculus is also apparently important (connects Superior Temporal cortex
with frontal lobe).
LIP neurons show strong responses at the onset of a visual stimulus in the RF. this is
modulated by attention. When still attending (working memory) to that RF, there is
activity. There is a dynamic remapping of visual memory trace
VIP neurons have RFs that are smaller around the mouth.
There are bimodal cells in ventral premotor area F4 that respond both to tactile
information and visual info of things approaching that sector of space around the
tactile receptive field. This moves with the body part, not the retina. The objects that
come close are typically 3D, graspable objects.
Extinction is when there is neglect only if there is a bilateral presentation of stimuli.
Cross-modal neglect is when a visual stimulus near the ipsilesional hand induces
tactile extinction of the contralesional hand. If the view of the hand is occluded, the
effect is reduced. The same effect us observed with a rubber hand.
Peripersonal space can be extended with the use of tools, like a rake. The visual
receptive field around the hand of the monkey will extend to the tip of the rake. You
can see inhumans with near space hemineglect that their near space extends if they
use a stick to bisect a line far away. But…if they use a laser pointer they can do it just
fine.
Stimulation of motor cortex can induce final posture regardless of initial. However,
the EMG profile is different depending on the start state. That means we are
stimulating intent and the intent gets completed regardless of starting position.
The velocity of these movements are the same as naturally spontaneous
movements.
You can induce defensive movements as well with VIP and F4 stimulation. Monkeys
will move in a manner that tries to protect the receptive field. Don’t just move the
body part with the RF, but also other parts to help block/avoid.
These movements can be evoked even under anesthesia.
There is a “hand-final-location” mapping where more dorsal representation for final
hand location is located more ventrally and vice versa. Models that try and figure
out the topographic continuity of actions ends up showing that the body motor map
in cortex is configured as such due to things necessary for coordinated movements.
hippocampal place cells are for particular locations in space and persist in darkness
and remap for new environments.
post subicular neurons are sensitive to head direction, even in dark, as long as
animal knows spatial orientation.
grid cells fire in regular hexagonal patterns across an environment. They can be
anchored to landmarks.
Attention
Covert orientation is attending to something without moving the eyes.
Slience drive exogenously driven attention.
Attending to a feature of an object tends to facilitate processing of other features of
the same objects.
Posner paradigm: spatial locations are cued with a peripheral cue or central cue(at
fixation with an arrow). Valid trials are when the actual stimulus goes to clued. RT
are faster for valid trials. The validity effect. RT gets penalized linearly with distance
of actual stimulus from cued location. This is the cost associated with switching the
direction of the covertly planned saccade.
The premotor theory of attention is that enhancement achieved by planning
(without executing) a saccade. This was tested by showing that attention shift a
saccade’s trajectory: saccades have trajectories that are not straight, with an initial
trajectory in the opposite direction of the cued location.
Both covert and overt attention recruit eye fields necessary for saccades (FEF, SFS,,
LIP, IPS, etc.). BOLD maps are overlapping.
However, peripheral cues can slow RT if the delay between cue and stimulus is long
enough(SOA: stimulus onset asynchrony). Known as Inhibition of Return, its
thought that this improves visual search by preventing return to previously
searched locations. Gaze cues don’t produce this IOR. IOR is thought to involve
cerebellum and superior colliculus since they involve peripheral cues (the only
situation where you observe this effect). These are olderstructures (and since IOR is
an old philosophy) it might follow that these regions evolved with the visual search
efficiency algorithms.
Gaze cues are important as well (looking at someone orient their eyes in a particular
direction). They are very powerful. They orient attention in an automatic manner.
Even when the gaze is not predictive of the target. Even when they are counter
predictive they will get faster RT to the gazed location.
FEF, SEF, and LIP(sometimes called PEF) are activated by both gaze and peripheral
cues.
Extrastriate areas are more active for gaze cues as compared to peripheral cues.
LIP neurons respond when stimulus is flashed onto its RF. However, does not
respond when a stimulus stable in the scene appears in the RF following a saccade.
Firing is high until right after saccade is executed. It does respond if the stimulus is
made salient by flashing it on and off, though. This suggests that LIP neurons
integrate bottom up and top down factors that modulate attention.
LIP neurons attend to task relevant or conscipuous objects by coding a salience
representation/priority map that specifies only a small number of stimuli.
Inactivating LIP or FEF produces deficits in target selection during search. Reward
expectations modulate this effect. However, they don’t change selectivity, but
instead more reliable response selectivity.
LIP neuronal responses are modulated by visual, cognitive, motivational, and motor
factors. The motor demands necessary for a task that involves the detection of
oriented stimuli in an LIP’s RF will be modulated by which motor response needs to
be made (left or right hand).
Attention modulated visual acuity changes activity in V4 before stimulus
presentation and during stimulation. Does attention modulate visual processing by
amplifying neural signals for attended stimuli? V1 doesn’t show any modulation.
Attention effects aremore pronounced in higher level processing areas. Top down
signals through feedback connections. Attentional effects have a 150ms delay in TE
and 230 in V1.
Attention modulates ERP by enhancing waveform in the opposite hemisphere.
Attention reduces signal ambiguities (separates the tuning curves of responses). Can
increase ability to discriminate orientations if attending to the RF where stimuli are.
Thus, attention increases discriminability by reducing variability of responses
(reduces the width of the response curve). This effect is stronger in interneurons
than in pyramidal cells. Perhaps this means that the effect of attention it so suppress
signal from distractors? Attention also reduces the degree to which ow-frequency
fluctuations in the neuron’s spiking were correlated with fluctuations in the activity
of other neurons. Spatial attention decorrelates intrinsic activity fluctuations in V4.
Decoupling intrinsic activity might help us dissociate things.
Furthermore, a decrease in correlation increases signal to noise ratio far more than
an increase in firing rate, as the neuronal pool size increases past 10.
If you stimulate FEF and record from V4 you can increase neuronal response in V1
to a receptive field that would be active if the saccade was made that would have
been made if the FEF stimulation was above threshold. This shows that preparing to
make a saccade to a location is similar to covert attention.
efficient search is pop out. ineddicient requires conjuntions of elementary elements
of which there are multiple of each type. each item in inefficient search adds
50msec.
Feature integration theory (FIT) is where elementary features are bound into
coherent objects. Pop out doesn’t require this. In inefficient, we need to integrate.
FIT assumes that object completion happens after attentional search. However,
completions that hide popouts (where occlusion could mistake someone for a
target) shows that object completion precedes attentional search. The
computational limits of search are after features are integrated into wholes. This is
probably because the RF of higher order visual areas are limited. Neurons cannot
simultaneously send signals about all the stimuli inside the RF.
Firing rate of infero-temporal neurons is increased more for preferred stimuli than
non-preferred.
If a stimulus is preferred and presented non competitively (isolated) or
simultaneously, there is no difference in V1, but there is in V4 where there is greater
activity than V1 (Bigger RFs, bigger effects).
Suppressive interactions among multiple stimuli are eliminated in extrastriate
cortex when they are presented in the context of pop-out displays, in which a single
item differs from the others, but not in hereogenous displays, in which all items
differ from eachother. This effect is exaggerated as you move up the visual stream.
Basically, there is a suppression when images are viewed in heterogeneous display
as opposed to in isolation. But, if the images have a pop-out, there is no relative
suppression.
Mirror Neurons
AIP-F5 are for grasping.
Mirror neuron system is generally PF, PFG-F5
Grasping neurons code for the goal, not specific movements. This changes whether
the intended grip is precision or whole hand grips.
There are excitatory mirror neurons where the cell fires when monkey grasps and
also watches experimenter grasp. There are also inhibitory where the cell stops
firing at the sight of the experimenter grasping and when the monkey grasps. The
firing rate, however, is different (perhaps preventing unwanted action). These cells
also fire when the monkey grasps in darkness (not strictly visual). These mirror
neurons still fire even if the intent of the grasping is clear, but not seen (due to a
screen put up to cover an object). The neurons won’t fire if the grasp was made in
vain (pantomime). Prior knowledge drives this.
Audiovisual mirror neurons will respond to the sounds of things that required
action/grasping. A common code for sender and receiver?
strictly congruent mirror neurons mirror the exact action. Broadly congruent mirror
exact actions and similar actions.
F5 neurons code for hand to mouth actions. STS actually codes a bit for vision of
observed actions (not motor). Could plausibly send visual info to F5. PFG connects
with both F5 and STS and could relay the signals of STS to F5.
Intention to eat are more powerful than just picking up and even picking up and
placing by the mouth.
Pliers study (reverse and normal grip) shows that n spite of completely different
movements, the cells fire in both situations suggesting that the coding is for an
action goal, rather than the finger movements.
Mirror neurons can be modulated by actions performed in peripersonal vs.
extrapersonal space. This is relevant for the potential interactions between observer
and observed action. However, this is more so for plausibility of interaction. If a
shield is put up where it removes the action from the monkey’s workspace (but still
would have been in peripersonal space), then you get responses similar to
extrapersonal space.
Piramidal tract mirror neurons can show excitatory reponses during action, but
inhibitory during observation which might be important for the inhibition of
movement during action observation.
We also see LIP mirror neurons for when a monkey observes another monkey
making a gaze in that neurons preferred direction.
There are also VIP mirror neurons for defensive movements.
Songbirds also show mirror neurons. identical patterns when listening to certain
note sequences as when producing them. Even disrupting auditory feedback during
singing does not alter the activity.
mirror neurons in humans have been found in SMA and MTL (for memory of making
the action when observing it)
humans usually have more prolonged responses (maybe a more complex neural
signature).
LIP mirrors eye and attention.
Dorsal PMC and M1 for reaching
VIP for peripersonal space
SMA movement initiation
MTL memory mechanisms
Watching grasping actions increases MEPs in muscles. This is also seen with
increases in contralateral premotor cortex. (just one synapse away from muscle).
Mirroring also suppresses central rhythms (beta and alpha) just like what happens
during execution.
fMRI shows overlapping activation for observed and executed. We also see
repetition suppression.
mirror neurons in ventral premotor/ posterior ifg (monkey F5) and inferior parietal
cortex (monkey PF) and superior temporal sulcus (for biological motion and action
observation).
Cross-modal repetition suppression is debated based on the model proposed earlier.
Mirror neurons do not adapt in response to observation of repeated actions.
Do we perceive speech by retrieving the motor plan necessary to emit the speech we
are listening to? Mirror neurons also appear to be important for social learning.
Empathy might interact with the core imitation network.
Mirror neurons might simulate the facial expression of seeing someone smiling,
without actually making the person smile…then pass info to insula then pass to
limbic to then feel the emotion….empathy.
activity of motor neurons in inferior frontal cortex, insula and amygdala correlates
with subject’s tendencies to empathize and their social competence.
Autism…defecit in mirroring? less mirror areas active during imitation of facial
emotional expressions.
Higher recover in stroke patients after action observation.
hearing speech and making speech activates same areas at the border of primary
motor and premotor. TMSing this region shows defecits in speech perception (even
though it is far from auditory/wernicke’s).
functional connectivity during speech perception show connectivity between
superior temporal and that same premotor/motor region.
Model: superior temporal implements acoustic analysis while premotor implements
the mirrored phoneme production necessary to repeat. This production gets
compared in superior temporal with the actual hearing. The error signal (if one)
would be sent back to premotor which would generate a corrected phoneme to be
used for categorization.
Download