Event Related Designs - Department of Psychology

advertisement
Jody Culham
Brain and Mind Institute
Department of Psychology
Western University
http://www.fmri4newbies.com/
Basics of Experimental Design
for fMRI:
Event-Related Designs
Last Update: February 22, 2013
Last Course: Psychology 9223, W2013, Western University
Event-Related Averaging
(can be used for block or event-related designs)
Event-Related Averaging
In this example an “event”
is the start of a block
In single-trial designs, an
event may be the start of a
single trial
First, we compute an event related average for the blue condition
• Define a time window before (2 volumes) and after (15 volumes)
the event
• Extract the time course for every event (here there are four events
in one run)
• Average the time courses across all the events
Event-Related Averaging
Second, we compute an event
related average for the gray
condition
Event-Related Averaging
Third, we can plot the average
ERA for the blue and gray
conditions on the same graph
Event-Related Averaging in BV
Define which subjects/runs to include
Set time window
Define which
conditions to
average (usually
exclude baseline)
We can tell BV where to put the
y=0 baseline. Here it’s the
average of the two circled data
points at x=0.
Determine how you want to define
the y-axis values, including zero
But what if the curves don’t have the
same starting point?
In the data shown, the
curves started at the same
level, as we expect they
should because both
conditions were always
preceded by a resting
baseline period
But what if the data looked
like this?
…or this?
Epoch-based averaging
FILE-BASED AVERAGING: zero baseline determined across all conditions (for 0 to 0: points in red circles)
In the latter two cases, we could simply shift the curves so they all start from the same (zero) baseline
EPOCH-BASED AVERAGING: zero baselines are specific to each epoch
File-based vs. Epoch-based Averaging
time courses may start at different
points because different event
histories or noise
e.g., set EACH curve
such that at time=0,
%BSC=0
0
0
0
File-based Averaging
Epoch-based Averaging
•
•
•
•
•
•
zero is based on average starting point
of all curves
works best when low frequencies have
been filtered out of your data
similar to what your GLM stats are
testing
•
each curve starts at zero
can be risky with noisy data
only use it if you are fairly certain
your pre-stim baselines are valid
(e.g., you have a long ITI and/or
your trial orders are
counterbalanced)
can yield very different conclusions
than GLM stats
What if…?
• This design has the benefit that each condition epoch is preceded
by a baseline, which is nice for making event-related averages
• However, we might decide that this design takes too much time
because we are spending over half of the time on the baseline.
• Perhaps we should use the following paradigm instead…?
• This regular triad sequence has some nice features, but it can
make ERAs more complicated to understand.
Regular Ordering and ERAs
• We might have a time course that looks like this
Example of ERA Problems
• If you make an ERA the usual way, you might get something that
looks like this:
File-Based (Pre=2, Post=10, baseline 0 to 0)
Intact
Scrambled
One common newbie
mistake is to make
ERAs for all
conditions, including
the baseline (Fixation).
This situation will
illustrate some of the
confusion with that
Fixation
• Initially some people can be confused how to interpret this ERA
because the pre-event activation looks wonky.
Example of ERA Problems
File-Based (Pre=2, Post=10, baseline 0 to 0)
File-Based (Pre=8, Post=18, baseline 0 to 0)
Intact
Scrambled
Fixation
• If you make the ERA over a longer
time window, the situation becomes
clearer.
• You have three curves that are merely
shifted in time with respect to one
another.
Example of ERA Problems
File-Based (Pre=2, Post=10, baseline 0 to 0)
Intact
End of
Intact
Scrambled
End of
Scrambled
End of
Fixation
Fixation
• Now you should realize that the different pre-epoch baselines result
from the fact that each condition has different preceding conditions
– Intact is always preceded by Fixation
– Scrambled is always preceded by Intact
– Fixation is always preceded by Scrambled
Example of ERA Problems
File-Based (Pre=2, Post=10, baseline 0 to 0)
Intact
Scrambled
Fixation
• Because of the different histories, changes with respect to baseline are
hard to interpret. Nevertheless, ERAs can show you how much the
conditions differed once the BOLD response stabilized
– This period shows, rightly so, Intact > Scrambled > Fixation
Example of ERA Problems
Epoch-Based (Pre=2, Post=10, baseline -2 to -2)
• Because the pre-epoch baselines are so different (due to differences in
preceding conditions), here it would be really stupid to do epoch-based
averaging (e.g., with x=-2 as the y=0 baseline)
• In fact, it would lead us to conclude (falsely!) that there was more
activation for Fixation than for Scrambled
Example of ERA Problems
• In a situation with a regular sequence like this, instead of making an
ERA with a short time window and curves for all conditions, you can
make one single time window long enough to show the series of
conditions (and here you can also pick a sensible y= 0 based on x=-2)
File-Based average for Intact condition only (Pre=2, Post23, baseline -2 to -2)
Intact
Scrambled
Fixation
Partial confounding
• In the case we just considered, the histories for various
conditions were completely confounded
– Intact was always preceded by Fixation
– Scrambled was always preceded by Intact
– Fixation was always preceded by Scrambled
• We can also run into problems (less obvious but with the same ERA
issues) if the histories of conditions are partially confounded (e.g.,
quasi-random orders)
• Intact is preceded by Scrambled 3X and by Fixation 3X
• Scrambled is preceded by Intact 4X and Fixation 1X
• Fixation is preceded by Intact 2X, by Scrambled 2X and
by nothing 1X
• No condition is ever preceded by itself
ERAs: Take-home messages
• ERAs can be a valuable way to look at activation
over time
• BUT
– you have to carefully select the baseline (file-based vs.
epoch-based; which time points to take as baseline)
– you have to know whether the history of your conditions is
counterbalanced
• if it’s not counterbalanced, ERAs can be misleading
• This problem of “history” will come up again…
Basics of Event-Related Designs
Block Designs
= trial of one type
(e.g., face image)
= trial of another type
(e.g., place image)
Block
Design
Early Assumption: Because the hemodynamic response delays and blurs
the response to activation, the temporal resolution of fMRI is limited.
Positive BOLD response
BOLD Response
(% signal change)
WRONG!!!!!
3
2
Overshoot
1
Initial
Dip
0
Post-stimulus
Undershoot
Time
Stimulus
What are the temporal limits?
What is the briefest stimulus that fMRI can detect?
Blamire et al. (1992): 2 sec
Bandettini (1993): 0.5 sec
Savoy et al (1995): 34 msec
2 s stimuli
single events
Data: Blamire et al., 1992, PNAS
Figure: Huettel, Song & McCarthy, 2004
Data: Robert Savoy & Kathy O’Craven
Figure: Rosen et al., 1998, PNAS
Although the shape of the HRF delayed and blurred, it is predictable.
Event-related potentials (ERPs) are based on averaging small
responses over many trials.
Can we do the same thing with fMRI?
Predictor Height Depends on Stimulus
Duration
= trial of one type
(e.g., face image)
= trial of another type
(e.g., place image)
Block
Design
Slow ER
Design
Rapid
Jittered ER
Design
Mixed
Design
Design Types
Detection vs. Estimation
% Signal Change
• detection: determination of whether
activity of a given voxel (or region)
changes in response to the
experimental manipulation
1
0
0
4
8
12
• estimation: measurement of the time
course within an active voxel in
response to the experimental
manipulation
Time (sec)
Definitions modified from: Huettel, Song & McCarthy, 2004,
Functional Magnetic Resonance Imaging
Block Designs: Poor Estimation
Huettel, Song & McCarthy, 2004, Functional Magnetic Resonance Imaging
Pros & Cons of Block Designs
Pros
• high detection power
• has been the most widely used approach for fMRI studies
• accurate estimation of hemodynamic response function is not as
critical as with event-related designs
Cons
• poor estimation power
• subjects get into a mental set for a block
• very predictable for subject
• can’t look at effects of single events (e.g., correct vs. incorrect
trials, remembered vs. forgotten items)
• becomes unmanagable with too many conditions (e.g., more
than 4 conditions + baseline)
Slow Event-Related Designs
Slow ER
Design
Convolution of Single Trials
Neuronal Activity
BOLD Signal
Haemodynamic Function
Time
Time
Slide from Matt Brown
BOLD Summates
Neuronal Activity
Slide adapted from Matt Brown
BOLD Signal
Slow Event-Related Design: Constant ITI
Block
2 s stim
vary ISI
Bandettini et al. (2000)
What is the optimal trial spacing (duration +
intertrial interval, ITI) for a Spaced Mixed
Trial design with constant stimulus
duration?
Event-related average
Source: Bandettini et al., 2000
Optimal Constant ITI
Source: Bandettini et al., 2000
Brief (< 2 sec) stimuli:
optimal trial spacing = 12 sec
For longer stimuli:
optimal trial spacing = 8 + 2*stimulus duration
Effective loss in power of event related design:
= -35%
i.e., for 6 minutes of block design, run ~9 min ER design
Trial to Trial Variability
Huettel, Song & McCarthy, 2004,
Functional Magnetic Resonance Imaging
How Many Trials Do You Need?
Huettel, Song & McCarthy, 2004, Functional Magnetic Resonance Imaging
•
•
•
standard error of the mean varies with square root of number of trials
Number of trials needed will vary with effect size
Function begins to asymptote around 15 trials
Effect of Adding Trials
Huettel, Song & McCarthy, 2004, Functional Magnetic Resonance Imaging
Pros & Cons of Slow ER Designs
Pros
• excellent estimation
• useful for studies with delay
periods
• very useful for designs with
motion artifacts (grasping,
swallowing, speech)
because you can tease out
artifacts
• analysis is straightforward
Example: Delayed Hand Actions
(Singhal et al., under revision)
Visual Delay
Action
Response
Execution
Grasp Go (G)
Reach Go (R)
Grasp Stop (GS)
Reach Stop (RS)
Actionrelated
artifact
Really long delay:
18 s
Cons
• poor detection power because you get very few trials
per condition by spending most of your sampling
power on estimating the baseline
• subjects can get VERY bored and sleepy with long
inter-trial intervals
Effect of this design
on our subject
“Do You Wanna Go Faster?”
Rapid
Jittered ER
Design
• Yes, but we have to test assumptions regarding
linearity of BOLD signal first
Linearity of BOLD response
Linearity:
“Do things add up?”
red = 2 - 1
green = 3 - 2
Sync each trial response
to start of trial
Not quite linear
but good enough!
Source: Dale & Buckner, 1997
Linearity is okay for events every ~4+ s
Why isn’t BOLD totally linear?
In part because neurons aren’t totally linear either
•“Phasic” (or “transient”) neural responses
•Adaptation or habituation… stay tuned
Spikes/ms
– May depend on factors like stimulus duration and stimulus intensity
Time (ms)
Ganmor et al., 2010, Neuron
Optimal Rapid ITI
Source: Dale & Buckner, 1997
Rapid Mixed Trial Designs
Short ITIs (~2 sec) are best for detection power
Do you know why?
Efficiency (Power)
Two Approaches
• Detection – find the blobs
– Business as usual
– Model predicted activation using square-wave predictor
functions convolved with assumed HRF
– Extract beta weights for each condition; Contrast betas
– Drawback: Because trials are packed so closely together,
any misestimates of the HRF will lead to imperfect GLM
predictors and betas
• Estimation – find the time course
– make a model that can estimate the volume-by-volume time
courses through a deconvolution of the signal
BOLD Overlap With Regular Trial Spacing
Neuronal activity from TWO event types with constant ITI
Partial tetanus BOLD activity from two event types
Slide from Matt Brown
BOLD Overlap with Jittering
Neuronal activity from closely-spaced, jittered events
BOLD activity from closely-spaced, jittered events
Slide from Matt Brown
BOLD Overlap with Jittering
Neuronal activity from closely-spaced, jittered events
BOLD activity from closely-spaced, jittered events
Slide from Matt Brown
Fast fMRI Detection
A) BOLD Signal
B) Individual Haemodynamic Components
C) 2 Predictor Curves for use with GLM (summation of B)
Slide from Matt Brown
Why jitter?
• Yields larger fluctuations in signal
When pink is on, yellow is off
 pink and yellow are anticorrelated
Includes cases when both pink and yellow are off
 less anticorrelation
• Without jittering predictors from different trial types
are strongly anticorrelated
– As we know, the GLM doesn’t do so well when predictors
are correlated (or anticorrelated)
GLM: Tutorial data
• Just as in the GLM for a block design, we have one
predictor for each condition other than the baseline
GLM: Output
Faces > Baseline
How to Jitter
= trial of one type
(e.g., face image)
= trial of another type
(e.g., place image)
TD = 2 s
ITI = 0 s
SOA = 2 s
TD = 2 s
ITI = 4 s
SOA = 6 s
Vary Intertrial Interval (ITI)
•
•
Stimulus Onset Asynchrony (SOA) = ITI + Trial Duration
may want to make TD (e.g., 2 s) and ITI durations (e.g., 0, 2, 4, 6 s) an integer multiple
of TR (e.g., 2 s) for ease of creating protocol files
Flat Distribution
Frequency
of ITIs
in Each
Condition
Another way to think about it…
Exponential Distribution
Frequency
of ITIs
in Each
Condition
0 2 4 6
0 2 4 6
ITI (s)
ITI (s)
Include “Null” Trials
= null trial
(nothing happens)
•
•
Can randomize or counterbalance distribution of three trial types
Outcome may be similar to varying ISI
Assumption of HRF is More Problematic for
Event-Related Designs
We know that the standard two-gamma HRF
is a mediocre approximation for individual Ss’
HRFs
Handwerker et al., 2004, Neuroimage
We know this isn’t such a big deal for block
designs but it is a bigger issue for rapid
event-related designs.
One Approach to Estimation:
Counterbalanced Trial Orders
•
Each condition must have the same history for preceding trials so that trial
history subtracts out in comparisons
•
For example if you have a sequence of Face, Place and Object trials (e.g.,
FPFOPPOF…), with 30 trials for each condition, you could make sure that the
breakdown of trials (yellow) with respect to the preceding trial (blue) was as
follows:
•
•
•
…Face  Face x 10
…Place  Face x 10
…Object  Face x 10
•
•
•
…Face  Place x 10
…Place  Place x 10
…Object  Place x 10
•
•
•
…Face  Object x 10
…Place  Object x 10
…Object  Object x 10
•
Most counterbalancing algorithms do not control for trial history beyond the
preceding one or two items
Algorithms for Picking Efficient Designs
Optseq2
Algorithms for Picking Efficient Designs
Genetic Algorithms
You Can’t Always Counterbalance
You may be interested in variables for which you can not control trial
sequence
e.g., subject errors can mess up your counterbalancing
e.g., memory experiments: remembered vs. forgotten items
e.g., decision-making: choice 1 vs. choice 2
e.g., correlations with behavioral ratings
Post Hoc Trial Sorting Example
Wagner et al., 1998, Science
Pros & Cons of Applying Standard GLM
to Rapid-ER Designs
Pros
• high detection power
• trials can be put in unpredictable order
• subjects don’t get so bored
Cons and Caveats
• reduced detection compared to block designs
• requires stronger assumptions about linearity
– BOLD is non-linear with inter-event intervals < 6 sec.
– Nonlinearity becomes severe under 2 sec.
• errors in HRF model can introduce errors in activation estimates
Design Types
Mixed
Design
= trial of one type
(e.g., face image)
= trial of another type
(e.g., place image)
Example of Mixed Design
Otten, Henson, & Rugg, 2002, Nature Neuroscience
• used short task blocks in which subjects encoded words
into memory
• In some areas, mean level of activity for a block
predicted retrieval success
Pros and Cons of Mixed Designs
Pros
• allow researchers to distinguish between staterelated and item-related activation
Cons
• sensitive to errors in HRF modelling
Deconvolution of Event-Related Designs
Using the GLM
Two Approaches
• Detection – find the blobs
– Business as usual
– Model predicted activation using square-wave predictor
functions convolved with assumed HRF
– Extract beta weights for each condition; Contrast betas
– Drawback: Because trials are packed so closely together,
any misestimates of the HRF will lead to imperfect GLM
predictors and betas
• Estimation – find the time course
– make a model that can estimate the volume-by-volume time
courses through a deconvolution of the signal
Convolution of Single Trials
Neuronal Activity
BOLD Signal
Haemodynamic Function
Time
Time
Slide from Matt Brown
Fast fMRI Detection
A) BOLD Signal
B) Individual Haemodynamic Components
C) 2 Predictor Curves for use with GLM (summation of B)
Slide from Matt Brown
DEconvolution of Single Trials
Neuronal Activity
BOLD Signal
Haemodynamic Function
Time
Time
Slide from Matt Brown
Deconvolution Example
• time course from 4 trials of two types (pink, blue) in a “jittered” design
Summed Activation
Single Stick Predictor
(stick predictors are also called finite impulse response (FIR) functions)
• single predictor for first volume of pink trial type
Predictors for Pink Trial Type
• set of 12 predictors for subsequent volumes of pink trial type
• need enough predictors to cover unfolding of HRF (depends on TR)
Predictor Matrix
• Diagonal filled with 1’s
.
.
.
Predictors for Pink Trial Type
Predictors for the Blue Trial Type
Predictor x Beta Weights for Pink Trial Type
• sequence of beta weights for one trial type yields an estimate of
the average activation (including HRF)
Predictor x Beta Weights for Blue Trial Type
• height of beta weights indicates amplitude of response (higher
betas = larger response)
Linear Deconvolution
Miezen et al.
2000
•
Jittering ITI also preserves linear independence among the hemodynamic
components comprising the BOLD signal.
Decon GLM
To find areas
that respond to
all stims, we
could fill the
contrast column
with +’s
…but that would
be kind of dumb
because we
don’t expect all
time points to be
highly positive,
just the ones at
the peak of the
HRF
14 predictors (time points)
for Cues
14 predictors (time points)
for Face trials
14 predictors (time points)
for House trials
14 predictors (time points)
for Object trials
Contrasts on Peak Volumes
We can search
for areas that
show activation
at the peak (e.g.,
3-5 s after
stimulus onset
Results: Peaks > Baseline
Graph beta weights
for spike predictors 
Get deconvolution
time course
Why go to all this
bother? Why not just
generate an eventrelated average?
…
Pros and Cons of Deconvolution
Pros:
• Produces time course that dissociates activation from trial history
• Does not assume specific shape for hemodynamic function
• Robust against trial history biases (though not immune to it)
• Compound trial types possible (e.g., stimulus-delay-response)
– may wish to include “half-trials” (stimulus without response)
Cons:
• Complicated
• Quite sensitive to noise
• Contrasts don’t take HRF fully into account, they just examine peaks
Not Mutually Exclusive
• Convolution and deconvolution GLMs are not
mutually exclusive
• Example
– use convolution GLM to detect blobs, use deconvolution to
estimate time courses
= trial of one type
(e.g., face image)
= trial of another type
(e.g., place image)
Block
Design
Slow ER
Design
Rapid
Jittered ER
Design
Mixed
Design
Design Types
Take-home message
• Block designs
– Great detection, poor estimation
• Slow ER designs
– Poor detection, great estimation
• Fast ER designs
– Good detection, very good estimation
– Excellent choice for designs where predictability is
undesirable or where you want to factor in subject’s behavior
Download