05_MEEG_StatsPractica

advertisement
Practical aspects of…
Statistical Analysis of M/EEG Sensor- and
Source-Level Data
Jason Taylor
MRC Cognition and Brain Sciences Unit (CBU)
Cambridge Centre for Ageing and Neuroscience (CamCAN)
Cambridge, UK
jason.taylor@mrc-cbu.cam.ac.uk | http://imaging.mrc-cbu.cam.ac.uk/meg (wiki)
14 May 2012 | London | Thanks to Rik Henson and Dan Wakeman @ CBU
Overview
The SPM approach to M/EEG Statistics:
A mass-univariate statistical approach to inference regarding
effects in space/time/frequency (using replications across
trials or subjects).
•
In Sensor Space:
2D Time-Frequency
3D Topography-by-Time
•
In Source Space:
3D Time-Frequency contrast images
4-ish-D space by (factorised) time
Alternative approaches: SnPM, PPM, etc.
Example Datasets
CTF Multimodal Face-Evoked Dataset | SPM8 Manual (Rik Henson)
CTF MEG/ActiveTwo EEG:
275 MEG
128 EEG
3T fMRI (with nulls)
1mm3 sMRI
2 sessions
Stimuli:
~160 face trials/session
~160 scrambled trials/session
Group:
N=12 subjects
(Henson et al, 2009 a,b,c)
Chapter 37 SPM8 Manual
Neuromag Faces Dataset | Dan Wakeman & Rik Henson @ CBU
Neuromag MEG & EEG Data:
102 Magnetometers
204 Planar Gradiometers
70 EEG
HEOG, VEOG, ECG
1mm3 sMRI
400600ms
8001000ms
1700ms
6 sessions
Stimuli:
faces & scrambled-face images
(480 trials total)
Group:
N=18
(Henson et al, 2011)
Biomag 2010 Award-Winning Dataset!
Neuromag Lexical Decision Dataset | Taylor & Henson @ CBU
Neuromag MEG Data:
102 Magnetometers
204 Planar Gradiometers
HEOG, VEOG, ECG
1mm3 sMRI
4 sessions
Stimuli:
words & pseudowords
presented visually
(480 trials total)
Group:
N=18
Taylor & Henson (submitted)
Neuromag Lexical Decision
The Multiple Comparison Problem
Where is the effect?: The curse of too much data
The Multiple Comparison Problem
•
The more comparisons we conduct, the more Type I errors (false
positives) we will make when the Null Hypothesis is true.
* Must consider Familywise (vs. per-comparison) Error Rate
•
Comparisons are often made implicitly, e.g., by viewing (“eyeballing”) data before selecting a time-window or set of channels for
statistical analysis. RFT correction depends on the search volume.
-> When is there
an effect in time
e.g., GFP (1D)?
time
The Multiple Comparison Problem
•
The more comparisons we conduct, the more Type I errors (false
positives) we will make when the Null Hypothesis is true.
* Must consider Familywise (vs. per-comparison) Error Rate
Comparisons are often made implicitly, e.g., by viewing (“eyeballing”) data before selecting a time-window or set of channels for
statistical analysis. RFT correction depends on the search volume.
-> When/at what frequency
is there an effect
an effect in time/frequency
(2D)?
frequency
•
time
The Multiple Comparison Problem
•
The more comparisons we conduct, the more Type I errors (false
positives) we will make when the Null Hypothesis is true.
* Must consider Familywise (vs. per-comparison) Error Rate
•
Comparisons are often made implicitly, e.g., by viewing (“eyeballing”) data before selecting a time-window or set of channels for
statistical analysis. RFT correction depends on the search volume.
x
-> When/where
is there an effect
in sensor-topography
space/time (3D)?
y
The Multiple Comparison Problem
•
The more comparisons we conduct, the more Type I errors (false
positives) we will make when the Null Hypothesis is true.
* Must consider Familywise (vs. per-comparison) Error Rate
•
Comparisons are often made implicitly, e.g., by viewing (“eyeballing”) data before selecting a time-window or set of channels for
statistical analysis. RFT correction depends on the search volume.
-> When/where
is there an effect
in source
space/time
x
(4-ish-D)?
z
y
The Multiple Comparison Problem
•
The more comparisons we conduct, the more Type I errors (false
positives) we will make when the Null Hypothesis is true.
* Must consider Familywise (vs. per-comparison) Error Rate
•
Comparisons are often made implicitly, e.g., by viewing (“eyeballing”) data before selecting a time-window or set of channels for
statistical analysis. RFT correction depends on the search volume.
Random Field Theory (RFT) is a method for correcting for multiple
statistical comparisons with N-dimensional spaces (for parametric
statistics, e.g., Z-, T-, F- statistics).
* Takes smoothness of images into account
Worsley Et Al (1996). Human Brain Mapping, 4:58-73
GLM: Condition Effects after removing variance due
to confounds
Each trial-type (6)
Confounds (4)
Each trial
W/in Subject
(1st-level) model
-> one image per trial
-> one covariate value
per trial
Also: Group Analysis
(2nd-level)
-> one image per subject
per condition
-> one covariate value per
subject per condition
beta_00* image volumes reflect
(adjusted) condition effects
Henson et al., 2008, NImage
Sensor-space analyses
Where is an effect in time-frequency space?
1 subject (1st-level analysis)
1 MEG channel
Faces > Scrambled
Faces
Morlet wavelet projection
Scrambled
1 t-x-f image per trial
Kilner et al., 2005, Nsci Letters
CTF Multimodal Faces
Where is an effect in time-frequency space?
18 subjects
1 channel
(2nd-level
analysis)
Faces > Scrambled
EEG
MEG (Mag)
6
6
Freq (Hz)
Freq (Hz)
90
90
100-220ms
8-18Hz
-500
+1000
Time (ms)
-500
+1000
Time (ms)
Neuromag Faces
Where is an effect in sensor-time space?
1 subject
ALL EEG
channels
3D
Topographyx-Time
Image
volume
(EEG)
time
Each trial,
Topographic
projection (2D) for
each time point (1D)
= topo-time image
volume (3D)
Faces vs
Scrambled
y
x
CTF Multimodal Faces
Where is an effect in sensor-time space?
Analysis over subjects (2nd Level)
NOTE for MEG: Complicated by variability in head-position
SOLUTION(?): Virtual transformation to same position (sessions, subjects)
- using Signal Space Separation (SSS; Taulu et al, 2005; Neuromag systems)
Without transformation to Device Space
With transformation to Device Space
Stats over 18 subjects, RMS of planar gradiometers
Improved: (i.e., more blobs, larger T values) w/ head-position transform
Other evidence: ICA ‘splitting’ with movement
Taylor & Henson (2008) Biomag
Neuromag Lexical Decision
Where is an effect in sensor-time space?
Analysis over subjects (2nd Level):
Magnetometers: Pseudowords vs Words PPM (P>.95 effect>0)
Effect Size
18
fT
-18
Taylor & Henson (submitted)
Neuromag Lexical Decision
Source-space analyses
Where is an effect in source space (3D)?
STEPS:
Estimate evoked/induced
energy (RMS) at each dipole for
a certain time-frequency window
of interest.
- e.g., 100-220ms, 8-18 Hz
- For each condition
(Faces, Scrambled)
- For each sensor type
OR fused modalities
Write data to 3D image
- in MNI space
- smooth along 2D surface
Smooth by 3D Gaussian
Submit to GLM
Henson et al., 2007, NImage
Where is an effect in source space (3D)?
RESULTS: Faces > Scrambled
fMRI
EEG
MEG
sensor fusion
E+MEG
Henson et al, 2011
Neuromag Faces
Where is an effect in source space (3D)?
RESULTS: Faces > Scrambled
fMRI
with group optimised
fMRI priors
EEG+MEG+fMRI
with fMRI priors
EEG+MEG+fMRI
mean
stats
Henson et al, 2011
Neuromag Faces
Where and When do effects emerge/disappear
in source space (4-ish-D: time factorised)?
Condition x Time-window
Interactions
Factorising time allows
you to infer (rather than
simply describe) when
effects emerge or disappear.
We've used a data-driven
method (hierarchical
cluster analysis) in sensor
space to define timewindows of interest for
source-space analyses.
* high-dimensional distance
between topography vectors
* cut the cluster tree where the
solution is stable over longest
range of distances
Taylor & Henson, submitted
Neuromag Lexical Decision
Where and When do effects emerge/disappear
in source space (4-ish-D)?
Condition x Time-window
Interactions
Factorising time allows you
to infer (rather than simply
describe) when effects
emerge or disappear.
source
timecourses
We've used a data-driven
method (hierarchical cluster
analysis) in sensor space to
define time-windows of
interest for source-space
analyses.
* estimate sources in each
sub-time-window
* submit to GLM with conditions
& time-windows as factors
(here: Cond effects per t-win)
Taylor & Henson, submitted
Neuromag Lexical Decision
Where and When do effects emerge/disappear
in source space (4-ish-D)?
Condition x Time-window
Interactions
Factorising time allows you
to infer (rather than simply
describe) when effects
emerge or disappear.
We've used a data-driven
method (hierarchical cluster
analysis) in sensor space to
define time-windows of
interest for source-space
analyses.
* estimate sources in each subtime-window
* submit to GLM with conditions
& time-windows as factors:
Cond x TW interactions
Taylor & Henson, submitted
Neuromag Lexical Decision
Alternative Approaches
Alternative Approaches
Classical SPM approach: Caveats
Inverse operator induces long-range error correlations (e.g., similar
gain vectors from non-adjacent dipoles with similar orientation),
making RFT correction conservative
Distributions over subjects/voxels may not be Gaussian (e.g., if
sparse, as in MSP)
(Taylor & Henson, submitted; Biomag 2010)
Alternative Approaches
Non-Parametric Approach
(SnPM)
p<.05 FWE
Robust to non-Gaussian
distributions
Less conservative than
RFT when dfs<20
Caveats:
No idea of effect size (e.g.,
for power, future expts)
Exchangeability difficult for
more complex designs
(Taylor & Henson, Biomag 2010)
SnPM Toolbox by Holmes & Nichols:
http://go.warwick.ac.uk/tenichols/software/snpm/
CTF Multimodal Faces
Alternative Approaches
Posterior Probability
Maps (PPMs)
p>.95 (γ>1SD)
Bayesian Inference
No need for RFT (no MCP)
Threshold on posterior
probabilty of an effect
greater than some size
Can show effect size after
thresholding
Caveats:
Assume Gaussian
distribution (e.g., of mean
over voxels)
CTF Multimodal Faces
Future Directions
• Extend RFT to 2D cortical surfaces (+1D time)?
(e.g., Pantazis et al., 2005, NImage)
• Novel approaches to controlling for multiple
comparisons
(e.g., 'unique extrema' in sensors: Barnes et al., 2011, NImage)
• Go multivariate?
To identify linear combinations of spatial (sensor or
source) effects in time and space
(e.g., Carbonnell, 2004, NImage; but see Barnes et al., 2011)
To detect spatiotemporal patterns in 3D images
(e.g., Duzel, 2004, NImage; Kherif, 2003, NImage)
-- The end --
• Thanks for listening
• Acknowledgements:
• Rik Henson (MRC CBU)
• Dan Wakeman (MRC CBU / now at MGH)
• Vladimir, Karl, and the FIL Methods Group
• More info:
• http://imaging.mrc-cbu.cam.ac.uk/meg (wiki)
jason.taylor@mrc-cbu.cam.ac.uk
Download