MEG Correlates of Visual Perception: The M170 Response to Degraded Face and Building Images jMASSACUSE1fiINSTITUTE by (y o TEiCHNCLOGY Valerie S. Morash NOV 13 2008 S.B., E.E. M.I.T., 2007 Submitted to the Department of Electrical Engineering and Computer Science in Partial Fulfillment of the Requirements for the Degree of Master of Engineering in Electrical Engineering and Computer Science at the Massachusetts Institute of Technology May 2008 Copyright 2008 M.I.T. All rights reserved. Author Depirtment of Electrical Engineering and Computer Science May 08, 2008 - Certified by_______________ C Pawan linha,Associate *r&" Professor Supervisor Accepted by Arthur C. Smith Professor of Electrical Engineering Chairman, Department Committee on Graduate Theses BARKER 1 2 Table of Contents M otivation for face research ............................................................................................... 5 Knowing how the brain processes faces is valuable to scientists, clinicians, and industry........5 The MI 70 represents at least one step in the face processing neural mechanism..................6 Background on M EG................................................................................................................8 M EG measures the magnetic fields associated with neural activity.......................................8 There are two types of M EG sensors: magnetometers and gradiometers...............................8 Interpretation of M EG data depends on the sensor type...................................................... 9 Magnetic field patterns and ECDs provide information about which brain areas are active.... 10 M EG is a fine method for fusiform gyrus research ................................................................ Background on the M 170 .................................................................................................... 11 12 The Ml 70 is reliably over occipitotemporal cortex about 170 ms after image presentation.... 12 The M l 70 is face-specific and face-generalizing............................................................... 13 The M 170 can be modulated by stimulus changes............................................................. 14 The M170 may or may not be related to the N170 and other face-specific neural signals....... 15 Pro sopagnosics do not always have or lack an M170, the elderly have a reduced M170 ........ 16 The M 170 reflects a beginning stage of face identification................................................ 17 Experiment..............................................................................................................................18 Intro du ction .......................................................................................................................... 18 M eth o d s................................................................................................................................2 0 Data Co llection ................................................................................................................. 20 MEG Data Noise Reduction ......................................................................................... 22 Selecting SOIs: Sensors that displayedan M 1 70 .......................................................... 23 Comparingthe M 1 70 Sensor TopographiesforBuildings and Faces.............................. 24 Examining the M1 70 Amplitude and Latency Across Conditions .................................... 25 R esu lts .................................................................................................................................. 27 Comparingthe M l 70 Generatorfor Buildings and Faces ............................................. 27 Examining the M l 70 Amplitude andLatency Across Conditions .................................... 28 D iscussio n ............................................................................................................................ 32 Contributions .......................................................................................................................... 34 References ............................................................................................................................... 35 3 4 Motivation for face research Knowing how the brain processes faces is valuable to scientists, clinicians,and industry How does seeing a face eventually lead to recognition: e.g. "that's Jill." Many reports (including this one) are interested in this question, due to its intrinsic scientific appeal and application-based benefits. From a science perspective, there is a natural desire to know how the brain does something it is good at: recognize faces. Humans are extremely adept at discriminating between faces, and there is evidence for specialized face-processing brain machinery (Kanwisher 2000). Additionally, knowing how the brain recognizes faces provides insight into how the brain generally works. It lays down some of the operations the brain is able and likely to perform. From a clinical perspective, knowing the steps the brain takes to recognize faces provides an opportunity to ask, at which step do things go wrong for people with face processing disorders, such as prosopagnosia and autism? Pinpointing a disorder's defective step illuminates the extent of the disorder, and provides a straightforward test for diagnosis (this step is fine, but the next step is affected). If this test can be done with non-invasive measurement on young children, it will provide an early diagnosis leading to preemptive treatment. From an industry perspective, computerized face identification systems are wanted for airports and casinos to catch unwanted trespassers. Computer scientists have been working diligently to produce such a product, but the current systems perform far worse than normal people (Willing 2003; Zhao et al. 2003). A possible solution is to create a computer system that copies the 5 human brain. Theoretically, a computer using the same steps a brain uses to recognize faces will be just as good as a person. This provides strong motivation for discovering the "neural algorithm" underlying face processing. The Ml 70 represents at least one step in the face processing neural mechanism Previous research has uncovered the basic route through the brain subserving face recognition (for a review see Haxby et al. 2002). Major structures have been identified, and their functions have been proposed (Fig. 1). However, further refinement of the pathway is needed. Namely, the major structures need to be better characterized, and non-major structures need to be identified. Figure 1. Flow chart of the neural steps involved in face processing. Note the role and context of the lateral fusiform gyrus, the neural area responsible for the M170. Reproduced from Haxby et al. 2002. Intraparietal Sulcus Spatially-directed attention Superior Temporal Sulcus Changeable aspects of faces / Auditory Cortex Prelexical speech perception Perception of eye gaze, expression, lip movement I Inferior Occipital Gyri Early perception of facial features Lateral Fusiform Gyrus Invariant aspects of faces Amygdala, Insula, Limbic System Emotion processing, emotional response Perception of unique identity I Anterior Temporal Personal identity, name, biographical information I 6 I Researchers have established that one of the important brain areas for face recognition, the lateral fusiform gyrus (AKA the fusiform face area, FFA), produces a neural signal called the MI 70. The MI 70 is a signal released from the brain about 170 milliseconds after viewing an image, and is measured with magnetoencephalography (MEG). The remainder of this report will focus on what neural step the Ml 70 may reflect. It will include a comprehensive review of past M170 research (preceded by some important facts about MEG), and a new experiment addressing two pivotal questions: 1) Why does the M170 respond to buildings if it is face-selective? 2) Does the M170 reflect categorization (that's a face or not) or identification (that's Jill's face)? 7 Background on MEG Magnetoencephalographic principles and techniques can be both complex and obscure. This is unfortunate, since they must be considered when comparing M170 studies, and MEG to nonMEG results. In this section, I present some of the MEG underpinnings, which are necessary to comprehend M170 results. MEG measures the magneticfields associatedwith neural activity Fundamentally, MEG measures the magnetic fields surrounding weak net currents in the brain. These currents arise from the post-synapses of thousands of pyramidal neurons firing in concert (Himihiinen et al. 1993). MEG is more sensitive to currents parallel/tangential to the skull, than currents orthogonal/radial to the skull, due to external field cancellation by volume/return currents in the orthogonal case (Hmiliinen et al. 1993). This sensitivity is opposite to that for electroencephalography (EEG), which measures the electric fields associated with neural currents. Although MEG is blind to purely radial currents, it is probabilistically unlikely for this to occur in the brain. Therefore, a tangential component of most neural sources can be detected with MEG (Lueschow et al. 2004). There are two types of MEG sensors: magnetometers and gradiometers Both magnetometers and gradiometers make measurements of the magnetic field's radial component. Magnetometers measure the magnetic field strength, whereas gradiometers measure the derivative of the field (which can be used to approximate the field strength). The magnetic 8 field derivative falls off faster with distance than the field itself, making gradiometers less sensitive to external noise and other distant magnetic generators (Clarke and Braginski 2003). Therefore, gradiometers are often preferred for neural recordings, and magnetometers for ambient field correction or as detectors of deep neural activity (Tarkiainen et al. 2002). Unlike magnetometers, gradiometers come in many flavors, which differ in derivative order and reference direction. The gradiometer order refers to the specific field derivative the sensor detects (e.g. a second-order gradiometer is sensitive to the magnetic field's second derivative). Radial/axial gradiometers measure the derivative with respect to the radial direction, and planar/tangential gradiometers measure the derivative with respect to a direction orthogonal to radial, tangential to the skull (Clarke and Braginski 2003). Interpretationof MEG data depends on the sensor type With a magnometer and radial gradiometer, a neural current will be measured as a magnetic sink and source, flanking the area directly above the current. Only one sink/source may be detected if the other falls outside the sensor range. With a planar gradiometer, a neural current will be measured as a single sink or source directly above the current. For all types of sensors, the signal will be strongest closest to the current, and the pattern produced is referred to as a "dipole pattern." Magnetic sinks are commonly denoted with negative amplitudes, and sources with positive amplitudes. For magnetometers and radial gradiometers, but not planar gradiometers, equivalent currents in opposite hemispheres will produce magnetic fields with opposite polarity. For 9 example, a current in the right hemisphere may produce a magnetic sink where the equivalent current in the left hemisphere would produce a magnetic source (Sams et al. 1997; Harris et al. 2005). Magneticfield patterns and ECDs provide information about which brain areas are active Examining magnetic field patterns can reveal differences in field generator locations (which brain area created the dipole pattern). Comparing magnetic iso-contours, and maximum M170 sensor locations are both adequate for this purpose (Watanabe et al. 1999; Taylor et al. 2001). In addition to this method, researchers can "back solve" the neural current for each hemisphere, which they refer to as an equivalent current dipole (ECD). The ECD is the current most likely to have created the magnetic field pattern. This process collapses the data across multiple sensors, acting like a spatial filter to reduce noise. It also provides insight into which neural areas are responsible for the magnetic signals (Tarkiainen et al. 2002). Sometimes more than one dipole is solved per hemisphere, but more often than not, only one dipole is solved per dipole pattern, even though it is likely that more than one neural area's activity created that pattern (Himiliinen et al. 1993). When activity from several neural areas is modeled as a single dipole, the dipole represents those areas' center of gravity (Sams et al. 1997). The presence of a single dipole pattern is not evidence that only one neural site is active, because the simultaneous activity from several neural areas, even if they are relatively widespread, will typically produce such a pattern (Swithenby et al. 1998). Dipole localization can be worsened by using too few trials, artifacts, or poor subject task (Defike et al. 2007). 10 MEG is a fine method for fusiform gyrus research Using MEG to investigating the role of the fusiform gyrus has some distinct advantages. It is a good method for detecting fusiform activity, since the fusiform gyrus currents are mostly tangential. Additionally, MEG is sensitive to neural variation on the order of milliseconds. This is its greatest advantage over fMRI and PET, which have sensitivities on the order of seconds (Watanabe et al. 1999). EEG has a similar time resolution as MEG, but poorer spatial resolution. In favorable conditions, MEG can discriminate neural sources on the order of millimeters (Lu et al. 1991). Unfortunately, the skull and scalp tissues smear emerging electric fields, reducing EEG's spatial resolution to the order of centimeters (Sams et al. 1997). 11 Background on the M170 Lu et al first detected the face-specific M170 response in 1991 (face-specific refers to the signal's preferential, but not exclusive response to faces). The group was inspired by the facespecific neural potential (called the N 170), to search for a similar face-specific magnetic signal. While recording MEG, they showed subjects many black and white images of faces, birds, and simple drawings (a square and a circle). As is now convention in M170 studies, the MEG data were averaged across trials with the same image type, so that average signal characteristics could be examined. In all of their subjects, Lu et al. found a magnetic peak, about 150 ms after image presentation, in the region of the occipitotemporal junction in both hemispheres. The peak was larger to faces than birds, and larger to birds than simple drawings. It was not affected by stimulus duration or luminance; and was qualitatively similar across hemispheres (Lu et al. 1991). Since its discovery, the M170 has been examined in roughly 30 published reports. Principally, these reports have established the M170 response characteristics, relationship to other facespecific neural signals, and clinical relevance. Finally, there is a preliminary understanding of the M170's neural purpose. The M1 70 is reliably over occipitotemporalcortex about 170 ms after image presentation Overall, the M170 is extremely reliable in terms of its generation and face-specificity. In most studies, it is reported for all subjects, even in studies with many participants (Harris et al. 2005). However, as can happen with any neural signal, there are some subjects who lack an M170, and others whose M170 is not face-specific (Sams et al. 1997; Watanabe et al. 1999). 12 The M170 location is widely established to be over occipitotemporal cortex (Lu et al. 1991; Watanabe et al. 1999; Taylor et al. 2001; Tarkiainen et al. 2002; Lueschow et al. 2004). A large amount of ECD evidence suggests that it is generated within, or near the fusiform gyrus (Halgren et al. 1995; Sams et al. 1997; Swithenby et al. 1998; Sato et al. 1999; Kato et al. 2004; Tanskanen et al. 2005; Barrie et al. 2006; Deffke et al. 2007; Schweinberger et al. 2007). The neural area responsible for the MI 70 has a graded response to images, instead of an all-ornothing type response (Tanskanen et al. 2007). Some researchers find the M170 stronger and more reliable over the right hemisphere (Halgren et al. 1995; Marinkovic et al. 1995; Swithenby et al. 1998; Watanabe et al. 1999; Tarkiainen et al. 2002). However, others find it to be equal over both hemispheres, or stronger over the left hemisphere (Lu et al. 1991; Sams et al. 1997; Liu et al. 2000; Defike et al. 2007). The Ml 70 response to faces can peak within a wide time window. Typically, it occurs between 160-180 milliseconds after image presentation; but has been reported as early as 125 milliseconds and as late as 220 milliseconds (Schweinberger et al. 2007; Tanskanen et al. 2007). The Ml 70 is face-specific and face-generalizing Of all the M170 characteristics, most notable is its face-specificity. The M170 is consistently larger, and sometimes earlier to images of faces than plants, animals, face parts, body parts, objects, houses, simple shapes, and meaningless scrambled stimuli (Lu et al. 1991; Halgren et al. 1995; Marinkovic et al. 1995; Sams et al. 1997; Swithenby et al. 1998; Watanabe et al. 1999; 13 Halgren et al. 2000; Liu et al. 2000; Taylor et al. 2001; Liu et al. 2002; Harris et al. 2005; Xu et al. 2005; Barrie et al. 2006). This specificity is restricted to faces, and does not extend to images of expertise objects (Xu et al. 2005). The face M170 is smaller when it is preceded by other face or house images. This adaptation is more severe when the preceding image was a face, insensitive to whether the face was a photo, Mooney face, or face drawing (Harris and Nakayama 2007). There is extra adaptation when the preceding image was of the same face as the second, however this effect dissipates if the two faces were from different perspectives (Ewbank et al. 2008). The M170 is not only face-specific, but also face-generalizing. Human front-view faces, human profile faces, animal faces, cartoon faces, and face-like stimuli all elicit stronger M I70s than non-faces. Some have found face-like and face M170s to be indistinguishable, but others have found that human-face M170s are larger (Marinkovic et al. 1995; Halgren et al. 2000; Liu et al. 2000; Downing et al. 2001; Tarkiainen et al. 2002; Kato et al. 2004; Schweinberger et al. 2007). The M1 70 can be modulated by stimulus changes A large number of reports have been dedicated to what parameters do and do not affect the M170. The M170 is not affected by low-level image characteristics, such as luminance, face color, gender, or expression (Lu et al. 1991; Marinkovic et al. 1995). It is also not affected by stimulus duration (when longer than 100 milliseconds), attention, or by what task the subject is performing (Marinkovic et al. 1995; Swithenby et al. 1998; Streit et al. 1999; Lueschow et al. 2004). The M170 for faces is delayed and decreased by degradation and negation of the image 14 (Linkenkaer-Hansen et al. 1998; Tarkiainen et al. 2002; Tanskanen et al. 2005; Itier et al. 2006; Harris and Nakayama 2007). It is delayed and amplified when the image is inverted. (Linkenkaer-Hansen et al. 1998). The M170 amplitude can also be modulated by face gaze direction, and changes in stimulus duration (when under 150 milliseconds) (Taylor et al. 2001; Tanskanen et al. 2007). M170 changes related to image degradation probably reflect efficiency changes in the processes underlying the M170 (Liu et al. 2000). Contrastingly, other M170 differences, due to inversion, eye gaze direction, and non-face stimuli, may reflect separate neural processes. The MI 70 for these stimuli may be generated by a separate group of non-face-specific cells (LinkenkaerHansen et al. 1998; Swithenby et al. 1998; Watanabe et al. 1999; Liu et al. 2000; Taylor et al. 2001; Tarkiainen et al. 2002; Itier et al. 2006; Kloth et al. 2006; Schweinberger et al. 2007). For example, images of motor-bikes, letter strings, and hands have M I70s generated by different brain activity in/around the fusiform gyrus than what generates the face MI 70 (Swithenby et al. 1998; Watanabe et al. 1999; Tarkiainen et al. 2002). The M1 70 may or may not be related to the N] 70 and other face-specific neuralsignals Many researchers have proposed that the N 170, a face-specific neural potential, may be equivalent to the M170. Indeed, both are face-selective, and seem to be generated in or around the fusiform gyrus (Halgren et al. 1995; Zhao et al. 2003; Harris et al. 2005; Defike et al. 2007). However, they differ in their response to experimental factors, indicating that they do not reflect equivalent processes (Taylor et al. 2001). It's possible that the Ml 70 reflects only one of the two 15 sources underlying the N170, or the N200 (Liu et al. 2002). The fact that EEG and MEG have dissimilar orientation sensitivities complicates the discussion (Sams et al. 1997). Although the M170 is generated within the vicinity of the FFA, its response pattern is not identical to that of fMRI revealed FFA activity (Liu et al. 2000). Whereas the FFA hemodynamic response is affected by attention and expertise, the M170 is not (Lueschow et al. 2004; Xu et al. 2005; Furey et al. 2006). It is possible the same group of cells responsible for the M170 is activated later, contributing to the fMRI signal. However, the M170 activity itself does not contribute to the FFA hemodynamic response. Preceding the M170 is the M100, which is the earliest face-specific signal detectable with MEG (Liu et al. 2002). The M170 is followed by the M220, another face-specific magnetic peak (Itier et al. 2006). Coincident with the M170 is a magnetic peak over posterior brain areas. This signal may reflect the processing of "simple visual features" (Sams et al. 1997). It is not merely the sink/source associated with the M170's occipitotemporal activity (which can often occur in the same area, measured with radial gradiometers) because it is detected with planar gradiometers. I am skeptical of the M170-coincident posterior signal reported by Itier et al, because it looks like the posterior sink/source of the M170 detected with radial gradiometers (Itier et al. 2006). Prosopagnosicsdo not always have or lack an M1 70, the elderly have a reduced Ml 70 Not much research has focused on the clinical relevance of the M170. It appears that the signal is intact in some prosopagnosic subjects, but not in others. There is no correlation between 16 Ml 70 selectivity and prosopagnosia symptoms (Harris et al. 2005). The only other clinically relevant finding is an M170 delay in elderly subjects compared to young subjects (Nakamura et al. 2001). This may reflect an efficiency loss for visual processing. The M1 70 reflects a beginning stage of face identification Based on previous evidence, the MI 70 has a role in face processing. Its amplitude correlates with face classification and identification, modulated by image degradation and duration (Tanskanen et al. 2005; Tanskanen et al. 2007). An earlier MEG signal (the M100) is also correlated with face classification, implying that the MI 70 is beyond the image category decision (Tanskanen et al. 2005). The M170 for personally familiar faces is larger than that for unknown faces (Liu et al. 2002; Kloth et al. 2006). However, the M170 is equivalent for celebrity and non-celebrity faces, suggesting that it does not reflect face identification per se (Kloth et al. 2006; Ewbank et al. 2008). This evidence suggests that the MI 70 reflects a beginning stage of face identification. 17 Experiment Introduction This study focuses on the M170, and its association with face processing in the brain. It is specifically directed towards the following questions: 1) Why does the M170 respond to buildings if it is face-selective? 2) Does the M170 reflect categorization (that's a face or not) or identification (that's Jill's face)? This experiment tackles these questions in four parts (two for each question). First, I searched for topographic evidence that the building and face M170s were generated in similar/separate neural areas. The building M170 may come from non-face-specific cells in the fusiform gyrus, similar to other non-face M I70s (Swithenby et al. 1998; Watanabe et al. 1999; Tarkiainen et al. 2002). Alternatively, the same group of cells may generate building and face M170s. This would explain why building images cause adaptation of the face M170. Second, I investigated the possibility that the building M170 is a "baseline" response. If it were not baseline, the building M170 could reflect a group of cells specific to buildings. I tested these possibilities by measuring the M170 response to clear and degraded building images. M170 decay and delay from image degradation would imply specificity. Third, I looked for difference in celebrity and non-celebrity MI70s, when the images were clear and degraded. I was motivated by fact that the M170 is smaller and later for degraded faces, but generalizes to face-like images. These phenomena seem somewhat contrary. Perhaps, the M170 18 is smaller and later to degraded faces because identification, not face-likeness, is compromised. This could mean that the MI 70 reflects, "This face could be identified," but not face identification by itself Deciding if an image could be identified is a prime opportunity for face memory to chime in. If the image is degraded, but meets some template for a known face, the image would still be a good candidate for identification. Therefore, degradation may affect celebrity Ml 70s less than non-celebrity Ml 70s. Fourth, I compared how much image degradation delays face identification, to the amount it delays the Ml 70. It is likely that the face identification delay is accumulated, through many steps in the brain. If the MI 70 delay can entirely account for the face identification delay, the Ml 70 must be associated with a later face-processing step. Otherwise, it must represent a fairly early step. 19 Methods MEG was recorded while subjects viewed images of celebrity faces, non-celebrity faces, and buildings subjected to no, low, and high blur. The M170 sensor topographies were compared to reveal M170 generator differences. The M170 amplitudes and latencies were compared across conditions to assess the impact of image type and blur. Data Collection Data were collected at the MIT/KIT MEG joint research laboratory on the MIT campus. Informed consent was obtained from all subjects, and the MIT Committee on the Use of Humans as Experimental Subjects (COUHES) approved the study. MEG recordings were done using a 157 sensor whole-head system with first-order gradiometer sensors (Kanazawa Institute of Technology, Japan). An additional 3 magnetometers recorded local noise for reference. Subjects were supine in the MEG machine, which was inside an actively shielded room (Vacuumschmelze, Germany). Stimuli were projected onto a screen above the subject via a mirror using homemade Matlab scripts (Mathworks, Natick MA). MEG recordings were done using MEG160 (Yokogawa Electric Corporation and Eagle Technology Corporation, Japan). Subjects performed a one-back task while viewing images. They were instructed to push a button with their left index finger if an image was identical to the one immediately before it, which had a 5% probability of happening. Response trials were excluded from MEG analysis. 20 The stimuli included images of 135 buildings, 135 celebrity faces, and 135 non-celebrity faces. Each non-celebrity face was matched to a celebrity face based on race, age, hair, and general attractiveness. All images were gathered from the Internet. The image backgrounds were deleted, so that only the person or building was present; they were rescaled, so that the face or major part of the structure occupied a central oval; they were desaturated and placed on a white background; and they were reduced to a resolution of 150 pixels wide by 113 pixels tall. The images subtended roughly 12 0 x8" visual degrees when viewed by the subjects. The images were divided into three equal groups per category and subjected to blurring. One group received no blurring. The other groups received Gaussian blurring, with a sigma of one pixel (low blur) and three pixels (high blur). Therefore, there were nine conditions: no blur, low blur, and high blur images of celebrity faces, non-celebrity faces, and buildings. The images selected for each blur group were randomized across subjects. In a separate study, with the same images but different subjects and image size, these blur levels corresponded to near perfect performance on face-building categorization. They were associated with average celebrity identification performance and latency: no blur: 95%, 700 milliseconds; low blur: 90%, 700 milliseconds; and high blur: 60%, 850 milliseconds (Cherian et al. 2008). Images were presented in blocks, with roughly 100 images designated to each block (100 for the first 3 blocks, 105 for fourth). Subjects were allowed to take breaks between blocks. Each image was presented in the context of a trial, which included a 300 ms fixation cross, the image for 300 ms, a 1500 ms fixation cross, and a 200 ms blank screen. To keep large chunks of trials 21 from containing only certain types of images (all celebrity faces, all non-celebrity faces, or all buildings), every 15 trials contained an equal number of each type. Besides this restraint, the images designated to each block, and the orders of the images within the blocks were randomized for each subject. After the MEG portion of the experiment, subjects were asked to fill out a celebrity recognition survey. They were given a list of names of the celebrities used in the experiment, and asked to indicate which ones they could recognize. They were told to answer "yes" if they were sure they could recognize the person by picture alone, and had no doubts. They were explicitly told to answer "no" if they thought they could maybe recognize the person. Images of non-recognized celebrities and their non-celebrity matches were excluded from MEG analysis. Subjects were excluded if they could not recognize over 60% of the celebrities, or did not performed above 70% on the one-back task. This retained 13 subjects, but two had noisy data due to machine malfunction, and one lacked M170s for celebrity faces and buildings. Therefore, I analyzed data from 9 subjects. MEG Data Noise Reduction All MEG data were cleaned using M160's implementation of the Continuously Adjusted LeastSquares Methods (CALM). This made use of the 3 orthongonally-oriented magnetometers to remove external noise from the data (Adachi et al. 2001). 22 Each trial's MEG data (from 200 ms before to 600 ms after image presentation) was automatically inspected for artifacts due to eye blinks, muscle movement, etc. A trial was marked "bad" if its MEG data surpassed an amplitude threshold of 2 pico-Teslas. If a particular sensor had more than 15 bad trials, it was excluded from future analysis. If a trial was found to be bad in any of the remaining sensors, that trial was excluded from future analysis. The remaining data were 20 Hz lowpass filtered (48 order Hamming-window based, linear-phase filter) using zero-phase digital filtering (this made the filter essentially 96 order). The low filter cutoff excluded more noise from the signal, but reduced the amplitude and sharpness of the Ml 70. I chose this cutoff because it had precedence in Ml 70 research, and because I was concerned about noise in the results due to relatively few trials per condition (45 before trial exclusion) (Lueschow et al. 2004). Selecting SOIs: Sensors that displayed an M170 An average MEG waveform was created for each condition, for each sensor. These were made by averaging trial data from 0 ms to 600 ms after image presentation, with a 200 ms baseline correction. As previously done in M170 studies, these waveforms were used to identify sensors that displayed an Ml 70 (sensors of interest, SOIs) (Linkenkaer-Hansen et al. 1998; Taylor et al. 2001). I identified SOIs using the waveforms belonging to no-blur celebrity face, non-celebrity face, and building conditions. A sensor was marked as an SOI if it had a predominant peak between 140ms - 230ms after image presentation. 23 I identified SOs by hand. I chose against identifying SOIs automatically with an amplitude threshold, which would have been the most appropriate alternative to my method, because I found that small deviations in the threshold resulted in large SOI changes (Harris et al. 2005). Comparing the M170 Sensor Topographies for Buildings and Faces I compared the SOIs for celebrities, non-celebrities, and buildings. This provided insight into the neural generators underlying these MI70s, and the general sensor topographies that can be expected in M170 experiments. First, I assessed whether the face SOIs were more similar to each other than to building SOIs. For each subject, I calculated the proportion of non-celebrity SOIs that were celebrity SOIs and vice versa. Similarly, I calculated the agreement between non-celebrity and building SOLs; and celebrity and building SOIs. This was done for each M170 area separately (left and right, temporal and occipital), and the results were averaged across areas before comparison. Results were compared using a paired t-test with Bonferroni correction. Second, I compared the celebrity, non-celebrity, and building SOIs with the greatest M170 amplitude. These SOIs were, theoretically, closest to the M170 neural generator. I found the sensor with the maximum amplitude for each subject, in each area (left and right, temporal and occipital), for no blur celebrities, non-celebrities, and buildings. I compared the agreement of these sensors (i.e. how many of the celebrity maximum SOIs were building maximum SOLs, etc.) within each subject and area. I also compared the locations of maximal SOIs projected onto the 24 scalp, to reveal any consistent differences between the SOI locations, such as building SOIs being more anterior than face SOIs. Results were compared within each area using a paired ttest with Bonferroni correction. Examining the M170 Amplitude and Latency Across Conditions The no-blur celebrity face, non-celebrity face, and building conditions were used to identify the sensors for M170 analysis. First, M170 peak amplitudes and latencies were identified for every SOI using the MEG waveforms. Second, these SOIs were divided, based on Ml 70 amplitude topography, into the four areas that constitute the M170's one dipole per hemisphere pattern: left temporal, right temporal, left occipital, and right occipital. Third, the four sensors with the largest Ml 70 amplitudes in each area were selected for analysis. If less than four sensors were available in an area, all of them were selected. The selected sensors' MEG waveforms were averaged together for all conditions (for areas, conditions, and subjects separately). I averaged across multiple sensors for the same reason I used a 20 Hz lowpass filter: because it had precedence in M170 research, and because I was concerned about noise in the results due to relatively few trials per condition (Barrie et al. 2006). I chose to use only the no blur conditions for sensor selection, based on my assumption that the Ml 70s to clear and blurry images of the same type were generated by the same neural activity. The M170 peak amplitudes and latencies were identified for each condition (for areas and subjects separately), using the MEG waveform averages (MEG waveforms collapsed across as many as four sensors). Multiplying the right hemisphere temporal and left hemisphere occipital 25 SOI amplitudes by -1 rectified M170 amplitudes (these SOs produce negative M170s). Results were compared in SPSS (SPSS Inc., Chicago), using repeated-measures ANOVA (GreenhouseGeisser corrected) with four factors: hemisphere (right, left), location (temporal, occipital), condition (celebrity face, non-celebrity face, building), and blur (no blur, low blur, high blur). Pairwise comparisons were done using a Bonferroni correction. 26 Results In reporting the results, standard errors are referred to as "SE." Unless otherwise noted, significance was at a level of p < 0.05, and non-significant effects were at a level of p > 0.1. Comparing the M170 Generator for Buildings and Faces Non-celebrity and celebrity SOIs agreed (mean 96%, SE 1.0%) more than both non-celebrity and building SOIs (mean 90.4%, SE 2.4%) and celebrity and building SOIs (mean 90.1, SE 2.4%). However, this effect was not significant (Bonferroni corrected t-test). Maximum SOIs tended to be the same for celebrities and non-celebrities (61% of celebrity maximum SOIs were non-celebrity maximum SOIs and vice versa). Maximum SOIs were less congruent for celebrities and buildings (28%), and non-celebrities and buildings (33%). Overall, the celebrity, non-celebrity, and building maximum SOI locations were not significantly different (Bonferroni corrected t-test) (Fig. 2). Figure 2. A scalp plot of the maximum M170 response sensors for each M170 area. Ovals are centered on the mean location, and extend outwards to indicate the standard deviation in the horizontal and vertical directions. Posterior non-celebrity sensors are completely covered by celebrity sensors. Celebrity Non-Celebrity c// Buildi 27 Examining the M170 Amplitude and Latency Across Conditions The M170 peak amplitude and latency were recorded (for each subject) in the four M170 areas: left temporal (LT), right temporal (RT), left occipital (LO), and right occipital (RO). The results are summarized in Figure 3 and detailed in Figure 4. Effects were found using a GreenhouseGeisser corrected repeated measures ANOVA, and pairwise effects were found using a Bonferroni correction. M170 latencies are reported in seconds, and amplitudes in 1013 Tesla. Figure 3. A schematic summary of the results, displaying the M170 mean amplitude and latency averaged across all M170 areas. Celebrity and non-celebrity M170s were similar (celebrity pictured is Martha Stewart). Face M170s were earlier and larger than building M170s. Blurring reduced and delayed the M170s for all image types. Degraded Images Clear Images 1.5 11.51 LI 1.3 ) 1.2. 1.3 . 1.2 1.1 1.1 . ~1 1.0 0.9 - 170 U 175 180 U 1.0 U U 185 190 ~ 0.9 U 195 200 LI H E 170 175 1 U I I U 180 185 190 195 200 Latency (ms) Latency (ms) 28 Figure 4. The average M170 peak amplitude (a) and latency (b) across subjects for the four M170 areas: left temporal (LT), right temporal (RT), left occipital (LO), and right occipital (RO). Error bars indicate the standard error. a 210 T T Celebrity Non-Celebrity 200 Building C 0 190 C., E, T T-rI 180 170 -J 160- LO RO b 2.5 2 CelebrityII TT Non-Celebrityinoj ITT 1.5 TT BuildingoJ IF T 1 T IT IT E 0.5 0 RT LO 29 RO M170 latencies showed no significant effect of hemisphere: left or right. There were significant effects of location: temporal or occipital; condition: celebrity, non-celebrity, or building; and blur level: no, low, or high. There were no significant factor interactions. M170 latencies (reported in seconds) were overall earlier in occipital areas (mean 189, SE 4.8) than temporal areas (mean 177, SE 2.4). The M170 latencies for celebrity faces (meanl8l, SE 3.0) and non-celebrity faces (mean 180, SE 2.8) were not significantly different, but were both significantly earlier than for buildings (mean 188, SE 3.6). The M170 latencies for no blur (mean 180, SE 3.0) and low blur (mean 181, SE 3.2) were not significantly different, but were both significantly earlier than for high blur (mean 188, SE 3.1). M170 amplitudes showed no significant effect of hemisphere: left or right. There were significant effects of location: temporal or occipital; condition: celebrity, non-celebrity, or building; and blur: no, low, or high. There was a marginally significant interaction of hemisphere and location (p < 0.7). We think this was due to artificial cancellation of the left occipital M170 by the much larger left temporal M170. There were no other significant factor interactions. M170 amplitudes (reported in 10-13 Tesla) were overall smaller in occipital areas (mean 0.86, SE 0.13) than temporal areas (mean 1.56, SE 0.21). The M170 amplitudes for celebrity faces (mean 1.28, SE 0.16) and non-celebrity faces (mean 1.27, SE 0.15) were not significantly different. The M170 amplitudes for non-celebrity faces were significantly larger than for buildings (mean 1.09, SE 0.13). The M170 amplitudes for no blur (mean 1.37, SE 0.18), low blur (mean 1.20, SE 30 0.15), and high blur (mean 1.10, SE .12) were not significantly different from each other (except no and low were marginally significantly different, p < 0.1). 31 Discussion In this report I addressed two questions: why there is an M170 response to buildings, which seems contrary to the M170 face-specificity; and what is the relationship between the M170 response and face identification. First, I discovered that different neural areas were responsible for building and face Ml 70s. Overall, the celebrity and non-celebrity SOIs (sensors containing an M170 response) were not more similar than the celebrity and building SOIs, nor the non-celebrity and building SOIs. However, this may be related to the fact the face M170s were stronger than building Ml70s, and therefore extended into more sensors. The locations of the maximum M170 SOI (the sensor closest to the M170 neural source) were similar for celebrities, non-celebrities, and buildings. These results indicated that the neural areas that subserve face and building M170s were identical or very close to each other (beyond sensor resolution). However, I found that celebrities and non-celebrities shared maximum SOIs more often than they did with buildings (60% agreement compared to 30%). I assumed that the same neural area generated celebrity and non-celebrity M170s. Therefore, this suggested that the face and building M170 generators were distinct. These findings were similar to what was found in a previous study with motor-bike and face images. The M170 generators for these stimuli seemed to be different, but very close to one another, and had no or little consistency in their relative positions (Swithenby et al. 1998). Hand 32 and letter string images also generated M I70s, which were subserved by different neural areas than the face M170 (Watanabe et al. 1999; Tarkiainen et al. 2002). Second, I found that the building MI 70 was building-specific (to a first approximation), and was not a "baseline" response. Just like for faces, degrading building images reduced and delayed the building Ml 70. Together with the finding that different neural areas generated building and face M I70s, this implied that there was a cluster of cells in/around the fusiform gyrus selective for faces, and another selective for houses (or larger category). This is congruent with the notion of the fusiform gyrus is divided into distinct areas, each specific to a particular visual image category (Watanabe et al. 1999). However, this possibility is somewhat troublesome. If houses and motor-bikes had representations, there must be hundreds or thousands of objects with their own representations in the fusiform gyrus. How many representations could this small brain area hold? Maybe, some part of the fusiform gyrus responds to many recognizable objects, instead of a single type. More research will be needed to address this conundrum. Third, I found no difference between celebrity and non-celebrity MI 70s at any blur level. This implied that the MI 70 was not influenced by recognition of the image. It reinforced the notion that the Ml 70 is a feed-forward pre-identification step in the face processing mechanism. Fourth, I observed that image degradation delays identification much more than the MI 70. Celebrity recognition was delayed about 100 milliseconds (in a separate study) by degradation that delayed the M170 about 10 milliseconds (Cherian et al. 2008). Additional delay must be 33 accumulated in neural processes after the M170, which occurs relatively early in the face identification process. Contributions In summary, this report contains four major contributions. First, it contains an overview of MEG techniques. This is written in a straightforward manner (not with equations), which is needed for science researchers using MEG. Second, it summarizes all previous M170 findings, which is another discussion that was previously missing from the literature. Third, with an experiment, it established that separate neural areas generate the face and building M170s. This has interesting implications for the organization and function of the fusiform gyrus brain area. Fourth, it confirmed that the M170 is a relatively early step in the face identification process, and is not impacted by face image recognizability. 34 References Adachi Y, Shimogawara M, Higuchi M, Haruta Y, Ochaiai M. Reduction of non-periodic environmental magnetic noise in MEGmeasurement by continuously adjusted least squares method. Applied Superconductivity, IEEE Transactions on 2001; 11:669-672. Barrie J, Herdman A, Tanaka J, M L (2006). A whole-cortex MEG study of face processing in healthy adults. Society for Psychophysiological Research Annual Meeting. Cherian T, Morash V, Sinha P. Performance and time costs associated with degraded images. In preparation 2008; Clarke J, Braginski Al. The SQUID handbook. Weinheim: Wiley-VCH, 2003. Defike I, Sander T, Heidenreich J, Sommer W, Curio G, Trahms L, Lueschow A. MEG/EEG sources of the 170-ms response to faces are co-localized in the fusiform gyrus. Neuroimage 2007; 35:1495-1501. Downing P, Liu J, Kanwisher N. Testing cognitive models of visual attention with fMRI and MEG. Neuropsychologia 2001; 39:1329-1342. Ewbank MP, Smith WA, Hancock ER, Andrews TJ. The Ml 70 reflects a viewpoint-dependent representation for both familiar and unfamiliar faces. Cereb Cortex 2008; 18:364-370. Furey ML, Tanskanen T, Beauchamp MS, Avikainen S, Uutela K, Hari R, Haxby JV. Dissociation of face-selective cortical responses by attention. Proc Natl Acad Sci U S A 2006; 103:1065-1070. Halgren E, Raij T, Marinkovic K, Jousmaki V, Hari R (1995). Magnetic fields evoked by faces in the human brain: 1. Topography and equivalent dipole locations. Society for Neuroscience Annual Meeting. 21. Halgren E, Raij T, Marinkovic K, Jousmaki V, Hari R. Cognitive response profile of the human fusiform face area as determined by MEG. Cereb Cortex 2000; 10:69-81. Himdlinen M, Hari R, Ilmoniemi RJ, Knuutila J, Lounsamaa OV. Magneto encephalographytheory, instrumentation, and applications to noninvasive studies of the working human brain. Reviews of Modem Physics 1993; 65:413-497. Harris A, Nakayama K. Rapid face-selective adaptation of an early extrastriate component in MEG. Cereb Cortex 2007; 17:63-70. Harris AM, Duchaine BC, Nakayama K. Normal and abnormal face selectivity of the MI 70 response in developmental prosopagnosics. Neuropsychologia 2005; 43:2125-2136. Itier RJ, Herdman AT, George N, Cheyne D, Taylor MJ. Inversion and contrast-reversal effects on face processing assessed by MEG. Brain Res 2006; 1115:108-120. Kanwisher N. Domain specificity in face perception. Nat Neurosci 2000; 3:759-763. Kato Y, Muramatsu T, Kato M, Shintani M, Yoshino F, Shimono M, Ishikawa T. An earlier component of face perception detected by seeing-as-face task. Neuroreport 2004; 15:225229. Kloth N, Dobel C, Schweinberger SR, Zwitserlood P, Bolte J, Junghofer M. Effects of personal familiarity on early neuromagnetic correlates of face perception. Eur J Neurosci 2006; 24:3317-3321. Linkenkaer-Hansen K, Palva JM, Sams M, Hietanen JK, Aronen HJ, Ilmoniemi RJ. Faceselective processing in human extrastriate cortex around 120 ms after stimulus onset revealed by magneto- and electroencephalography. Neurosci Lett 1998; 253:147-150. Liu J, Harris A, Kanwisher N. Stages of processing in face perception: an MEG study. Nat Neurosci 2002; 5:910-916. 35 Liu J, Higuchi M, Marantz A, Kanwisher N. The selectivity of the occipitotemporal M170 for faces. Neuroreport 2000; 11:337-341. Lu ST, Hamalainen MS, Hari R, Ilmoniemi RJ, Lounasmaa OV, Sams M, Vilkman V. Seeing faces activates three separate areas outside the occipital visual cortex in man. Neuroscience 1991; 43:287-290. Lueschow A, Sander T, Boehm SG, Nolte G, Trahms L, Curio G. Looking for faces: Attention modulates early occipitotemporal object processing. Psychophysiology 2004; 41:350360. Marinkovic K, Raij T, Halgren E, Hari R (1995). Magnetic fields evoked by faces in the human brain: 2. Cognitive profile. Society for Neuroscience Annual Meeting. 21. Nakamura A, Yamada T, Abe Y, Nakamura K, Sato N, Horibe K, Kato T, Kachi T, Ito K. Agerelated changes in brain neuromagnetic responses to face perception in humans. Neurosci Lett 2001; 312:13-16. Sams M, Hietanen JK, Hari R, Ilmoniemi RJ, Lounasmaa OV. Face-specific responses from the human inferior occipito-temporal cortex. Neuroscience 1997; 77:49-55. Sato N, Nakamura K, Nakamura A, Sugiura M, Ito K, Fukuda H, Kawashima R. Different time course between scene processing and face processing: a MEG study. Neuroreport 1999; 10:3633-3637. Schweinberger SR, Kaufmann JM, Moratti S, Keil A, Burton AM. Brain responses to repetitions of human and animal faces, inverted faces, and objects: an MEG study. Brain Res 2007; 1184:226-233. Streit M, loannides AA, Liu L, Wolwer W, Dammers J, Gross J, Gaebel W, Muller-Gartner HW. Neurophysiological correlates of the recognition of facial expressions of emotion as revealed by magnetoencephalography. Brain Res Cogn Brain Res 1999; 7:481-491. Swithenby SJ, Bailey AJ, Brautigam S, Josephs OE, Jousmaki V, Tesche CD. Neural processing of human faces: a magnetoencephalographic study. Exp Brain Res 1998; 118:501-510. Tanskanen T, Nasanen R, Montez T, Paallysaho J, Hari R. Face recognition and cortical responses show similar sensitivity to noise spatial frequency. Cereb Cortex 2005; 15:526534. Tanskanen T, Nasanen R, Ojanpaa H, Hari R. Face recognition and cortical responses: effect of stimulus duration. Neuroimage 2007; 35:1636-1644. Tarkiainen A, Cornelissen PL, Saimelin R. Dynamics of visual feature analysis and object-level processing in face versus letter-string perception. Brain 2002; 125:1125-1136. Taylor MJ, George N, Ducorps A. Magnetoencephalographic evidence of early processing of direction of gaze in humans. Neurosci Lett 2001; 316:173-177. Watanabe S, Kakigi R, Koyama S, Kirino E. Human face perception traced by magneto- and electro-encephalography. Brain Res Cogn Brain Res 1999; 8:125-142. Willing R (2003). Airport anti-terror systems flub tests. USA Today. Xu Y, Liu J, Kanwisher N. The M170 is selective for faces, not for expertise. Neuropsychologia 2005; 43:588-597. Zhao W, Chellappa R, Phillips PJ, Rosenfeld A. Face Recognition: A Literature Survey. Computing Surveys 2003; 35:399-459. 36