Philip, Stacy.

advertisement
Listening to the Blink of an Eye: Using the Sonification of EEGs to
Recognize and Classify Types of Brain Activity
by
Stacy Philip
AN HONORS THESIS
for the
HONORS COLLEGE
Submitted to the
Honors College
at Texas Tech University in
partial fulfillment of the
requirement for
the degree designation of
HIGHEST HONORS
MAY 2015
Approved by:
__________________________________________________
Carl Seaquist, Ph.D., Thesis Director
Associate Professor, Department of Mathematics & Statistics
______________
Date
__________________________________________________
Brock Williams, Ph.D., Second Reader
Professor, Department of Mathematics & Statistics
______________
Date
__________________________________________________
Keira V. Williams, Ph.D., Thesis Instructor
Assistant Professor, Honors College
______________
Date
__________________________________________________
Michael San Francisco, Ph.D.
Dean, Honors College
______________
Date
The author approves the photocopying of this document for educational purposes.
ABSTRACT
A series of clinical tests were conducted at a Lubbock hospital, from which ten patient
electroencephalogram (EEG) recordings were selected for further analysis. These recordings
consisted of two normal and eight abnormal EEGs, which were stripped of all personal
identifying information. Recent publications indicate that sonification (converting data to sound)
allows the human ear to analyze series data and detect irregularities that might otherwise go
unnoticed. Since brain rhythms are typically lower than the human hearing range, signalprocessing techniques, including modulation, Fourier transforms, wavelet analysis, and digital
filtering, can be applied in an attempt to convert EEG signals to sound in the audible human
hearing range. In particular, this exploratory study strives to sonify blink artifacts and alpha
rhythms in an excerpt of an EEG recording from a ‘normal’ patient using modulation to produce
an A-C-E chord. The objective is to demonstrate that in addition to traditional visual analysis,
auditory acuity may be useful in the analysis of EEGs and in the identification of important EEG
features, without the loss of information.
ii
ACKNOWLEDGEMENTS
This author would like to extend gratitude to the Proactive Recruitment in Introductory
Science and Mathematics (PRISM) undergraduate research program for funding this research
and for all of the doors it has opened through research in regards to providing a wholesome and
enriching college experience. This author is also very thankful for Dr. Carl Seaquist for allowing
her to work on such an intriguing and fun project, while being a wonderful mentor throughout
this entire endeavor. This author is also grateful for fellow researcher Kate Ehnis, for the
partnership, friendship, and teamwork without which this venture would not be a success. Also,
this author appreciates Dr. Renato Gonik and Frederick Ramirez for their contributions and
assistance to make this project efficacious. In addition, special thanks to Dr. Ram Iyer for his
succors and support of this research. Also, this author thanks Dean Michael San Francisco, Dr.
Keira Williams, and Dr. Brock Williams for their helpfulness, mentorship, and guidance
throughout the Honors Thesis writing process. This author would also like to extend thanks to
the Texas Tech University Mathematics and Statistics Department and the Honors College.
iii
TABLE OF CONTENTS
TITLE PAGE.……………………...………………………………………………………………i
ABSTRACT………………………………………………………………………………………ii
ACKNOWLEDGEMENTS………………………………………………………………………iii
TABLE OF CONTENTS ………………………………………………………………………...iv
Chapter 1…………………………………………………………………………………………..1
Introduction………………………………………...………………………………...……1
Background………………………………………………………….…………….………2
Electroencephalograms…………………………………………………….……...3
Eye Blink Artifacts…………………………………………………….………….3
Listening to Data…………………………………………………….…………….5
Chapter 2…………………………………………………………………………………………..7
Previous Sonification Work…………………………………………………………….....7
Chapter 3…………………………………………………………………………………………15
Methodology……………………………………………………………………………..15
Data Selection...………………………………………………………………….15
MATLAB and EEGLAB………………………………………………………...17
Channel Selection………………………………………………………………..17
Montages…………………………………………………………………………18
Filtering………………………………………………………………………..…20
Resampling, Modulation, and Chord Selection………………………………… 21
Chapter 4…………………………………………………………………………………………25
Results……………………………………………………………………………………25
Conclusion……………………………………………………………………………….26
Chapter 5…………………………………………………………………………………………27
Future Work……………………………………………………………………………...27
REFERENCES…………………………………………………………………………………..29
APPENDIX A (MATLAB CODE)...……………………………………………………………32
iv
Chapter 1
This chapter discusses introductory information that will be further discussed throughout this
paper for basic knowledge and reference.
Introduction
Electroencephalograms (EEGs) monitor the brain’s electrical functions by comparing
differences in surface voltages. The analysis and display of EEGs for visual analysis is
ubiquitous. When one is able to understand a normal EEG, it is much easier to identify
abnormalities and better understand how the brain works. While a normal EEG does not exclude
a clinical diagnosis, an abnormal finding on EEGs supports a diagnosis, indicates cerebral
dysfunction, or may have nothing to do with the reason that the study was performed (Tatum,
2008). Therefore, EEGs are quite utilitarian and can be an advantageous diagnostic tool in the
identification of abnormal brain functions.
Visually learning how to classify various types of brain activity is a complex, rigorous,
and time-consuming process. Recent work indicates that auditory analysis may be useful in
detecting irregularities in EEGs and assessing EEG data apart from traditional visual analysis
(Khamis, Mohamed, Simpson, & McEwan, 2012). Since brain waves consist of frequencies
inaudible to the human ear, it is necessary to use digital signal processing techniques to convert
the signals to sound without losing valuable information. The auditory analysis of EEG signals
will aid in the early detection of abnormal EEG signals to better understand the human brain, and
hopefully improve diagnosis and treatment of neurological abnormalities in the future. Thus,
training non-experts how to recognize various types of brain activity through simple auditory
means will allow those with neural irregularities to gather information about their condition to
determine appropriate medical regimes as treatment in a timely and efficient manner.
1
Background
Research on the sonification of electroencephalograms has been attempted in the past,
though the work is still novel in this developing field. Dale Purves’s Neuroscience (2008)
textbook was used over the course of this exploration because this text addresses complex
neuroscience topics in an approachable manner, using the analytical tools of genetics, molecular
and cell biology, systems anatomy and physiology, behavioral biology, and psychology.
Therefore, it was helpful for the study of neuroanatomical and neurophysiological concepts that
were necessary during the research process to better understand brain structure and function.
Electroencephalograms
To begin work in this area, Tatum’s Handbook of EEG Interpretation was referenced in
order to learn about key features of electroencephalograms and their functional uses throughout
the course of this research (Tatum, 2008). Tatum explains that the basis of cerebral potentials
relies upon the nervous system’s intrinsic electrophysiological properties. Being able to identify
the generator source(s) and electrical field(s) of propagation provides the foundation for
recognizing electrographic patterns that denote the expression of the “brain waves” as normal or
abnormal. In general, the majority of regular EEGs recorded at the surface of the scalp represent
pooled electrical activity produced by large numbers of neurons. Electrical signals are created
when electrical charges move within the central nervous system. Neural function is normally
maintained by ionic gradients established by neuronal membranes. Sufficient duration and length
of small amounts (measured in μV) of electrical currents of cerebral activity are required to be
amplified and displayed for interpretation (Tatum, 2008).
From the patient’s scalp, electrodes conduct electrical potentials to an electrode box, also
known as a jackbox. Subsequently, a montage selector allows EEG signals to pass through
2
amplifiers before filtering and auxiliary controls regulate the signal output. Data display follows
acquisition and processing and has a wide variety of data presentations for EEG interpretation.
Electrode placement has been standardized by an international 10–20 system that uses
anatomical landmarks on the skull. These sites are then subdivided by intervals of 10% to 20%
and to designate the site where an electrode will be placed. The designations for where on the
scalp electrodes are assigned in this 10-20 system are: Fp (frontopolar), F (frontal), T (temporal),
O (occipital), C (central), and P (parietal). Afterward, numbers combined following the letters
for location designate either the left (odd numbers) or right (even numbers) hemisphere of
electrode placement, where the “z” designation reflects midline position (i.e., Cz = central
midline) as seen in Figure 1 from the Handbook of EEG Interpretation below.
Figure 1. The black circles indicate the 10-20 electrode placement system.
3
Eye Blink Artifacts
The Handbook of EEG Interpretation aided in the identification of artifacts in the EEG
signals and then interpreting them to distinguish between normal and abnormal EEGs. Due to the
fact that recording the brain’s electrical activity may come with noncerebral interference, diverse
generators of nonphysiological and physiological artifacts may deceive the interpreter to believe
that the apparent sources are abnormal or epileptiform. When in doubt, an EEG interpreter must
assume that the source is an artifact until proven otherwise (Tatum, 2008).
Specifically, the Handbook of EEG Interpretation text was exceptionally useful to
identify eye blink artifacts, which are generated by the electrical potential produced by vertical
movement of the eye. Normally, the eye functions as an electrical dipole with a relative
positivity of the cornea compared to the retina. The potential created is a DC potential of higher
amplitude (mV) than the amplitude produced by the brain (μV). The artifact is produced in the
electrodes around the eye (FP1/2) during vertical eye movements. With an eye blink, the cornea
rolls up with resultant positivity in the FP1/2 electrodes relative to the F3/4 electrodes and
creates a downward deflection during the normal Bell’s phenomenon (Tatum, 2008). Bell’s
phenomenon is a medical sign that enables an upward and outward movement of the eye to be
observed when an attempt is made to forcefully close the eyelids (Francis & Loughhead, 1984).
Tatum (2008) states that by having electrodes record above and below the eye helps distinguish
the brain as the “generator,” where every channel is the same polarity, as opposed to it being
distinguished as an artifact, where there is opposite polarity in electrode sites above and below
the eye (Tatum, 2008).
Having a vertical eye blink artifact present helps determine the state of the patient as
being awake. In the same way, during drowsiness, slow rolling (lateral) eye movements are
4
beneficial. Lateral eye movements are usually easily recognized because they create phase
reversals in the anterior temporal derivations that are of opposite polarity on opposite sides of the
scalp EEG. When the eyes move to the left yielding a positive phase reversal in F7 due to the
cornea polarity, the homologous F8 electrode site demonstrates a negative phase reversal from
the retina. The positive phase reversals noted at the F8 derivation is due to the proximity of the
cornea. The homologous F7 electrode site is negative due to the conjugate effect from the retina
(Tatum, 2008).
Eye movements can be detected by using a single channel connecting the right upper
lateral eyebrow and the left lower lateral eyebrow. However, due to the fact that vertical eye
movements are usually the source of confusion, bilateral infraorbital electrodes referred to the
ipsilateral ear as a reference may better represent the eye as a dipole and demonstrate phase
reversals that are out-of-phase with cerebral activity when due to eye movements (Tatum, 2008).
Listening to Data
Transforming EEG data signals that are traditionally only analyzed visually by a trained
expert into the auditory spectrum requires interpreting the data sequence, which is usually a time
series, through signal processing as an audio waveform. Listening to data provides a means by
which large datasets can be explored by mapping properties of the EEG data points into music.
Then, by hearing this continuous, dynamic, and broad-spectrum data, a new platform is provided
in which to interpret the signals. Mills (2012), explains that this idea of “Electroacoustics has
been at the forefront of signal engineering and signal processing since ‘the transducing 1870s,’
when the development of the telephone marked the first successful conversion of a sensuous
phenomenon (sound) into electrical form and back again.” Thus, hearing is a simple technique to
perceive data, because there is already a strong relationship between hearing and electronic
5
signal processing.
Signal Processing for Communications (2008) was referred to in order to learn more
about how to understand and modulate digital signals. The authors Martin Vetterli and Paolo
Prandoni, produced a text from course notes developed over a period of a dozen years teaching
undergraduates the basics of signal processing or communications at École Polytchnique
Fédérale de Lausanne. This textbook has valuable information on the basis of various digital
signal-processing techniques, which were consulted over the course of the research project.
6
Chapter 2
This chapter discusses former work conducted by individuals in the field of sonification and how
this specific research team sought to conduct its own sonification study using EEG signals.
Previous Sonification Work
In 1999, Barrass and Kramer introduced the many benefits that sonification, also referred
to as audification, provides for a plethora of disciplines, including practical uses in the medical
field, in assistive technologies to help blind individuals be more independent, in assisting
engineers analyze computer models and simulations of complex systems, in scientific
visualizations, in emergency services, and in aircraft cockpits, among many other examples. The
authors also survey and outline the many methods to designing and realizing a sonification.
Kramer classifies methods for designing an auditory display along a spectrum between analogic
and symbolic. An analogic representation has an immediate and intrinsic correspondence
between the sort of structure being represented and the representation medium. There is a one-toone mapping between points in the data space and points in representation space. On the
contrary, a symbolic representation categorically denotes the thing being represented and
relations within the representation do not necessarily reflect intrinsic relationships in what is
being represented. The authors also discuss the semiotic distinctions (syntactic, semantic and
lexical) that have also been used to classify the best-known methods of auditory display
(earcons, auditory icons and parameter mapping).
Barrass and Kramer (1999) also present the issues that have arisen from this range of
approaches and applications of converting data into sound. For example, listeners need to
correctly deduce data relations from the sounds. They say that “one of the reasons that
audification fails for arbitrary data, such as stock market figures or daily temperatures, is even a
7
slow playback rate of 8000 samples/s requires many data points to make a sound of any duration.
The resultant short sound does not reveal valuable information to the listener. Even for long
sequences, the sounds do not resemble natural or familiar noises, because arbitrary data points
seldom obey the physics of natural sounds transmitted through air.” In addition, some
classifications of sound events tend to be categorical. They say that relevant changes in data
should insure a change in what is perceived and signify meaningful changes in the data. In
addition, the authors relate sonification to other areas where similar issues have arose including
human-computer interaction, sound design for commercial products, and computer music.
In 2002, Hermann, Meinicke, Bekel, Ritter, Müller, and Weiss conducted a study that
presented different techniques in order to produce acoustic representations for EEG data. Their
data was received from “psycholinguistic experiments where subjects are exposed to three
different conditions based on different auditory stimuli.” The stimuli included Spoken Language
(story) with an Austrian-German speaker (mean duration of two minutes); Pseudo Speech, which
consisted of auditory patterns generated by amplitude and frequency modulation using a base
frequency of 200 Hz and an amplitude envelope that resembles the real spoken sentences; and
EEGr, in which the EEG was recorded for two minutes during rest with open eyes. They utilized
nineteen scalp electrodes that were placed according to the 10-20 positioning system and
measured against the averaged signal of both earlobes. Before beginning their analysis, the
signals were band-pass filtered between 0.3 Hz to 35 Hz and digitized using a 16-bit quantization
and a sampling rate of VSR = 256 Hz. They hoped to unearth elements of neural processes that
correlate with high-level cognitive activity. The three types of sonification described in their
paper include spectral mapping sonification, which provides a direct assessment of the recorded
data; distance matrix sonification, which enables nonlinear long range correlations to be detected
8
at high time resolution; and differential sonification, which summarizes the comparison of EEG
measurements where each subject experiences different conditions. Since EEG data consists of
multiple time series, it contains a lot of noise, which complicates automatic pattern detection that
requires high-developed human auditory skills in signal/noise separation. The authors believe the
trained listener can conclude coarsely which electrodes and what frequency bands were affected
by the relevant condition.
In 2004, Hinterberger and Bair introduced a device known as POSER, which stands for
the Parametric Orchestral Sonification of EEG in Real-Time, for the self-regulation of brain
states. This device is implemented in the Thought-Translation-Device TTD, which is a braincomputer interface designed for the communication with completely paralyzed patients. The
device enables auditory feedback of multiple EEG characteristics in real time. Specifically, each
of six frequency bands is assigned an instrument of a MIDI device. The time-dependent
parameters extracted from each frequency band modulate the timing, pitch, and volume of the
instrument.
In a pilot study, Hinterberger and Bair tested the ability of subjects to perform a simple
discrimination task using this parametric sonification in real time. The EEGs in the study were
sampled at a rate of 256 per second in an amplitude range of +/-1mV and a resolution of 16 bits.
The pilot study was carried out with the ProComp+ amplifier system. The raw EEG signal is
stored to disk and complemented with status information such as a code for the feedback task
and other events during a trial. Their data processing was performed at 256 samples/s, while
sonification was paced at a rate of 127 per second, thus offering a sufficient update rate and
timing resolution. For excellent real-time behavior, a 1.6 GHz Pentium IV machine running
under MS Windows XP was utilized. In regards to processing, after calibration of the EEG-
9
signal into μV, the spatial filter offers the user to arrange the EEG-channels by a linear
combination of incoming channels. In the standard setting, three channels of EEG were used,
including the Cz-mastoids, C3-mastoids, and C4-mastoids. They roughly corrected eye
movement artifacts by subtraction of a fixed factorized amount of the vertical Electrooculogram.
Afterwards, they then band-passed the EEG signal into several different frequency bands and
used an FIR (finite impulse response) filter with either 127 or 63 coefficients used. In order to
sonify EEG signals that were below 12 Hz, they triggered the touches of a note at the maxima of
a wave. In addition, the potential differences between subsequent extrema (maxima minus
previous minima or minima minus previous maxima) were calculated. The authors used the
inverse time difference between consecutive maxima of a band pass filtered signal to serve as an
estimate of the “instantaneous” frequency of a signal. In order to lower the update rate of the
sounds to half the sampling rate, they down-sampled all output signals.
For the study, ten healthy participants, five male and five female, between the ages of 20
and 44, were selected to receive full orchestral feedback of their brain activity from one channel
of EEG (Cz versus C3). The subjects all appreciated the characteristic timbres of instruments and
found the stereo presentation helpful for distinction, e.g. between theta and alpha rhythms. The
participants were instructed to focus their attention alternately on two different sounds. The
sequence of the tasks was defined by the computer in a balanced order and symbolized on a
screen by two vertically arranged rectangles. A red colored upper rectangle asked participants to
focus attention on the rhythmic vibraphone sounds (alpha, theta, and gamma). A red colored
lower rectangle asked them to focus attention on the smooth sounds (SCP, delta, and beta). T-test
revealed significant differentiation between the tasks in at least one parameter for all participants.
The use of the POSER device in a closed-loop feedback application leads to a new level of
10
interaction with one’s state of consciousness.
In 2006, Baier, Hermann, Sahle, and Stephani, described a technique used to sonify
rhythmic activity of epileptic seizures as measured by human EEG, particularly using eventbased mapping of parameters, which relates information in regards to auto- and crosscorrelations of the multivariate data. They used a group of patients with childhood absence
seizures (because they only show one epileptic EEG symptom), and they found consistent intrapatient conservation of the rhythmic pattern as well as inter-patient variations, especially in terms
of cross-correlations. Therefore, they were able to determine that sonification in clinical
monitoring is now possible. However, for real-time application, one fixed set of sonification
parameters might not be sufficient, so one would need to adjust the sonification interactively
according to features such as the state of the patient, the pathologic features being detected, and
the amount of artifacts. This could be solved by providing a graphical user interface for real-time
interaction with the main parameters involved in the sound rendering.
The following year, in 2007, Baier, Hermann, and Stephani introduced a multivariate
event-based sonification that, in addition to displaying salient rhythms, uses pitch and spatial
location to provide such cues in a multi-speaker environment. Using clinical recordings with
epileptic seizures, they have demonstrated how the spatio-temporal characteristics of EEG
rhythms can be perceived in such sonifications. They support the argument that one of the
advantages of auditory displays stems from how the human sense of listening deals with multiple
events that arrive in one sound stream, which they say can be particularly suited for data mining
and data exploration of multivariate data sets. They also agree that there have not been any
explicit medical applications of the auditory representation of space in the case of the human
EEG. The data was used in a source density montage, which creates an individual reference for
11
each electrode defined as the weighted sum of the potentials from its nearest neighbors. This
reference acts as a spatial high-pass filter that improves localization and reduces spurious crosscorrelations compared to common references, particularly in the case of generalized rhythms.
They were successful in obtaining an EEG sample in real time. This work is significant in that it
shows that the sonification of EEGs has been successfully executed; however, it only deals with
the spacio-temporal aspect of the human brain rather than an all-encompassing approach.
Also in 2007, Campo, Hoeldrich, Eckel, & Wallisch described two software
implementations for EEG data screening and realtime monitoring via sonification. Their study
explains how the alpha range is between 8 and 13 Hz, and it is associated with “general state of
relaxedness, and non-activity of the brain region for visual tasks; it is most prominent with eyes
closed.” For each of the two software designs, the EEG signals were split into frequency ranges
which closely correspond to the traditional EEG bands constant relative bandwidths were used of
one octave each to attain a more equal energy balance between the bands. The ranges are:
deltaL(ow) from 1-2 Hz, deltaH(igh) from 2-4 Hz, theta from 4-8 Hz, alpha from 8-16 Hz,
which is wider than the traditional value of 13, and includes the mu-rhythm band, which is a bit
wider than the traditional value of 13, beta from 16-32 Hz, and gamma from 32-64 Hz.
Trevisan and Jones’s (2010) study provides a simple and portable system that is able to
generate MIDI output based data collected through an EEG collecting device. This device is
beneficial because the therapeutic effects of listening to the music created by the brain waves
have been seen in different documented cases to successfully treat health problems. This paper
seeks to make the most accurate representation of brainwaves, which had yet to be performed.
Several neuroscientists have adopted MIDI protocol in order to monitor brain waves in real time.
Therefore, the authors alternated several variants of positioning, including power spectrum,
12
event-related potential, and spectral centroid. This study contributes to the field of neurofeedback
by providing criteria tools for assessment.
Finally, “A Review of Real-Time EEG Sonification Research” was also used in the
research process in order to obtain a thorough and extensive overview on the work that has been
conducted on the sonification of EEGs thus far. The authors present an overview and critical
review of the principal research to date in EEG data sonification. First, the authors identify
several sub-domains of real-time EEG sonification and discuss their diverse approaches and
goals. Then, they describe the search and inclusion criteria, and present a synoptic summary
table spanning over fifty different research projects or published research findings. Thirdly, they
analyze sonification approaches to the various EEG data dimensions, such as time-frequency
filtering, signal level, and location, before going on to consider higher order EEG features.
Lastly, they discuss future application domains, which may benefit from new capabilities in the
real-time sonification of EEG data. They hope that their critical review may help to reduce
research fragmentation and may aid future collaboration in this emerging multidisciplinary area,
which was definitely made useful throughout the course of this study.
The limitations of this study are reflected in the fact that there is a lack of common
terminology due to the high level of multidisciplinary collaborations in the field of EEG
sonifications. Also, many papers lack technical details on the sonifications used, the EEG
recording equipment, and the processing techniques. Future papers addressing the topic of EEG
sonification, in any of the application domains, would benefit from providing more
documentation on the EEG methodology and sonifications used. EEG recording methodology
description should include: equipment used (amplifier models, electrodes), electrode placement,
channels used and referencing technique, EEG sampling rate, applied low and high pass pre-
13
filtering, signal processing details (e.g., FFT block size), artifacts reduction. In addition, there
has not been proper validation done on sonification methods in order to determine which is the
most efficient for representation of EEG data for particular purposes with proper control
methodologies.
14
Chapter 3
This chapter discusses the methodology used to conduct this study as a test of proof to determine
if EEG signals could successfully be transformed into sound.
Methodology
Ten patient Electroencephalogram (EEG) recordings were selected from a group of
clinical EEGs conducted at a Lubbock hospital. These recordings include two normal and eight
abnormal EEGs, which were stripped of all personal identifying information. The patients that
these recordings were taken from include individuals who were ‘normal,’ meaning that they did
not have any apparent medical conditions, to those who were comatose, to some who were
epileptic.
Data Selection
For the initial study, to test if it was feasible to convert the electroencephalogram signals
into the auditory spectrum, a normal patient’s recordings were selected for analysis. In particular,
a seventeen-second excerpt (from 79 seconds to 96 seconds) from the EEG recording was chosen
for further exploration. This selection contained blink artifacts, followed by the eyes closing,
leading to alpha rhythms, as seen in Figure 2. The alpha rhythms are characterized by their
frequency between 7.5 and 13 Hz, and usually appear when the eyes are closed and the subject is
relaxed.
15
16
Figure 2. Raw data of blink artifacts, eyes closing, and alpha rhythms viewed in EDF Browser
MATLAB and EEGLAB
These raw signals were read into a numerical computing software program known as
MATLAB, specifically using a compatible EEGLAB toolbox, which can process data from
electroencephalography, among other signals through the aid of coding. While using EEGLAB,
the normal patient’s EEG data, saved file in EDF format, is uploaded in MATLAB.
Channel Selection
From the EEG recording, electrodes were positioned on the patient’s scalp using the 1020 placement system to obtain signals from the brain’s frontal, occipital, and parietal regions.
The odd numbered electrodes correspond with the left side of the brain, while the even numbered
ones correspond to the right side of the brain. Utilizing a multi-platform viewing application
known as EDF Browser, the data received from the electrodes placed on these regions can be
visualizes as displayed in Figure 2. Though several electrodes were placed on the patient’s head
during the EEG recording to obtain data, the specific channels used for this study’s analysis were
those from the Fp1, Fp2, F3, F4, A1, A2, O1, and O2 electrodes (although data from all
electrodes is initially all loaded into MATLAB). The researchers chose to use strictly these
channels for further analysis, because according to the Handbook of EEG Interpretation, these
key channels are those that are primarily associated with the blinking of an eye. Therefore, this
would eliminate having to exclude extraneous data retrieved from EEG signals that gather data
from other parts of the brain later on in the analysis stage.
Based on suggestions from Niedermeyer’s Electroencephalography: Basic Principles,
Clinical Applications, and Related Fields, the optimal way to detect alpha rhythm amplitude is
by using a referential montage to the ipsilateral ear, taking into careful consideration the
17
interelectrode distances. The authors of this text, Schomer and Lopes da Silva, say, “the
maximum alpha voltage is usually over the occipital region as such, but a bipolar montage with a
parasagittal array may obscure rather than reveal the true alpha maximum.” They have found that
“alpha amplitude may be quite small in the channels displaying P3-O1 and P4-O2 because of
massive homophasic activity, which results in canceling out” (Schomer and Lopes da Silva,
2010).
Montages
Representations of EEG channels can be exemplified through various montages, also
known as arrays, which are combinations of multiple derivations or electrode recording
combinations. Most commonly, the two different types of montages that are used are the
Monopolar (Referential) and the Bipolar. Referential montages include using the common
average reference, the Cz reference, or an ear reference. Bipolar montages consist of an adjacent
pair of electrodes of the 10-20 system of electrode placement and potentials are localized based
on phase reversal; the types of bipolar montages include Longitudinal (anterior to posterior) and
Transverse (left to right).
For this study, a referential montage was used to display the EEG signal data, because it
provides a perspective in which there is a common reference electrode for all of the channels.
The reference is placed in an electrically quiet spot on the patient in order not to contaminate the
other leads. A specific type of referential montage uses ear reference (contralateral and
ipsilateral), which provides an advantage in that there is little to no cancellation because there are
long interelectrode distances, which allow potentials to appear with maximal amplitude waves,
making it easier to read. Both bipolar and ipsilateral ear referential montages were used during
18
this study, as seen in Figure 3 below, with the Fp1, Fp2, F3, F4, A1, A2, O1, and O2 electrodes
highlighted.
Figure 3. Bipolar and referential montages with primary electrodes chosen for sound file
highlighted
Specifically, a bipolar (double banana) montage is particularly useful at blink detection
so it takes readings of Fp1-F3 and Fp2-F4 and a referential montage (ipsilateral ear) of channels
aids in alpha rhythm detection using readings of O1-A1 and O2-A2. Therefore, these eight
channels were picked off for further analysis. When the raw signals from EDF Browser using the
bipolar and referential montage were read as code into MATLAB, raw, unfiltered data for each
of the eight electrodes was obtained, as seen in Figure 4.
19
Figure 4. Raw data using bipolar and referential (ipsilateral) montages in MATLAB
Filtering
The montages were used for further exploration and filtering to remove extraneous noise.
In particular, a Butterworth filter, also known as a maximally flat magnitude filter, was used
because it is a signal processing filter that has the flattest frequency response possible in the
passband, as seen in Figure 5. Therefore, undesired frequencies that are too high and too low are
removed, leaving a uniform sensitivity for the desired frequencies.
20
Figure 5. Butterworth (BandPass) Filter applied to the bipolar and referential ipsilateral
ear montage in MATLAB
Resampling, Modulation, and Chord Selection
The bandpass filtered signals were then resampled at a higher frequency (sped up) of
44,100 Hz, as seen in Figure 6.
21
Figure 6. Bipolar and referential ipsilateral ear montage filtered and resampled at 44,100
Hz
Then, a wave file was written out for the resampled signal. Since this frequency is too high to be
heard by the human ear (humans can only hear between 20 – 20,000 Hz), these EEG signals
were modulated into the auditory range, using the tones of A3 (frequency of 220 Hz), A4
(frequency of 440 Hz), C5 (frequency of 523.25 Hz), and E5 (frequency of 659.25 Hz), thus
creating an A-C-E chord. These specific frequencies were chosen, as they seemed to produce a
harmonious sound to the ear.
22
Figure 7. Bipolar and referential ipsilateral ear montage filtered and modulated at 44,100
Hz with 16 bit
New files were written out, as seen in Figure 7, and a digital audio workstation known as Adobe
Audition was used to mix the sound files. To do so, the newly written out and modulated EEG
signal files were read into the software. A sampling rate of 44,100 with a bit depth of 16 bits,
was selected under the Multitrack options. Then, the .wav files were dragged into tracks 1
through 4. The volume for each track was adjusted to where all tracks were equal. Then, all
tracks except the Master were deleted. Ultimately, this produces a sound file that can be listened
to for further analysis and playback, as seen in Figure 8.
23
Figure 8. Modulated EEG signals read into Adobe Audition
24
Chapter 4
This section discusses the results and conclusions obtained from this study.
Results
By using the Adobe Audition digital audio software, it was relatively simple to place the
code that was written to take the EEG signals from EDF Browser and place it through an
interface where waveforms could easily be entered and edited to produce the desired outcome.
This was particularly beneficial because having the capacity to work with multi-tracks enabled
the data from multiple electrode channels to be read and played simultaneously. The particular
workstation software made available through Adobe Audition has enabled real-time and nondestructive processing and also supplied plugins for noise reduction and FFT equalization. Thus,
as seen in Figure 7, the modulated signals were indeed able to translate the original blink
artifacts and alpha rhythms seen in the EDF Browser into the auditory spectrum through Adobe
Audition. It does not appear that any data was lost throughout this conversion process and excess
noise does not interfere with being able to identify when the blinks are occurring or when the
alpha rhythms are detected.
Therefore, it now possible to detect blinks and determine whether the eye is open or
closed simply by listening to appropriate EEG channels. This then sufficiently supports the test
of proof conducted to determine if sonification of visual data was indeed possible. Through these
methods, it can be seen that it is feasible to determine when the blink of an eye occurs through
listening to the sound produced when the appropriate channels are first converted into auditory
range frequencies through code and then transformed into an acoustic arena. By demonstrating to
the general public what this noise sounds like and training them to be able to identify this sound,
it would be possible for non-specialists to classify this type of brain activity when hearing the
25
correct pattern of noise. This then revolutionizes what trained physicians have traditionally been
able to classify only through visual techniques into something in audible form that can be offered
to non-specialist individuals to classify brain patterns.
Conclusions
In this study, the Adobe Audition Software was in fact able to keep the original blink
artifacts and alpha rhythms intact when the signals were read through as a sound file, as
evidenced by a pattern very similar to the one seen with the raw data from the EEG excerpt in
EDF Browser. Although the blinking of an eye is visually an easily detectable artifact in an EEG,
it is now possible to detect blinks and to determine whether the eye is open or closed simply by
listening to the correct EEG channels.
26
Chapter 5
This chapter discusses the potential prospective projects that will be further explored from the
basis of the results of this study to make additional discoveries in this area of sonification.
Future Work
By sonifying EEG signals, it appears that new insight into EEG analysis is now
achievable. Since the sonification of EEG signals was made possible by successfully being able
to “listen to the blink of an eye,” further exploration seeks to go beyond listening to and
assessing blinking towards the analysis of other situations. For example, it would be very
beneficial to be able to detect and classify various types of seizures through auditory methods.
Hopefully, this insight about brain signals will be helpful in the early detection of abnormalities
and thus the ability to take better care of patients. The ultimate goal is to enable specialists and
non-specialists to be able to recognize important features of an EEG, such as certain kinds of
seizures, acoustically.
Although significant progress has been made on the sonification of EEGs using clinical
data, the addition of a Cadwell Easy III EEG with 32 channels in the laboratory has enabled
further information to be gathered in order to record data to test the algorithms for sonification
that are currently being developed. Due to the fact that initially the algorithms to convert EEG
data to sound will be issue-specific, the acquisition of real EEG data is crucial. It is critical to
find the means by which information from the EEG machine can be obtained, analyzed, and
converted into sound in real-time for convenient real-world applications. For this reason, a
particular topic of interest to this study would be electrooculography, the monitoring of the
subset of the electrodes of the EEG that show eye and eyelid position (Seaquist, 2014). For this
work, a volunteer human subject would be required to have an EEG performed in the laboratory
27
while cues are given requesting that the subject perform various actions. These requests include
opening and closing his or her eyes, singularly and together, as well as blinking several times.
Therefore, more relevant data can be attained from the sonification of these signals.
The results of this study also introduce the question of using blink detection in work
related to the use of contact lenses. This would be beneficial because when rigid contact lenses
are placed on the surface of the eye, the eye blinking causes the rigid contact lens to move out of
place, which can cause irritation and the inability to see. Being able to detect the specific
movements of the rigid contact lenses when the eyes blink, and determining how to combat or
account for this movement, can allow for a more comfortable and the ideal fit for the wearer.
Therefore, there are several broad implications of this sonification of EEGs research, which can
be applied to different fields of study, furthermore than the current clinical diagnoses of brain
functions.
28
References
Baier, G., Hermann, T., & Stephani, U. (2007). Multichannel Sonification of Human EEG.
Proceedings of the 13th International Conference on Auditory Display, 491 – 496.
Retrieved from http://pub.unibielefeld.de/luur/download?func=downloadFile&recordOId=2017124&fileOId=2456079
Baier, G., Hermann, T., Sahle, S., & Stephani, U. (2006). Sonified Epileptic Rhythms.
Proceedings of the 12th International Conference on Auditory Display, 148 – 151.
Retrieved from
http://www.sonification.de/publications/media/BaierHermannSahleStephani2006SER.pdf
Barrass, S. & Kramer, G. (1999). Using Sonification. Multimedia Systems 7, 23-31. Retrieved
from https://ccrma.stanford.edu/courses/120-fall-2005/using_sonification.pdf
Benitez, R., Holland, S., Jordà, S., Marimon, X., Mealla, S., Oliveira, A., Steffert, T., &
Váljamäe, A. (2013). A Review of Real-Time EEG Sonification Research. International
Conference on Auditory Display. 85 – 93. Retrieved from
http://oro.open.ac.uk/38478/1/Väljamäe-Steffert-2013-ICADA%20REVIEW%20OF%20REALTIME%20EEG%20SONIFICATION%20RESEARCH.pdf
Campo, A., Hoeldrich, R., Eckel, G. & Wallisch, A. (2007). New sonification tools for EEG data
screening and monitoring. Proceedings of the 13th International Conference on Auditory
Display, 1-7. Retrieved from
http://iem.kug.ac.at/fileadmin/media/iem/altdaten/projekte/publications/paper/new/new.p
df
29
Francis & Loughhead. (1984). Bell’s Phenomenon: A Study of 508 Patients. Australian Journal
of Ophthalmology, 15-21. Retrieved from
http://onlinelibrary.wiley.com/store/10.1111/j.1442-9071.1984.tb01119.x/asset/j.14429071.1984.tb01119.x.pdf;jsessionid=3F85E3D3BB77324A61CD828C5144E2A5.f04t01
?v=1&t=i5qigoe9&s=bd30facece6bb8cc99cecc147a41c0a4ab40bab5
Hermann, T., Meinicke, P., Bekel, H., Ritter, H., Muller, H.M., Weiss, S. (2002). Sonifications
for EEG Data Analysis. Proceedings of the 2002 International Conference on Auditory
Display, 1-5. Retrieved from
http://www.icad.org/websiteV2.0/Conferences/ICAD2002/proceedings/22_Thomas_Her
mann_EEG.pdf
Hinterberger, T. & Baier, G. (2004). POSER: Parametric Orchestral Sonification of EEG in Real
Time for the Self-Regulation of Brain States. Proceedings of the International Workshop
on Interactive Sonification, 1-6.
Khamis, H., Mohamed, A., Simpson, S. & McEwan, A. (2012). Detection of temporal lobe
seizures and identification of lateralisation from audified EEG. Clinical Neurophysiology,
123 (9). 1714-1720. doi: 10.1016/j.clinph.2012.02.073.
Mills, M. (2012). Do Signals Have Politics? Inscribing Abilities in Cochlear Implants. The
Oxford Handbook of Sound Studies. New York, New York: Oxford University Press.
Prandoni, P., & Vetterli, M. (2008). Signal Processing for Communications. Lausanne,
Switzerland: EPFL Press. Retrieved from http://www.sp4comm.org/docs/sp4comm.pdf
Purves, D. (2008). Neuroscience. Sunderland, MA: Sinauer Associates.
Schomer, D.L. & Lopes da Silva, F. H. (2010). Niedermeyer’s Electroencephalography: Basic
30
Principles, Clinical Applications, and Related Fields. Philadelphia, PA: Lippincott
Williams & Wilkins.
Seaquist, Carl. (2014). Sonification of Electroencephalograms IRB. Lubbock, TX.
Tatum, W.O., Husain, A.M., Benbadis, S., & Kaplan, P.W. (2008). Handbook of EEG
Interpretation. New York, New York: Demos Medical Publishing.
Trevisan, A. A., & Jones, L. (2010). A Low-end Device to Convert EEG Waves to Music. 128th
Convention of the Audio Engineering Society. 1-7. Retrieved from
http://menteheadband.com/wp-content/uploads/2014/02/Low-end-device-that-convertsEeg-to-Music.pdf
31
Appendix A
MATLAB Code
% How to get started:
% load EEGLAB
% >eeglab;
% Now go to EEGLAB Menu and pull down from the
% File Menu --->Import Data-->Using EEG functions and Plugins-->
%
From EDF/EDF+/GDF files (BIOSIG Toolbox)
%
When prompted for file click on subject.edf...A box will
%
appear...hit return to allow defaults to be in force. A new box
%
will appear...name what you will.
% Pull down Plot-->Channel data (scroll) to look at data.
%
%%
eegdata=EEG.data; % Save all channels from subject.
save('subject1eegdata.mat','eegdata','-mat'); % From now on loaddata from
this file
%% If data is saved start here.
clear;
load('subject1eegdata.mat')
%%
% Channel
% Event :
% Fp1
:
% Fp2
:
% F7
:
% F3
:
% Fz
:
% F4
:
% F8
:
% T3
:
% C3
:
% Cz
:
% C4
:
% T4
:
% T5
:
% P3
:
% Pz
:
% T4
:
% T6
:
% O1
:
% O2
:
% A1
:
% A2
:
% others:
% read in all eegdata(channel #, time frames)
Numbers
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23-43 (Pg1, Pg2, Oz, X1-X8, DC1-DC8, OSAT, PR)
32
% Recall that odds are left side and evens are right side.
%
%
%
%
%
%
%
For this study Looking at blinks and alpha rhythmes we are interested in
bipolar montage (double banana) of channels Fp1-F3 and Fp2-F4 for blink
detection and
in referential montage (ipsi ear) of channels O1-A1 and O2-A2 for alpha
detection. (See page 185 of Niedermeyer's Electroencephalography, 2011.)
Thus we want to pick off
Fp1:2, F3:5, Fp2:3, F4:7, O1:19, A2:22, O2:20, and A1:21
% For this study we are interested in the blinks and alpha
% in the 17 seconds from 79secs to 96 seconds.
%% Pick off data
%
%
channels=[2 5 3 7 19 21 20 22];
% Channels of interest and names
channelnames={'Fp1';'F3';'Fp2';'F4';'O1';'A2';'O2';'A1'};
montagenames={'Fp1-F3';'Fp2-F4';'O1-A1';'O2-A2'}; % Montages of interest
freq=512;
% frequency of sampling of original data
starttime=60+19;
% start time of interest: 1 minute 19 seconds
endtime=60+36;
% end time of interest:
1 minute 36 seconds
istrt=freq*starttime;
iendt=freq*endtime;
newdata=eegdata(channels,istrt:iendt);
%% plot the raw data to compare with the EDF file to makesure correct data is
being used.
figure('Name','Original data e unfiltered: 1:19-1:36; i.e., 17secs');
ticks={'0';'1';'2';'3';'4';'5';'6';'7';'8';'9';'10';'11';'12';'13';'14';'15';
'16';'17'};
for i=1:18
ta(i)=(i-1)*freq+1;
end
for i=1:8
subplot(4,2,i); plot(newdata(i,:),'black');
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(channelnames(i));
end
%% Create the actual signals to be studied
Fp1mF3=newdata(1,:)-newdata(2,:);
Fp2mF4=newdata(3,:)-newdata(4,:);
O1mA1=newdata(5,:)-newdata(6,:);
33
O2mA2=newdata(7,:)-newdata(8,:);
%% Plot the data with correct montage...should again compare with original
EDF
figure('Name','Montage (Double Banana + Ipsi Ear) Unfiltered')
subplot(2,2,1); plot(Fp1mF3,'black')
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(montagenames(1));
subplot(2,2,2); plot(Fp2mF4,'black')
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(montagenames(2));
subplot(2,2,3); plot(O1mA1,'black')
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(montagenames(3));
subplot(2,2,4); plot(O2mA2,'black')
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(montagenames(4));
%% At this point we might wonder how to filter the data.
%
%
%
%
%
%
One approach might be to use the EEGLAB menu and use (Under the Edit Pull
down) Select Data. This will prompt for Time Range, e.g. [79 96] and
Channel range, e.g., 2 3 5 7 15 17 19 20. The resulting file/dataset can
be written out to file...I used fselectnewdata.
I then thought I would try to run an order 5 highpass butterworth filter
at 1Hz.
x(1,:)=Fp1mF3;
x(2,:)=Fp2mF4;
x(3,:)=O1mA1;
x(4,:)=O2mA2;
%% High Pass Filter over 1Hz
34
[zh,ph,kh]=butter(5,1/(0.5*freq),'high');
[sosh,gh]=zp2sos(zh,ph,kh);
Hd=dfilt.df2tsos(sosh,gh);
% hh=fvtool(Hd);
% set(hh,'Analysis','freq');
for i=1:4
y(i,:)=filter(Hd,x(i,:));
end
figure('Name','Montage (Double Banana + Ipsi Ear) High Pass Filter (1Hz)');
for i=1:4
subplot(2,2,i); plot(y(i,:),'black');
set(gca,'XTick',ta);
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(montagenames(i));
end
%% Low Pass Filter under 70Hz
[zl,pl,kl]=butter(7,70.0/(0.5*freq),'low');
[sosl,gl]=zp2sos(zl,pl,kl);
Ld=dfilt.df2tsos(sosl,gl);
% h=fvtool(Ld);
% set(h,'Analysis','freq');
for i=1:4
z(i,:)=filter(Ld,y(i,:));
end
figure('Name','Montage (Double Banana + Ipsi Ear) Band Pass Filter
(1Hz/70Hz)');
for i=1:4
subplot(2,2,i); plot(z(i,:));
set(gca,'XTick',ta);
set(gca,'XTickLabel',ticks);
xlabel('Secs');
ylabel('uV');
title(montagenames(i));
end
35
%% Resample at 44100
fsample=44100;
for i=1:4
v(i,:)=resample(z(i,:),fsample,freq);
end
lth=(endtime-starttime)*fsample;
ibg=22;
for i=1:4
w(i,:)=v(i,ibg:ibg+lth-1)/200.0;
end
%% Plot Resampled
figure('Name','Montage (Double Banana + Ipsi Ear) Filtered and Resampled at
44,100Hz');
for i=1:18
ta(i)=(i-1)*fsample+1;
end
for i=1:4
subplot(2,2,i); plot(w(i,:));
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
title(montagenames(i));
end
%% Write out wave file of modulating signal
for i=1:4
a=['xFnewa',char(montagenames(i)),'mod.wav'];
audiowrite(a,w(i,:),44100)
end
%% Set up cords and x
clear x
clear y
clear z
sample=44100;
interval=2^(1/12);
36
A3=220;
A4=440;
C5=440*interval^3;
E5=440*interval^7;
freq(1)=A3;
freq(2)=A4;
freq(3)=C5;
freq(4)=E5;
l=length(w(1,:))/sample;
x=0:(1/sample):l-(1/sample);
%% Build tones
for i=1:4
m(i,:)=sin(2*pi*x*freq(i));
end
%% Modulate tones with EEG signals
for i=1:4
z(i,:)=m(i,:).*w(i,:);
end
%% Plot Modulated
figure('Name','Montage (Double Banana + Ipsi Ear) Filtered and Modulated at
44,100Hz');
for i=1:18
ta(i)=(i-1)*fsample+1;
end
for i=1:4
subplot(2,2,i); plot(z(i,:));
set(gca,'XTick',ta)
set(gca,'XTickLabel',ticks);
title(montagenames(i));
end
%% Write files
for i=1:4
a=['yFnewa',char(montagenames(i)),'tones.wav'];
audiowrite(a,z(i,:),44100);
37
end
%% Now use Audition to mix the sound files
%
% Fire up Audition and read in the yFXXXXX files. These are the ones that
% are the original EEGs that have been modulated.
% You can listen to individual ones by selecting and playing.
%
% Select Multitrack (Upper left corner) and fill in box with name, file to
% same in, and select a sample rate of 44100 and bit depth of 16 bits.
%
% Drag the .wav files into tracks 1 to 4.
%
% You can adjust volume of each track separately by using 'volume' knob on
the
% far left of each track. Try to keep all tracks in the Green.
%
% Delete other tracks except Master
%
% Adjust balance on each track and volume on master. This is easier if you
% loop the sound and hit mixer tab
38
Download