Paper Title (use style: paper title)

advertisement
Non-Invasive EEG Feature extraction and
recognition of epileptiform activity using Adaptive
resonance theory
Aswathappa.P 1 , S.Ranjitha2 Dr .H. N. Suresh3
3
Abstract : The recent research in artificial neural networks
(ANN) have provided us many information about a natural
learning strategy. Artificial neural network consists of densely
interconnected simple neuron like elements. Information is
represented in the strengths of the connections (synapses)
between elements and is processed in parallel. Different neural
network models are specified by the topology of connections,
characteristics of processing elements and training rules. The aim
of this study is to develop a simple model for characterization of
Epileptical seizure. Emerging Adaptive Resonance Theory
(ART2 Neural Network) are used in this study to detect and
generate epileptical seizure.
Keywords: Adaptive Resonance Theory,
networks (ANN, epileptical seizure.
artificial
neural
ASWATHAPPA P
Asso. Professor
Dept. of IT, BIT Bangalore
I.INTRODUCTION:
In this study, adaptive resonance theory (ART 2) neural
network are investigated for on-line unsupervised recognition
(spikes for epilepsy monitoring) and also epileptical seizures
are generated from output nodes of the network. ART 2
networks are self organizing systems that clusters data into
different classes. Learning is completed when these classes are
labeled by an expert and are saved in a look-up table.
Recognition of new data is accomplished by finding the class
assigned by ART 2 and searching through the table to get the
proper labels. Unrecognized inputs are put into a new class for
future labeling. In this study an ART 2 neural network with 32
inputs was developed and trained using EEG data containing
spike and non-spike waveforms. For comparison a 32 input
multilayer perceptron is also constructed and evaluated for
comparison.
Aswathappa.P 1
1
Research scholar, Dept. Of ECE, Karpagam University,
Coimbatore, Tamil nadu, , India.
paswathappa@yahoo.com1
S.Ranjitha2
Final BE(ECE), Bangalore Institute of Technology,vv puram
,Bangalore-04,India
2
Dr .H. N. Suresh3
Professor, Bangalore Institute of Technology, Dept.of
Elecronics and Instrumentation Engg.,
Bangalore04,Karnataka,India
hn.suresh@rediffmail.com, 2
Evaluation with three sets of EEG patterns indicates that the
ART 2 Neural Network's performance is better than that of
multilayer perceptron showing its high potential. Considering
that ART 2 can be trained with one or few iteration (compared
to thousands required for back propogatation networks), hence
this can used for on line trainining. Such systems are required
for long-term EEG monitoring due to large variations of spike
waveforms that occur among individual patients during long
recording time periods.
Many researchers tasks include the recognition and
generation of abnormal electrical wave patterns. Traditional
algorithmic methods are insufficient to characterize epilepsy
seizure. Medical diagnosis is one of the prime examples of this
problem. Although many expert systems or so called
"intelligent" devices have been developed using traditional
Artificial Intelligence (AI) approaches. Intelligent systems are
developed using symbol and rule based AI techniques. In
essence, traditional AI requires the existence of a domain
expert who knows the answer to many questions and can
specify the rules to a knowledge engineer. This is however,
hardly the case in medicine since most of our knowledge is
incomplete. This research focuses on the problem of
epileptical EEG spike detection for the diagnosis and to
construct an ART memory to evaluate epilepsy.
Although many rules are proposed for the detection of
spikes (e.g. width, amplitude, sharpness, etc.) they are not up
to the satisfaction. Many EEG technicians are not aware of
these rules consciously, yet they can identify spikes in
ongoing EEG easily and accurately. Humans learn to
recognize EEG spikes by seeing a number of such waveforms
and generalizing their characteristics. Better success can be
expected if a natural learning strategy is used by humans and
adapted in machines. Multilayer perceptrons trained with back
propagation (BP) [1] are the best known and most utilized NN
models. Pattern recognition and classification by neural
networks have several major advantages over traditional
methods. Neural networks use massively parallel, nonalgorithmic processing which are in some ways similar to
those used by the human brain. As a result, many competitive
hypotheses have been explored simultaneously. Neural
networks store the patterns presented to them by modifying
their states. Pattern elements are not memorized individually
hence a large numbers of complex patterns can be stored.
Page 1 of 6
Neural networks have other advantages over current
traditional systems that they are fault-tolerant and relatively
insensitive to the underlying statistical distributions of patterns
to be recognized. Neural network applications in EEG have
been primarily focused on sleep staging, anesthesia
monitoring and transient event detections such as spikes and
K-complexes [2,3].
All these studies used multilayer
perceptrons to train various EEG waveforms. These studies
have shown that transient EEG events such as spikes and Kcomplexes can be successfully recognized [4]. These networks
are trained off line with a fixed set of labeled data using the
backpropagation algorithm and do not adapt to new conditions
during monitoring. Retraining with new patterns generally
degrades the performance of the network in recognizing
previously learned patterns. This property of the multilayer
perceptron results in a rigid, non adaptive system that cannot
be modified or customized.
To solve this problem new neural network architecture,
which adapts to new circumstances is needed. Once the
pattern recognition task has been completed then the network
behavior can be modified by new information. As new
patterns are added to the knowledge base (e.g., as data become
available from more patients) the network may adapt itself to
improve its recognition rate. Adaptive Resonance Theory
(ART) architecture was originally developed by Grossberg
and Carpenter [5]. The main thrust of this research is to
develop on line adaptive methods for the detection of EEG
spikes using ART neural networks. A brief description of
ART neural networks is presented in this section. We have
developed techniques for the detection of EEG spikes. These
techniques can be used for long term monitoring of seizure
patients by continuously training the system to recognize new
waveforms on line. After development of these techniques,
this can be used for the detection of other transient EEG
waveforms and biomedical signals.
input patterns. This type of Neural Network architecture has
the ability to learn new patterns in one or a few iterations
without forgetting the old ones (learned patterns). This
property of adapting to new information makes ART 2 NN
ideal for on line training with individual patients for long term
monitoring, since EEG waveforms obtained from each patient
are different. Originally two ART networks, ART l and ART
2, have been developed. ART 1 and ART 2 differ essentially
in the nature of input patterns. ART l network works only
with binary inputs while ART 2 works with analog patterns or
binary inputs. In this research we are using ART 2 networks
since EEG spike patterns are analog in nature. A detailed
diagram of the ART 2 model used in this study is shown in Fig
2.1 .A short description of the network is given below. A more
detailed description of the network can be found in [6]. As
seen in Fig. 2.1, ART2 neural network consists of two major
parts attentional subsystem and orienting subsystem.
Orientingsubsystem and Attentional sub system
II. ART 2 Neural Network
Architecture
ART neural networks are unsupervised, self organized,
and self stable types of NN architecture. ART is the type of
architecture that remains plastic (adaptive) in response to new
input patterns and yet remains stable, i.e., it preserves its
previous learned patterns. This Neural Network models
function as follows. First an input pattern is given to the
network on a feed forward (bottom-up) basis, which produces
a feed backward pattern (top-down) of activations, over one of
the existing node classifier. These feed backward and feed
forward patterns are then compared. If a match occurs
(resonance) within specified limits (vigilance), then this
pattern is classified as a member of that winning node. If not,
the remaining nodes are searched, failure to find an
appropriate node results in the formation of a new node for
this new pattern.
ART type neural networks are particularly suited for
pattern classification problems that cannot be defined by static
Fig. 2.1 - ART2 system diagram
The attentional subsystem consists of F1 and F2 layers. All the
nodes in F1 and F2 are fully interconnected with each other and
the weights associated with them are in the form of long-term
memory (LTM). F1 layer is composed of many sub layers
whose interconnections are depicted in Fig. 2.1.All the
processing elements in ART 2 models are governed by the
neuronal equation.
€ d ( X k ) / dt = - A Xk + ( 1 - B Xk ) Jk + - ( C + D Xk ) J k (2.1)
Where, Xk = activation function of the kth unit,Jk + =
excitatory input to the kth unit ,Jk- = inhibitory input to the kth
unit ,Under steady state (asymptotic) conditions (using B = C
= 0 for ART 2) , this equation reduces to:
X k = J k + / ( A + D J k -)
(2.2)
Activation functions of each sub layer of F 1 and the orienting
subsystem are calculated using Table 5.1 followed by the
Page 2 of 6
interconnections of the nodes, shown in Fig. 1.1 .The
function f(x) is a noise filter which is described as
f(x) = 0
for 0 ≤ x ≤ Ө
( 2.3)
= x
for
x ≥Ө
where ‘ Ө ‘ is the noise factor which determines the degree of
suppression of noise.Bottom-up (F1 to F2) equation is given
by, T j = ∑ Pi Z ji
(2.4)
A contrast enhancement is used to calculate the activation
output g(yj) of the F 2 nodes g(yj)= d
for T j = Max (T k )
for all k = 0
otherwise
(2.5)
Tk does not include the node recently reset by the orienting
system and d is a constant determined experimentally. The
weights z ij and z ji, corresponding to the interconnections
between y and p nodes form the LTM of the system are
updated as follows zij = zji = u / (1-d)
(2.6)
Only the weights corresponding to the winning node in F2
layer are updated as designated by the J. G node is a gain
control unit and is in-charge of sending an inhibitory signal to
each unit on the layer it is connected. This gain control signal
simulates the ‘2/3’ rule used in ART 1 architecture.
III. STM and LTM Equations and
Synaptic Plasticity
The STM activity xk of any node vk in F1 or F2 obeys a
membrane equation of the form
 d(xk) /dt= -xk + (1-Axk)Jk+ - (B+Cxk)Jk-
(3.1)
Jk+
Where
is the total excitatory input to vk . Jk- is the total
inhibitory input to vk and all the parameters are non negative.
If A>0 and C>0, then the STM activity xk(t) remains within
finite interval [-BC-1 to A-1] no matter even for large values
of nonnegative inputs J k+ and Jk-..We denote nodes in F1 by vi
where i =1,2…M. We denote nodes in F2 by vj where
j= M+1, M+2…..N. Thus,
 d(xi) / dt = -xi + (1-A1 xi)Ji+ - (B1 +C1 xi)Ji- -- (3.2)
 d(xj)/dt= -xj + (1-A2 xj)Jj+ - (B2 +C2 xj)Jj-
(3.3)
The notation is, the F1 activity pattern x = (x1, x2,…., xM) and
the F2 activity pattern
y = (x M+1, xM+2,.., xN).The
input Ji+ to the ith node vi of F1 is a sum of the bottom-up input
Ii and the top-down template input vi
vi = D1  f(xj) zji
And
Ji+
(3.4)
= Ii + vi
Where f (xj) is the signal generated by activity xj of vj and zji is
the LTM trace in the top-down pathway from vj to vi. In the
notation, the input pattern I=(I1, I2, …, IM), the signal pattern
u=(f (xM+1), f (xM+2), , f(xM)), and the template pattern v=(v1,
v2,…, vM). output patterns are described by S=(h(x1), h(x2), ,
h(xM)) and input pattern T= (T M+1, TM+2, ,TM). In this study the
model system is recurrent networks of excitatory and
inhibitory type of integrate and fire neurons. Excitatory
neurons are connected by plastic synapses having two values
of
potentiated and depressed efficacy (PSP), which are
assumed to be stable on the long time scale. Long term
synaptic dynamics, i.e transitions between these two states is
driven by neural pre synaptic and post synaptic activities
through a fast, spike driven dynamics which affects the
internal variability of the synapse. Up or down regulation of
the internal variable occurs if a pre-synaptic spike comes
within (beyond) t. When the internal variable crosses a
threshold from below, the efficacy gets potentiated or
depressed. Because of the stochasticity of the spikes in the
network (due to features like the random pattern of
connectivity) such synaptic transitions are themselves
stochastic, and are characterized by a frequency dependent
‘transition probability’. Such a synaptic device, driven by
spike timings, acts as pre- and post- synaptic neurons, and
implements a rate dependent memory. For a network with N
neurons the analysis of the coupled neural synaptic dynamics
requires to solve a system of O (N+N2) coupled equations.
Standard approaches to the numerical integration of those
equations imply a prohibitive O (N3) computational load.We
therefore use the event driven approach to stimulate the
complexity of O(N2). On the other hand, predictive, analytical
tools are needed in order to choose interesting regions in the
huge parameters space of the neural and synaptic dynamics.
As long as the characteristic times of synaptic changes are
much longer than those neural dynamics, it still serves as a
precious tool in the present scenario to analyze the working
memory. Assuming that synapses undergoing potentiation and
depresses homosynaptically, a compact representation can be
deviced, exposing in the plane (JP and JD ) of the stable
collective states predicted by MFT . We will be seeking
learning trajectories bringing the network from a totally
unstructured synaptic state (the lower right corner in the plane)
to a state supporting a low rate. An unspecific “spontaneous”
state coexisting with a selective higher rate state is interpreted
as neural substrate of a “working memory”. To take STD into
account, we adopt the model introduced in [7,8] and calculate
the corrections to the mean and variance of the apparent
currents (the needed ingredients for MFT) due to inclusion of
STD [9].
The role of LTD and STD with feasibility, it is clear that
despite the wide region allowing for coexistent spontaneous
and selective activity, its learning trajectories have to cross
from below, has a very steep rise with respect to J p (Jp is
candidate source of instability). If the increase of J p is not
adequately balanced by increased depression, the spontaneous
activity is easily destabilized. The stabilizing role of the LTD
is such that the lower Jp /Jee (higher depression) the wider the
regions of the synaptic space in which the spontaneous and
selective activities coexist, and the region of instability
shrinks. The inclusion of STD, the spontaneous activity is
virtually unaffected and is kept stable in much wider region of
the plane while the rates of the selective states have a very
smooth dependences on Jp,
suggesting that, simulations
confirm on line learning, the ongoing neural synaptic
dynamics(diamonds in the plot) meet the above expectations.
Keeping spontaneous activity stable during learning implies
that synapses driven by low neural activity should not change.
Page 3 of 6
During learning a sufficiently intense synaptic long term
depression should take place, in order to compensate for
synaptic potentiation [10].
A multilayer ART 2 model with
varying number of input nodes and output nodes using neural
network simulator running with MAT Lab and GENESIS
simulator is used. During training small amount of noise is
added to the original training input to create large number of
slightly different patterns. During the ID thousands of the inter
neurons synchronously undergo an unusual large
depolarization on which burst of action potential is
superimposed and neuronal inhibition occurs in areas
surrounding the focus [11,12].
IV. Training and Testing
Procedure for ART 2 Network:
Since the ART2 network is essentially an unsupervised
learning system, the output clusters have to be labelled using
some kind of an external event or supervision. Such labeling
can simply be done by a trainer who assigns desired classes to
the different clusters as shown in Fig. 4.1. In the training
phase, the system trainer assigns class names (spike vs. nonspike) to clusters and a node classification table is generated.
In the testing phase, this table is used to classify the node
generated by ART2. New nodes generated by ART2 during
the testing phase are labeled as unclassified nodes.
are used for training the network. Each EEG waveform was
included in only one set in order to ensure that testing was
performed on new waveforms not used during learning. All
these networks were trained by the sets of EEG data and tested
with all other standard three sets of data. Percentage correct
classification over the whole test sets gave the overall success
of the network in comparison to expert ART network. In order
to evaluate the potential of the unsupervised ART2 network,
the same data were evaluated by using the standard back
propagation (BP) neural network. This BP network consisted
of 32, 8 and 2 nodes in the input, hidden and output layers.
V. Results
The clustering of a set of EEG data containing 30 spike
and 30 non-spike patterns is used for training the network. As
seen, the network is iterated to produce 8 nodes for spikes and
15 nodes for non-spikes. As expected more nodes were
generated for non-spikes since they can be of any shape
observed in EEG. Surprisingly it was also observed certain
seizure waveforms at the output nodes of the constructed
ART2 module iterated with long term synaptic memory as
illustrated in Fig. 5.2 to Fig. 5.3. The number of nodes and the
accuracy of the network is strongly influenced by the vigilance
factor ( ρ ) and to a lesser degree by the noise factor ( Ө ). In
this study, we have chosen a value for the vigilance factor by
experimentally incrementing it in small steps and scanning
through an estimated range. The noise factor is recommended
to be set at l/√n where n is the number of input nodes.
Generally, for higher Ө, more nodes are generated and
accuracy goes down. On the other hand, for lower Ө, fewer
nodes are generated but accuracy goes down. Also, the
network becomes unstable. We have determined Ө = 0.223.
The overall results comparing the supervised network (BPNN) with the unsupervised network (ART2-NN) are shown in
Table 5.1 and 5.2. ART 2 neural network generally produce
lower accuracies since these networks are not trained under
supervision. However, the ART 2-NN results were very close
to the performances of the BP-NN nets indicating the high
future potential of this approach.
Fig.4. 1. Training and testing procedures using
unsupervised neural architecture.
The ART neural network system described above is
implemented. EEG data is recorded from epileptic patients
undergoing seizure treatment were used for training and
testing. Digitized temporal sequences (32 points) of EEG data
were organized in three sets. Each set consists of an equal
number of spike and non-spike waveforms as determined by
the standard EEG data.
These sets consists of 60 and 150 EEG waveforms
obtained from 22,24 and 26 patients,
(ASC 60/22,
ASA150/24, ARA 150/26 are named as data files) respectively
Page 4 of 6
and accuracy level achieved are clinically acceptable and
comparable to those of human oral observations.
Fig. 5.2 Detection of spikes(S) and artifacts (A)
Fig. 5.3 Linear combination for spike and wave detection
1
Fig. 5.1 Spike and non-spike wave patterns generated by
ART 2 at its output nodes.
Table: 5.1 Spike and wave detection
VI. Discussions and Conclusion
It has been concluded that the ART 2 neural network
model reproduces many of the observed activities of
epileptical seizure. Epileptical seizure data can be recalled and
its characterization can be studied using ART 2 networks. We
have outlined a systematic methodology, which reduces the
construction of an iterative neural network, which closely
models time series data from a given chaotic system, to a
relatively automatic process. The dependence of the number of
nodes and classification accuracy on the vigilance factor and
threshold noise gives an opportunity for the user to optimize
the network by fine tuning these parameters during online
testing and thus ART 2 module can itself act as a memory to
generate certain observed behaviors of epileptic disorders. The
added stability and plasticity feature of ART 2 networks, and
the speed at which network converges has proven that, this
method of constructing working memory to characterize
epileptical EEG seizure is best compared to other conventional
neural network algorithms.
References:
Table 3.2 Performance comparisons of supervised BP and
unsupervised ART 2 NN.architectures using the training
and testing files.
Performance is measured in percentage. Testing of BPNN with ASC and ASA gave about 90% accuracy. This
performance, however, dropped to about 80% when an
inflexible set of data (ARA) containing waveforms from
many new patients was tested. As previously described,
performance of the networks are generally lower when they
are subjected to totally new waveforms from new subjects.
Overall, the test results indicate that it is possible to automate
spike and seizure detection of epilepsy with ART2 network
1. Victor, J. D., and Canel, A., “A Relation Between the Akaike Criterion and
Reliability of Parameter Estimates, with Application to Nonlinear
Autoregressive Modelling of Ictal EEG”, Annals of Biomedical Engineering,
vol. 20, pp. 167-180, pp. 167-180, 2011
2. Wang J., and Wieser, H. G., “ Regional Rigidity of Background EEG
Activity in the Epileptogenic Zone, Epilepsia”, vol. 35 (3), 2004, pp. 495-504.
3. Webber, W. R. S., Litt B., Wilson K., and Lesser R. P., “ Practical
Detection of Epileptiform Discharge (ED's) in the EEG using an Artificial
Neural Network: A Comparison of Raw and Parameterized EEG data”,
Electroencephalography and Clinical Neurophysiology, vol. 91, pp. 194-204,
2014.
4. Wendling F., Shamsollahi, J. M., Bellanger J. J., “ Time-Frequency
Matching of Warped Depth-EEG Seizure Observations”, IEEE Transactions
on Biomedical Engineering, vol. 46, no. 5, pp. 601-605, May 1999.
5. Werman M., and Peleg, S., “Min-max Operators in Texture Analysis”,
IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 7, pp.
730-733, 2010.
Page 5 of 6
6. Williams, W. J., Zaveri, H. P., and Sackellares, J. C., “ Time-Frequency
Analysis of Electrophysiology Signals in Epilepsy”, IEEE Engineering in
Medicine and Biology, Mar/April, 2009, pp. 133-143.
7. Witte, H., Eiselt, M., Patakova, I., Petranek, S., Griessbach, G., Krajca, V.,
and Rother, M., “ Use of Discrete Hilbert Transformaion for Automatic Spike
Mapping: a Methodological Investigation”, Medical & Biological Engineering
& Computing, vol. 29, pp. 242-248, 2010.
8. A.Barreto, J.Andrian, N.Chin and J.Riley (2011). “Multiresolution
characterization of interictal spikes based on a wavelet transformation”.
Wellesley-Cambridge Press, Wellesley MA.
9.. A.Graps (1995). “ An introduction to wavelets” IEEE Computational
Science and Engineering,Los Alamitos, USA.
10.E.Niedermeyer and F.L.Silva (2008). “Electroencephalography:Basic
principles, clinical applications and related fields”. Williams and
Wilkins,Baltimore, Maryland, USA.
11. G.Strang, T.Nguyen (2009). “Wavelets and Filter Banks”. WellesleyCambridge Press, Wellesley.
12. J.M.H.Karel, R.L.M.Peeters and R.L.Westra (2005). “Optimal discrete
wavelet design for cardiac signal processing”. accepted, EMBC 2013.
First Author P.Aswathappa :
At present he is Working as a Research scholar, Dept. Of ECE,
Karpagam University,coimbotore,T.N,, India. Also he is working as
a Assistant Professor in the Dept. of IT, Bangalore Institute of
Technology, Bangalore Affiliated to Visveswaraya Technical
University. he is good exposure in the field of signal
processing, ,Image processing, etc. he got more 16 years of
Teaching experience in various field.
Dr. H.N. Suresh received his BE
(E&C) from P.E.S College of
Engineering, Mysore University,
Karnataka, India, in the year 1989
and completed his M.Tech (Bio
Medical Instrumentation) from
SJCE
Mysore
affiliated
to
University of Mysore., in the year
of 1996 and since then he is actively involved in teaching and
research and has Twenty six years of experience in teaching.
He obtained his PhD (ECE) from Anna university of
Technology.He worked at various capacities in affiliated
University Engineering Colleges. For Visveswaraya Technical
University and Bangalore University he worked as a Chairman
for Board of Examiners, and member of Board of Studies etc.
At present he is working as Professor and Research and PG
co-ordinator in Bangalore Institute of Technology, Bangalore
,Affiliated to Visveswaraya Technical University. He has good
exposure in the field of signal processing, Wavelet
Transforms, Neural Networks,Pattern recognition,Bio Medical
Signal Processing, Netwoking and Adaptive Neural network
systems. He has published more 45 research papers in the
refereed international journals and presented contributed
research papers in refereed international and national
conferences. He is a member of IEEE, Bio Medical Society of
India, ISTE,IMAPS & Fellow member of IETE.
Page 6 of 6
Download