constructing models for early chronic kidney disease detection and

advertisement
CONSTRUCTING MODELS FOR EARLY CHRONIC KIDNEY
DISEASE DETECTION AND RISK ESTIMATION
Ruey Kei Chiu, Fu Jen Catholic University, Taiwan
rkchiu@mail.fju.edu.tw
Renee Yu-Jing Chen, Fu Jen Catholic University, Taiwan
497746077@mail.fju.edu.tw
ABSTRACT
In this paper, we aim to develop a best-fitting neural network model which can be employed
to detect various severity levels of CKD based on the two different sets of influence factors as
the inputs for model learning and developing. Firstly, the influence factors of CKD which
most physicians employ in the computational formula are used as the inputs to three
preselected fundament neural network models to detect (classify) various levels of chronic
kidney disease. Secondly, the potential key influence factors of CKD which are identified by
reviewing the CKD relevant literatures as well as conducting expert interviews are used as
the inputs to three models as well. Three fundamental neural network models selected for
detection and comparison include back-propagation network (BPN), generalized feedforward
neural networks (GRNN), and modular neural network (MNN). Furthermore, three hybrid
neural network models which are modeled by embedding genetic algorithm (GA) into each
respective fundamental neural factor are employed for conducting CKD detection and
comparison. The comparison among network models are based the CKD detection
performance measured by the computation accuracy, sensitivity, and specificity in model
training and testing.
The results of model experiment show that almost all models employed in experiment may
gain near 100% accuracy in CKD detection in the training stage regardless of which sets of
influence factors used in model training. However, if the model testing is further conducted, it
is found that the network models with the inputs of influence factors of CKD used by
physicians employed in the computational formula always shows better detection
performance in all three aspects of measures. The BPN gain the highest 94.75% accuracy
measures in the testing stage among three fundamental neural network models while GFNN
gain only 86.63% in accuracy measure which is the lowest performance in three models.
Through further observations from experiment results, it is found the hybrid network model of
GFNN plus GA embedded may significantly enhance detection performance in all three
measures from its fundamental model in the testing although the the detection performance
are degraded in BPN and MNN.
We conclude that BPN might be the best-fitting factor Among three fundamental neural
network models employed in detecting CKD while the models with GA embedded in BPN and
GFNN respectively might be the best-fitting hybrid models for CKD detection.
Keywords: Chronic Kidney Disease, Disease Detection, Risk Estimation, Artificial Neural
Networks, Genetic Algorithms.
INTRODUCTION
As people’s life level become more and more modernized and life span become longer in our
society, chronic kidney disease (CKD) becomes more common which may result in
developing different levels of ill-function and damage of patient kidneys. Once a person
gets CKD, he/she will suffer from the disease which may decrease his/her working ability as
well as live quality. It is also very likely to develop other chronic diseases like high blood
pressure, anemia (low blood count), weak bones and cause poor nutritional health and nerve
damage. In the meantime, kidney disease increases patient risk of contracting heart and blood
relevant diseases. Chronic kidney disease may be caused by other chronic disease such as
diabetes, high blood pressure and other disorders. High CKD risk groups include those who
have diabetes, hypertension, and family history of kidney disease. Early detection and
treatment can often keep chronic kidney disease from getting worse. When kidney disease
progresses, it may eventually lead to kidney failure which even requires dialysis or even
proceeds to kidney transplant to maintain patient life.
According to the statistical data announced by the Department of Health of Taiwan’s
government in 2010 (DOH, 2010), the mortality caused by kidney disease has been ranked in
the 10th place in all causes of death in Taiwan and thousands of others are at increased risk.
The mortality caused from kidney disease is estimated as 12.5 in every 100,000 peoples. As a
result, it costs as high as 35 percent of health insurance budget to treat the chronic kidney
disease (CKD) patients with the age over 65 years old and end stage kidney disease patient in
all ages. It occupies a huge amount of expenditures in national insurance budget.
Regarding to the measurement of serious levels of CKD, presently glomerular filtration rate
(GFR) is the most commonly measuring indicators used in health institutions to estimate
kidney health function. The physician in the health institution can calculate GFR from
patient’s blood creatinine, age, race, gender and other factors depending upon the type of
formal-recognized computation formulas (Firman, 2009; NKF, 2002; Levey, 2005) is
employed. The GFR may indicate how health of a patent’s kidney and can be also taken to
determine whether a people has kidney disease. The physician may also depend upon this
indicator to determine the stage of kidney disease of a patient and help him to set a treatment
plan for the patient (Firman, 2009; Levey et al., 2005).
The objectives of this paper aims to develop decision models by applying artificial neural
networks and fuzzy expert system in artificial intelligence which attempts to provide
physicians an alternative method to detect chronic kidney diseases and estimate the risk
levels of a patient in early CKD stage. This study of this paper was cooperatively carried out
with a district teaching hospital in New Taipei City located in the northern district of Taiwan.
In the first stage of this paper, we aim to develop a best-fitting neural network model which
can be employed to identify various levels of CKD based on two different sets of influence
factors as the inputs for model learning and developing. To reach this develop objective,
firstly the set of influence factors of CKD which most physicians employ in the
computational formula to diagnose CDK are determined for being used as the inputs to three
preselected fundament neural network models. Next, another set of potential key influence
factors of CKD which are identified by reviewing the CKD relevant literatures as well as
conducting expert interviews are used as the inputs to three models as well. Three
fundamental neural network models selected for detection and comparison include
back-propagation network (BPN), generalized feedforward neural networks (GRNN), and
modular neural network (MNN). Furthermore, three hybrid neural network models which are
developed by embedding genetic algorithm (GA) into each respective fundamental neural
factor Are employed for conducting CKD detection and comparison. The comparison among
network models is based on the CKD detection performance measured by the measurement
of factor Accuracy, sensitivity, and specificity respectively in model training and testing. The
input data for the training, testing, and experiment of ANN models is collected from the
people’s health examination which is periodically carried out by collaborative hospital.
METHODS AND LITERATURES REVIEW
In this section, firstly the different methods of measuring CKD followed by two sets of
influence factors used as the inputs for network modeling are reviewed. Next, the
investigation of the criteria of CKD classification (i.e., detection) and risk evaluation of CKD
is conducted. The three different neural network models employed for detection and
comparison in this paper are surveyed and briefly introduced as well.
The Review of the Methods of Measuring Chronic Kidney Disease
The GFR is a most-commonly method used to measure kidney health function. It refers to the
water filterability of glomerular of people’s kidney. The normal value of measurement ought
to be between 90 and 120 ml/min/1.73m2. There are three common computation methods of
GFR, which are (1) Removing rate of 24-hour urine creatinine (i.e., Creatinine Clearance
Rate, Ccr) ; (2) Cockcroft-Gault formula (i.e., known as C-G formula) ; (3) Modification of
Diet in Renal Disease formula (i.e., known as MDRD) (Firman, 2009). The CKD is
categorized into five stages by making using of GFR to measure kidney function. Although
the course of the change from stage one to five may usually last for years, it sometimes may
enter into fifth stage pretty soon resulted in the necessity of dialysis or kidney transplant
(NKF, 2002; Levey et al., 2005).
Again, according to the guideline developed by the National Kidney Foundation’s Kidney
Disease Outcomes Quality Initiative (KDOQI) (NKF, 2002), an independent not-for-profit
foundation governed by an international Board of Directors in 2002 providing a new
conceptual framework for a diagnosis of CKD independent of cause and developing a
classification scheme of kidney disease severity based on the level of glomerular filtration
rate (GFR). The new system represented a significant conceptual change, since kidney
disease historically had been categorized mainly by cause. The diagnosis of CKD relies on
markers of kidney damage and/or a reduction in GFR. Stages 1 and 2 define conditions of
kidney damage in the presence of a GFR of at least 90 ml/min/1.73 m2 or 60 to 89
ml/min/1.73 m2, respectively, and stages 3 to 5 define conditions of moderately and severely
reduced GFR irrespective of markers of kidney damage. The summarized of this guideline is
shown in Table 1. The common definition of CKD has facilitated comparisons between
studies. Thus, this new diagnostic classification of CKD has likely been one of the most
profound conceptual developments in the history of nephrology. Nevertheless, there are
limitations to this classification system, which is by its nature simple and necessarily arbitrary
in terms of specifying the thresholds for definition and different stages. When the
classification system was developed in 2002 (NKF, 2002; Levey et al., 2005), the evidence
base used for the development of this guideline was much smaller than the CKD evidence
base today. Therefore, this guideline is constantly revised from then on by Kidney Disease
Outcomes Quality Initiative (KDOQI) with the modification and endorsement by Kidney
Disease: Improving Global Outcomes (KDIGO). The currently available methods to estimate
GFR and ascertain kidney damage are evolving (Levey et al., 2005).
Table 1: Classification of CKD as Defined by KDOQI and Modified and Endorsed by
KDIGO.
Stage Description
Classification by Severity
By GFR (ml/min/1.73m2)
1
Kidney damage with normal or GFR increasing
GFR >=90
2
Kidney damage with mild decreasing in GFR
GFR of 60-89
3
Moderate with decreasing in GFR
GFR of 30-59
4
Severe with decreasing in GFR
GFR of 15-29
5
Kidney failure
GFR <= 15 (or dialysis)
One of most recognized versions of the classification and risk evaluation of CKD was
released by National Kidney Foundation (NKF) of US written by Firman in 2009 (Firman,
2009). Its severity classification and risk evaluation criteria are shown in Table 2 and 3,
respectively. It is properly the most referable guideline regarding the classification and risk
evaluation of CKD.
Table 2: Classification and Risk Evaluation of Chronic Kidney Disease.
Stage
Description†
GFR (ml per minute Action plans
per 1.73 m2)
-
At increased risk
> 60 (with risk factors
Screening, reduction of risk factors for chronic
for chronic kidney
for chronic kidney
kidney disease
disease
disease)
Kidney damage
> 90
1
Diagnosis and treatment, treatment of comorbid
with normal or
conditions, interventions to slow disease
elevated GFR
progression, reduction of risk factors for
cardiovascular disease
2
Kidney damage
60 to 89
Estimation of disease progression
30 to 59
Evaluation and treatment of disease
with mildly
decreased GFR
3
Moderately
decreased GFR
4
Severely decreased
complications
15 to 29
GFR
5
Kidney failure
Preparation for kidney replacement therapy
(dialysis, transplantation)
< 15 (or dialysis)
Kidney replacement therapy if uremia is
present
Table 3: Risk Factors for Chronic Kidney Disease and Definitions
Type
Definition
Examples
Susceptibility
Factors that increase
Older age, family history of chronic kidney disease,
factors
susceptibility to kidney damage
reduction in kidney mass, low birth weight, U.S.
racial or ethnic minority status, low income or
educational level
Initiation
Factors that directly initiate
Diabetes mellitus, high blood pressure, autoimmune
factors
kidney damage
diseases, systemic infections, urinary tract
infections, urinary stones, obstruction of lower
urinary tract, drug toxicity
Progression
Factors that cause worsening
Higher level of proteinuria, higher blood pressure
factors
kidney damage and faster decline
level, poor glycemic control in diabetes, smoking
in kidney function after kidney
damage has started
End-stage
Factors that increase morbidity
Lower dialysis dose (Kt/V)*, temporary vascular
factors
and mortality in kidney failure
access, anemia, low serum albumin level, late
referral for dialysis
*-In Kt/V (accepted nomenclature for dialysis dose), "K" represents urea clearance, "t"
represents time, and "V" represents volume of distribution for urea.
In Taiwan, the Taiwan Society of Nephrology (TSN) also presented the self-detecting method
of kidney for the public. The MDRD formula (Levey et al., 2000; Levey et al., 2005), which
is recognized as a more accurate with mostly adopted method by kidney physicians to
estimate glomerular filtration rate (GFR) from serum creatinine, is adopted for the use in
diagnosis. Therefore, in this paper of study we also take MDRD to calculate the GFR. The
method of calculation formula is shown in formula 1. The input data for the GFR calculation
of each individual patient case in health examination are provided by the collaborative
hospital of this research. We also use the calculated results as the desired (targeted) value to
train our neural network models in experiment.
GFR = 186 × Creatinine−1.154 × age−0.203 (ml/min/1.73m2 )…….(1)
Note: For female the result should be multiplied by a factor of 0.742。
Snyder and Pendergraph (2005) indicated that the MDRD was suitable to be used to the
group of diabetes mellitus, older age (age above 51), patient with kidney replacement therapy
(dialysis, transplantation). It also takes less number of factors than other two popular methods,
which are Creatinine Clearance Rate computation and Cockcroft-Gault formula. It becomes
easier and simple but still maintains the result effectiveness for calculating GFR. Hence, this
paper adopts MDRD to conduct the calculation of GFR directly from the data collection of
creatinine, age, and sex. The three factors, which are ‘creatinine’, ‘age’, and ‘sex’, considered
in MDRD calculation also majorly compose the first set of input influence factors for
modeling neural network models.
Artificial Neural Network
An artificial neural network (ANN), usually called neural network (NN), is a mathematical
model or computational model that is inspired by the structure and/or functional aspects of
biological neural networks. Kriesel (2011) indicates neural networks are a bio-inspired
mechanism of data processing that enables computers to learn technically similar to
human-being brain. The generic structure of ANN is shown in Figure 1 (Nagnevitsky, 2005).
A neural network consists of an interconnected group of artificial neurons, and it processes
information using a connectionist approach to computation. In most cases an ANN is an
adaptive system that changes its structure based on external (input) or internal information
that flows through the network during the learning phase. Modern neural networks are
non-linear statistical data modeling tools. They are usually used to model complex
relationships between inputs and outputs or to find patterns in data
(http://en.wikipedia.org/wiki/Artificial_neural_network). It is properly the most prestigious
and adoptable model in all application models in the field of artificial intelligence.
Figure 1: A Generic Structure of ANN.
Input
layer
First
hidden
layer
Second
hidden
layer
Output
layer
There are many types of neural network models derived from the generic structure of ANN.
Three fundamental neural network models employed for the experiment in this paper are
briefly introduced in next sections
Back-Propagation Neural Network
The back-propagation neural network is aslo known as a "feed-forward back-propagation
network". It is a supervising learning multi-layer perceptron (MLP) neural network model
(Negnevitsky, 2005). A training set of input patterns is presented to the network. The network
computes its output pattern, and if there is an error - or in other words a difference between
actual and desired output patterns - the weights are adjusted to reduce this error. Since the
real uniqueness or 'intelligence' of the network exists in the values of the weights between
neurons, we need a method of adjusting the weights to solve a particular problem. It learns by
example, that is, we must provide a learning set (or called as training set) that consists of
some input examples and the known-correct output (desired output) for each case. So, we use
these input-output examples to show the network what type of behavior is expected, and the
back-propagation learning algorithm allows the network to adapt. Hence, this neural
network is called as back-propagation neural network (BPNN). A three-layer architecture of
BPNN is illustrated in Figure 2 (Negnevitsky, 2005).
Input signals
x1
x2
1
2
2
xi
i
1
y1
2
y2
k
yk
l
yl
1
wij
j
wjk
m
xn
n
Input
layer
Hidden
layer
Output
layer
Error signals
Figure 2: Three-layer Back Propagation Neural Network.
This model is properly the most popular and applicable model in artificial neural network
used in academic research and the development practical decision support applications in
terms of forecasting, classification, regression, data mining, and so on.
Generalized Feedforward Neural Network
The feedforward neural network was the first and arguably most simple type of artificial
neural network devised. In this network the information moves in only forward direction
which is simply from the input neuron data goes through the hidden neuron and to the output
nourons. There are no cycles or loops in the network where the data flow from input to output
units is strictly feedforward. The data processing can extend over multiple layers of neurons,
but no feedback connections are present. Its generalization model, which is konwn as
generalized feedforward neural networks (GFNN). It has the same architecture as
back-propagation neural network and also adopts the same learning methods to train the
network. The differences are that GFNN adopts generalized distributed-flow neurons in its
hidden layers and take perceptron as its inhibitory neuron. Arulampalam and Bouzerdoum
(2000) mentioned that the GFNN architecture used as the basic computing unit is a
generalized shunting neuron (GSN) model which includes as special cases the perceptron and
the shunting inhibitory neuron. GSNs were capable of forming complex, nonlinear decision
boundaries. This allows the GFNN architecture to easily learn some complex pattern
classification problems. Arulampalam and Bouzerdoum (2003) applied GFNNs to several
benchmark classification problems, and their performance was compared to the performances
of multilayer perceptrons such as a back-propagation neural network. Experimental results
show that a single GSN can outperform multilayer perceptron networks. An exemplified
architecture of GFNN is illustrated as Figure 3.
Generalized
Shunting
w00, c00 Neurons
Input Neurons
x0
Output Neurons
(Perceptorns)
wk0
‧
‧
‧
wm0, cm0
‧
‧
‧
o0
‧
‧
‧
xm
wjm, cjm
‧
‧
‧
‧
‧
‧
ok
wkj
xi
wji, cji
Excitatory Synapses
Inhibitory Synapses
Figure 3: The Generic Architecture of Generalized Feedforward Neural Networks.
Modular Neural Network
The modular neural network (MNN) was initially presented by Jacobs, et al. in 1991 (Jacobs
et al. 1991). The architecture of this network is shown in Figure 4. It is characterized by a
series of independent neural networks moderated by some intermediary. Each independent
neural network serves as a module like an expert and operates on separate inputs to
accomplish some subtask of the task the network hopes to perform. The intermediary takes
the outputs of each module and processes them to produce the output of the network as a
whole (Azam, 2000). The intermediary only accepts the modules’ outputs—it does not
respond to, nor otherwise signal, the modules. As well, the modules do not interact with each
other. One of the major benefits of a modular neural network is the ability to reduce a large,
unwieldy neural network to smaller, more manageable components. There are some tasks it
appears are for practical purposes intractable for a single neural network as its size increases.
The following are benefits of using a modular neural network over a single all-encompassing
neural network.
Expert 1
×
x
×
Expert 2
Σ
y
×
Expert k
g3
g2
g1
Gate
Figure 4: The Generic Architecture of Modular Neural Networks
(Source: http://en.wikipedia.org/wiki/Modular_neural_networks).
RESEARCH MATERIALS AND METHODS
Based on the literature reviews and expert interviews, this research takes three neural network
models, back propagation neural network, feedforward neural network, and modular neural
network, respectively to generate the detection model for chronic kidney disease. The input
data set for neural networks training and testing are collected from the population data of
health examination provided by collaborative hospital. The key influence factors input for the
each factor Are determined and verified from professional kidney experts and physicians. The
set of influence factors derived from GFR computation formulas such as MDRD is shown in
Table 4 and another set of key influence factors selected and determined in this paper as the
inputs of Neural Networks is shown in Table 5, respectively. The classification performance
of two sets input for model training and testing are compared in model experiment.
Table 4: The Influence Factors Derived from Formula Computation for GFR.
Factors
Description
Mean
Maximum
Minimum
Sex
0→Female;1→Male
---
---
---
Age
Integer
77.38
93
67
Creatinine
Real
0.88
4.28
0.37
Weight
Real
60.719
98
32
GFR Levels
0→Negative;1, 2, 3,4,5→Positive
---
---
---
measured in levels
Table 5: The Influence Factors Investigated Determined by This Paper.
Factors
Description
Mean Max. Min.
Creatinine
Real
0.88
4.28
0.37
Glucose (GLU)
Real
109.63
271
82
Systolic Pressure (SP)
Real
135.78
195
88
Protein in Urine (UP)
0→Negative
---
---
---
or called proteinuria
1→Trace; 2→+
---
---
---
3→++; 4→+++
Hematuria
Blood Urea Nitrogen (BUN)
Real
15.80
80.1
5.4
GFR-Level
0→Negative;1, 2, 3,4,5→Positive
---
---
---
measured in levels
The Learning Data and Preprocessing
The input learning data set for neural networks are collected and sampled from the cases of
the elderly people with age over 65 years old of health examination provided by collaborative
hospital in past few years. The areas of sampled data cover the hospital neighboring cities and
counties located in northern part of Taiwan. There are 1161 heath examination case records
are sampled. Before they are input for training and testing network models, a data
preprocessing is processed to remove duplicate, correct the error, inconsistency and missing
fields, numerification of class field in each case record.
By this process, we can ensure the accuracy, completeness and integrity of the input data.
After data preprocessing, only 945 sampled cases are left as the use of learning data set.
Meanwhile, initially there are 21 fields contained in each source of case record. Only 6 fields
specified in Table 5 are selected and maintained for each case.
Model Learning and Experiments
We take NBuilder of neural solution, a well-known neural network modeling tool, to train
and test our neural network model through a series of simulations in experiment. In
consideration of the tool restriction for building the hybrid model of neural network and
genetic algorithm which is also studied in this research, we only take the most accurate 430
cases from 945 sampled cases left after preprocessing in learning process. By this selection,
we can ensure there are equal numbers of cases used in the leaning process of each factor
Building. As it is shown in Table 6, among 430 cases, 300 cases are selected for training, 100
records are used for testing, and 30 cases for cross verification. The desired output of
diagnosis is identified by the professional physicians of chronic kidney disease from the
collaborative hospital with the aid of MDRD calculation formula.
Table 6: Data Distribution for Model Training, Testing, and Cross Verification.
Sampled Data
Record #
Desired Output #
Training Data
300
101 Negative
199 positive in various
levels
Testing Data
100
34 negative
66 positive
Cross Verification
30
10 negative
20 positive
EXPERIMENTS AND COMPARISONS OF NEURAL NETWORK MODELS
The experiment for a specific network model can be viewed as a series of modeling and
simulation through leaning process. We attempt to pursue a best model through the selection
of different combination of modeling parameters for conducting learning in different network
models. The parameters combination shown in different model experiments have been
filtered from a series of learning process.
The Experiment of Back-propagation Neural Network
Table 7 shows the parameters set for the experiment of back-propagation neural network.
Table 7: Parameters Selected for Back-propagation Neural Network.
Parameters
Value Setting
# of Hiding Layers
1、2、3
# of Neurons in Hiding Layer
5、8、12
Learning Rules
Step、Momentum、ConjugateGradient、
LevenbergMarquar、Quickprop、DeltaBarDelta
Transformation Functions
TanhAxon、SigmoidAxon、LinearTanhAxon、
LinearSigmoidAxon、SoftMaxAxon、BiasAxon、LinearAxon、
Axon
Weight Update Methods
Batch
Criterion for Termination
MSEmin:0.0001; Epochsmax:3000
In this paper, we also take the hybrid models of combining each neural network factor And
genetic algorithm (GA) in experiment and comparison. The model parameters selected for
genetic algorithm are shown in Table 8. GA will be taken to combine with back-propagation
neural network, feedforward neural network, and modular neural network, respectively in
model experiment and comparison as well.
Table 8: The Parameters Selected for Genetic Algorithm.
Network Parameters
Value Setting
Method of Selection
Roulette
Method of Crossover
One Point
Rate of Crossover
0.9
Rate of Mutation
0.01
Termination Criterion
MSEmin:0.0001
Epochsmax:100
Population: 20
Generations:50
Through a series of simulation by taking the different sets of parameter combination selected
from Table 7, we gain that the best model of backpropagation neural network in terms of their
respective network parameters with respect to two models, which are the factor By adopting
the key factors used in computation formula (called as Factor A hereafter) and the factor By
adopting the key factors selected in this paper (called as Factor B hereafter), respectively, are
listed in Table 9. The classification accuracy, sensitivity, and specificity with respect to factor
A and B, respectively is shown in Table 10.
Table 9: The Best Models with Backpropagation Network Parameters Setting.
Network Parameters
By Adopting the
Key Factors Used
in Computation
Formula (Factor
A)
By Adopting the
Key Factors
Selected in This
Paper (Factor B)
Learning Rule
LevenbergMarquar
LevenbergMarquar
Transformation Functions
TanhAxon
SigmoidAxon
# of Hiding Layer
2
2
# of Neurons in Hiding Layer
5
8
Criterion for Termination
MSEmin:0.0001
MSEmin:0.0001
Epochsmax:3000
Epochsmax:3000
Batch
Batch
Methods for Weight Update
Table 10: The Experiment Results from Backpropagation Neural Network Models.
BPN
Factor A
Factor B
Pure BPN
Training
Test
Training
Test
Accuracy of
100%
94.75%
100%
72.42%
Sensitivity
100%
97.06%
100%
70.59%
Specificity
100%
92.42%
100%
74.24%
BPN+GA
Test
Test
Accuracy of
91.71%
77.68%
Sensitivity
97.06%
81.82%
Specificity
86.36%
73.53%
Classification
Classification
In Table 10, “accuracy” is used to measure the classification accuracy with the proportion of
the sum of the number of true positives and the number of the true negatives which are
correctly identified. “Sensitivity” is used to measure the proportion of actual positives
which are correctly identified as such (e.g. the percentage of sick people who are correctly
identified as having the condition). “Specificity” is used to measure the proportion of
negatives which are correctly identified (e.g. the percentage of healthy people who are
correctly identified as not having the condition). These two measures are closely related to
the concepts of type I and Type II. From the results shown in Table 10, our experiment may
gain very good results with BPN in model training stage but it shows a significant drop
measured in model testing stage both with BPN and BPN plus GA. By the results shown in
Table 10, BPN may obtain better results in accuracy measure, but GPN plus GA gain better
result in sensitivity measure. The result also shows the model with the adoption of the key
factors used in computation formula gains better results in all three indicators measures.
These results tell us that a hybrid model with the combination of BPN and GA does not
improve the model performance in accuracy measure but it is helpful to improve sensitivity
measure.
The Experiment of Generalized Feedforward Neural Network
Table 11 shows the parameters set for the experiment of generalized feedforward neural
network (GFNN). The classification accuracy, sensitivity, and specificity with respect to
factor A and B, respectively is shown in Table 12. As the results shown in Table 12, pure
GFNN obtain a perfect classification percentage in three measures including accuracy,
sensitivity, and specificity in training stage, but GFNN plus GA apparently may gain better
results in testing. However, they also show the results gained from GFNN are not so good in
all three measures as those gained from BPN and BPN plus GA, respectively.
Table 11: The Best Models with Generalized
Feedforward Neural Network Parameters Setting.
Network
Parameters
Factor A
Factor B
Learning Rule
LevenbergMarquar
LevenbergMarquar
Transformation
SigmoidAxon
TanhAxon
2
2
8
8
Criterion for
MSEmin:0.0001
MSEmin:0.0001
Termination
Epochsmax:3000
Epochsmax:3000
Methods for
Batch
Batch
Functions
# of Hiding
Layer
# of Neurons
in Hiding
Layer
Weight Update
Table 12: The Experiment Results from Generalized Feedforward Neural Network Models.
GFNN
Factor A
Factor B
Pure GFNN
Training
Test
Training
Test
Accuracy of Classification
100%
86.63%
100%
71.03%
Sensitivity
100%
82.35%
100%
61.76%
Specificity
100%
90.91%
100%
80.30%
GFNN+GA
Test
Test
Accuracy of Classification
91.09%
78.39%
Sensitivity
88.24%
76.47%
Specificity
93.94%
80.30%
The Experiment of Modular Neural Network
Table 13 shows the parameters set for the experiment of modular neural network (MNN). The
classification accuracy, sensitivity, and specificity with respect to factor A and B, respectively
is shown in Table 14. As the results shown in Table 13, pure MNN same as prior two
models which may obtain a perfect classification percentage in three measures in training
stage, but MNN plus GA apparently may gain worse results in testing stage which are
different from the results shown in prior two models.
Table 13: The Best Models with Modular Neural Network Parameters Setting.
Network
Parameters
Factor A
Factor B
Learning Rule
LevenbergMarquar
LevenbergMarquar
Transformation
SigmoidAxon
SigmoidAxon
# of Hiding Layer
2
2
# of Neurons in
4
8
Criterion for
MSEmin:0.0001
MSEmin:0.0001
Termination
Epochsmax:3000
Epochsmax:3000
Methods for
Batch
Batch
Functions
Hiding Layer
Weight Update
Table 14: The Experiment Results from Modular Neural Network Models.
MNN
Factor A
Factor B
Pure
MNN
Training Test
Training Test
Accuracy of
100%
93.23%
100%
70.99%
Sensitivity
100%
97.06%
100%
64.71%
Specificity
100%
89.39%
100%
77.27%
MNN+GA
Test
Test
Accuracy of
88.82%
78.34%
Sensitivity
89.39%
79.41%
Specificity
88.24%
77.27%
Classification
Classification
CONCLUSIONS
The performances of three neural network models developed for comparison in this paper in
terms of detection (i.e., classification) accuracy measure are summarized and shown in Table
15 and 16, respectively. Table 15 shows the detection accuracy in three pure (i.e., fundamental)
neural network models while Table 16 shows the detection accuracy with GA embedded in
three respective models. We found BPNN may obtain best accuracy regardless of in factor A
or in factor B which are measured with 94.75% and 72.42% accuracy, respectively in model
testing. However, factor A, which is the test by adopting the key factors used in computation
formula, may gain much better performance than factor B, which is the test by adopting the
key factors selected in this paper. The same result is gained with the GA embedded in each
fundamental model in experiment. As the results observed from Table 16, both BPN plus GA
and GFNN plus GA gain close results in accuracy measure which is much better than MNN
plus GA.
Table 15: Detection Accuracy in Three Pure Neural Network Models.
Network
Models
By Adopting the
Key Factors Used
in Computation
Formula (Factor A)
By Adopting the
Key Factors
Selected in This
Paper (Factor B)
BPN
94.75%
72.42%
GFNN
86.63%
71.03%
MNN
93.23%
70.99%
Table 16: Detection Accuracy in Three Pure Neural Network Models Plus GA.
Network
Models
Factor A
Factor B
BPN+GA
91.71%
77.68%
GFNN+GA
91.09%
78.39%
MNN+GA
88.82%
78.34%
By further observation from the results of model experiment shown in section 4, it is found
that almost all models employed in experiment may gain near 100% accuracy in CKD
detection in the training stage regardless of which sets of influence factors used in model
training. However, if further model testing is conducted, it is found that the network models
with the inputs of influence factors of CKD used by physicians employed in the
computational formula always shows better detection performance in all three aspects of
measure including accuracy, sensitivity, or specificity measures. The BPN gain the highest
94.75% (shown in Table 10 and 15) accuracy measure in the testing stage among three
fundamental neural network models while GFNN gains only 86.63% (shown in Table 12 and
16) in accuracy measure which is the lowest performance in three models. Through further
observations from experiment results, it is found the hybrid network model of GFNN plus GA
embedded may significantly enhance detection performance in all three measures from its
fundamental model in the testing although the reversed enhancement are shown in BPN and
MNN. Again, by the experiment figures shown in section 4, although pure neural network
models plus GA shows degraded performance improvement for factor A in testing stage in
terms of three measures in accuracy, sensitivity, and specificity but it shows significant
improvement for factor B. As a result, we conclude a hybrid model indeed may improve
detection performance. This result may conclude that GA provides no benefit in the yield of
better detection performance with the adoption of the key factors used in computation formula
as the inputs to all models in testing.
Above all, we conclude that BPN might be the best-fitting factor Among three fundamental
neural network models employed in detecting CKD while the models with GA embedded in
BPN and GFNN respectively might be best-fitting hybrid models for CKD detection but only
GFNN plus GA can gain enhancement in detection performance in both sets of influence
factors as inputs. These conclusions should be further verified and compared with other
models selected for experiment in later studies before they can be assured.
We believe that the neural network models respectively developed for CKD detection and risk
level estimation in this paper may equip medical staff with the ability to make precise
diagnosis and treatment to the patient. Meanwhile, public users may take advantage of these
system instruments to conduct self-detection of the risk level of causing CKD in order to take
necessary precaution in advance to avoid any risk of causing CKD or prevent their condition
from getting worse.
REFERENCES
1. Azam F. (2000). Biologically Inspired Modular Neural Networks. Ph.D. dissertation,
Virginia Polytechnic Institute and State University.
2. Arulampalam G., & Bouzerdoum, A. (2003). A Generalized Feedforward Neural Network
Classifier. Proceedings of the International Joint Conference on Neural Networks, 2,
1429-1434.
3. Arulampalam G., & Bouzerdoum A. (2000). Training shunting inhibitory artificial neural
networks as classifiers. Neural Network World, 3(10), 333-350.
4. Arulampalam G., & Bouzerdoum A. (2001). Application of shunting inhibitory artificial
neural networks to medical diagnosis. Proceedings of the Seventh Australian and New
Zealand Intelligent Information Systems Conference (ANZIIS 2001), Perth, Australia,
89-94.
5. Chiu, R.K., Huang C.I., & Chang Y.C (2010). The Study on the Construction of
Intelligence–Based Decision Support System for Diabetes Diagnosis and Risk Evaluation.
Journal of Medical Informatics, Taiwan Association for Medical Informatics, 19(3), 1-24.
6. Department of Health, Taiwan (DOH) (2010). The Statistical Data of the Causes of
Civilian Mortality, Retrieved from http://www.doh.gov.tw/CHT2006/DM/DM2_2.aspx?
now_fod_list_no=11122&class_no=440&level_no=3.
7. Eknoyan, G. (2007). Chronic kidney disease definition and classification: The quest for
refinements. Kidney International, 72, 1183-1185.
8. Kochurani, O.G., Aji, S., & Kaimal, M.R.(2007). A Neuro Fuzzy Decision Tree Model for
Predicting the Risk in Coronary Artery Disease. Proceedings of 22nd IEEE International
Symposium on Intelligent Control, Singapore, 166-171.
9. Firman G. (2009), Definition and Stages of Chronic Kidney Disease (CKD), Released on
Feb. 28, 2009, Retrieved from http://www.medicalcriteria.com/site/index.php?option=com
_content&view=article&id=142%3Anefckd&catid=63%3Anephrology&Itemid=80&lang
=en.
10. Kriesel, D. (2011). A Brief Introduction to Neural Networks, Retrieved from
http://www.dkriesel.com.
11. National Kidney Foundation (NKF) (2002). K/DOQI Clinical Practice Guidelines for
Chronic Kidney Disease: Evaluation, Classification and Stratification. American Journal
of Kidney Disease, 39, S1-S266.
12. Levey, A.S., Greene, T., Kusek, J.W, Beck, G.L., et al. (2000). MDRD Study Group. A
simplified equation to predict glomerular filtration rate from serum creatinine. J Am Soc
Nephrol, 11:155A.
13. Levey, A.S. Eckardt, K.U., & Tsukamoto, Y. (2005). Definition and classification of
chronic kidney disease: A position statement from Kidney Disease: Improving Global
Outcomes (KDIGO). Kidney International, 67, 2089-2100.
14. Snyder, S., & Pendergraph, B. (2005). Detection and Evaluation of Chronic Kidney
Disease. American Family Physician, 72(9), 1723-1732.
15. Jacobs, R., Jordan, MNowlan, S., & Hinton, G. (1991). Adaptive Mixtures of Local
Experts. Neural Computation, 3(1), 79-87.
Download