International Journal on Advanced Research in Science

advertisement
INTERNATIONAL JOURNAL ON ADVANCED RESEARCH IN
SCIENCE, ENGINEERING AND TECHNOLOGY
IJSET is a scholarly peer-reviewed journal that contributes for the contemporary trends in the
field of Science, Engineering and Technology published quarterly by Vankatram Learning Centre
(Central Library), Sri Krishna College of Engineering and Technology. Responsibility for the
contents rests upon the authors and not upon the IJSET. For copying or reprint permission, write
copyright department, IJSET, Vankatram Learning Centre, Sri Krishna College of Engineering and
Technology, Kuniamuthur Post, Coimbatore – 641 008, Tamil Nadu, India.
Patron
Smt.S.Malarvizhi
Advisor
Dr.S.Sundararaman
Editor in Chief
Dr.S.Annadurai
Editor
Dr. A.F.Rahman
Managing Trustee
Chief Executive Officer
Principal
Librarian
Editorial Board
1. Dr.K.Duraisamy
Professor, KSR College of Tech., Tiruchengode
2. Dr.L.Ganesan
Professor, Alagappa Chettiar College of
Engg. & Tech., Kariakudi, Tamilnadu
3. Dr.S.Thamarai Selvi,
MIT, Dean CT, Anna University, Chennai
4. Dr.R. Shanmugalakshmi
Associate Professor, GCT, Coimbatore
5. Dr.T.Purusothaman
Associate Professor, GCT, Coimbatore
6. Dr.V. Gopalakrishnan
Associate Professor, GCT, Coimbatore
7. Dr.Narasiman
Professor, Alagappa Chettiar College of
Engg. & Tech., Kariakudi, Tamilnadu
8. Dr.Esther Annlin Kala James
Principal, GCE, Srirangam, Trichy
9. Dr.P.Geetha
1. Prof. Palesi Maurizio
Associate Professor
University of Enna “Kore”, Italy
2. Prof. Leo Mathew,
Professor, Texas University, Austin
3. Dr. Selwyn Piramuthu
Professor, University of Florida, USA
4. Dr.Henry Selvaraj
Professor, University of Nevada, USA
5. Dr. Sameer Hanna Khader
Associate Professor,
Palestine Polytechnic University, Palestine
6. Prof. Vijay K Arora
Professor, Wilkes University, USA
7. Dr.Kevin Demmitt,
Provost and Vice President,
Clayton State University, Georgia
8. Prof. Scott E.Mclntyre,
Professor, University of TOLEDO, USA
Associate Professor, Anna University, Chennai
9. Dr. Hairul Azhar Abdul Rashid
Professor, Sri Ramakrishna Engg. College, CBE
Director Research Management Centre,
Multimedia University, Malaysia
10. Dr.Grace Selvarani
11. Dr.Anna Saro
Director and Head, Dept. of MCA,
SNR Sons College, CBE
10.Dr. Teerapat Sanguankotchakorn
Associate Professor
Asian Institute of Technology, Thailand
12. Dr.Arulanandan
INTERNATIONAL JOURNAL ON ADVANCED RESEARCH IN
Professor, PSG Tech, Coimbatore
SCIENCE, ENGINEERING AND TECHNOLOGY
INTERNATIONAL JOURNAL ON ADVANCED RESEARCH IN
SCIENCE, ENGINEERING AND TECHNOLOGY
INTERNATIONAL JOURNAL ON ADVANCED RESEARCH IN
SCIENCE, ENGINEERING AND TECHNOLOGY
International Journal on Advanced Research in Science, Engineering and Technology is
published as a quarterly publication.
The journal is a scholarly peer-reviewed journal that contributes for the contemporary trends
in the field of Science and Technology.
IJSET or editorial board accepts no responsibility for the statements and opinions advanced by
contributors. The Editorial Committee’s decision regarding suitability of contributions for
publication will be final.
All correspondence and manuscripts of contributions should be sent to
Dr.S.Annadurai
Principal & Editor-in-Chief (IJSET)
International Journal on Advanced Research
in Science, Engineering and Technology
Sri Krishna College of Engineering and Technology
Kuniamuthur Post
Coimbatore : 641 008
Tamil Nadu, India.
Marking a copy to
Dr.A.Fazlur Rahman
Librarian & Editor (IJSET)
International Journal on Advanced Research
in Science, Engineering and Technology
Vankatram Learning Centre
Sri Krishna College of Engineering and Technology
Kuniamuthur Post
Coimbatore : 641 008
Tamil Nadu, India.
E-Mail : librarian@skcet.ac.in
drfrlib@gmail.com
© 2016, IJSET, Vankatram Learning Centre, Sri Krishna College of Engineering and
Technology, Coimbatore.
INTERNATIONAL JOURNAL ON ADVANCED RESEARCH IN
SCIENCE, ENGINEERING AND TECHNOLOGY
MAY – AUGUST 2016
VOLUME 1
NUMBER 1
IJARSET
(ISSN
)
CONTENTS
Sl.No.
Title
Page
No.
1
A Novel Approach for Identifying Common Signs of CT Pathological Lung
Images………………………………...………… T. Revathi, and P. Geetha
01
2
Ensemble Approach on Maize Leaf Nutrition Deficiency Identification Using
Enhanced Multivariate Partial Least Square…………………………………..
……………………………………….. S. Sridevy, and Anna Saro Vijendran
09
3
Low Power Multiplier Using Algorithmic Noise Tolerant Architecture
.…………………………………………………...... S. Fathima, and S. Uma
16
4
A Survey on Compression in Encrypted Domain …………………………...
……………………………………….. T. Suguna, and R. ShanmugaLakshmi
20
5
A Survey on Assistive Technology using Natural User Interface(NUI)
computing to support and augment therapeutic rehabilitation ………………..
.. ...…… S. J. Syed Ali Fathima, S. Shanker, and A. Ahamed Thajudeen
27
6
Experimental Investigation on Weld Strength of Copper Wire – Al8011 Sheet
Using Ultrasonic Metal Welding .....................................................…………
…………………. R. Karthik, P. Ashoka Varthanan, and V. S. Gowthaman
32
7
Effect of Vibratory treatment on Metallurgical and Mechanical properties of
Aluminum alloys weldments……… K. Balasubramanian, and V. Balusamy
40
A Novel Approach for Identifying Common Signs of
CT Pathological Lung Images
T. Revathi
Dr.P.Geetha
PG Student
Dept. of Computer Science & Engg
College of Engineering, Guindy
Anna University, Chennai.
revathiarasu.t@gmail.com
Associate Professor
Dept. of Computer Science & Engg
College of Engineering, Guindy
Anna University, Chennai.
geetha@cs.annauniv.edu
Abstract-Accurate identification of pathological lungs
& its sign is a challenging task in medical imaging
applications. The existing methods fail to provide a better
solution for images with dense pathologies and they
falsely exclude some regions from the lung parenchyma.
To tackle this problem, a novel pathological identification
system is required. It should work in a way that it
encounters almost all common pathologies & its sign
present in the image set. This proposed approach is fully
based on Volume based pathologies identification. It has
two stages. Stage one is for Volume detection & Volume
difference
test,
where
Fuzzy-Connectedness(FC)
segmentation and morphological - Rib cage Volume
detection method is used. FC is used for segmenting lungs
with diverse range of lung abnormalities. Refinement
steps are conducted in stage two by using Random Forest
(RF) algorithm which further refines the images such as
pathological existence and its sign by using textural
information of images. It classifies the lung images into
two main classes namely pathological class and nonpathological class. Pathological class has nine sub
categories such as GGO(Ground Grass Opacity),
Lobulation, Air Bronchogram, Cavity, Plueral
Indentation(PI), Spiculation, Calcification, Obstructive
Pneumonia and Bronchial Mucus Plugs. So these
subcategories of signs are also identified for effective
diagnosis of lung diseases.
Some wide variety of imaging techniques are available today
such as magnetic resonance imaging (MRI), computed
tomography (CT), etc. But interpreting a CT image is very
challenging because the image is complicated by
superimposed anatomical structures. Also some current lung
pathological identification algorithms work well for certain
lung pathologies present in moderate amounts, they fail to
perform when dense pathologies exist. Accurate
segmentation of the pathological lung is challenging since
lung pathologies hold different appearances from the normal
lung tissues. And CT scans of patients with severe lung
diseases often include diverse imaging patterns. Thus, it is
difficult to adapt the current methods due to their limited
ability to be applied on finding diverse imaging patterns with
dense pathologies. However misdiagnosis the image will
affect patient care and treatment. So, further studies are
necessary in the development of medical imaging software to
produce more details of the image and diseases. Therefore the
aim of this work is to target this challenge and provide a
generic solution for lung pathology & its sign identification
from CT scans.
There were so many diseases identification
techniques had been developed, but each approach
concentrates on particular diseases detection like respiratory
problem, lung diseases, pulmonary diseases, nodule diseases
& pleural effusion diseases etc. Mostly earlier studies were
concentrated on lung cancer detection. There is no such
generic approach which finds all the lung abnormalities at a
single stretch. Therefore accurate segmentation of the
pathological lung is challenging. So, there is a need for such
approach to recognize all the lung abnormalities in a single
system.
Index Terms—Lung Segmentation, Fuzzy Connected
-ness, Random Forest.
I.
INTRODUCTION
Medical imaging plays an important role in medical
world nowadays. Pulmonary Lung diseases and disorders are
one of the major causes of deaths and hospitalization around
the world. The American Lung Association estimates that
about 90% of deaths occur per year in the United States from
lung diseases[2].
Computed tomography is the current modality used in
clinics for detecting and diagnosing the lung diseases[2].
Since manual inspection is tiresome and time consuming.
II.
LITERATURE SURVEY
The study of various lung segmentation techniques and
classification techniques are described in below sections.
A. Lung Segmentation Techniques
The aim of medical image segmentation is to extract
quantitative
information
(e.g.,volumetric
data,
morphological data, textural information) with regard to an
1
organ of interest or a lesion within the organ. Some major
lung segmentation methods are Threshold-based, Regionbased Segmentation methods.
Threshold based segmentation method segments the
images or distinguishes the image regions with respect to
contrast / intensities present in the images. Threshold - based
methods [2], [3] are often used for their efficiency. However,
such methods have limited applicability as they fail to
consider the intensity variations due to pathologies presence
or even under normal conditions. In [4] used multi-level
thresholding for identifying the regions of interest (ROI) of
lung nodule. 5% –17% of the lung nodules are undetected.
In[5], Adaptive threshold method is used for segmenting the
lung region. It fails to deal with attenuation variations.
Images with non uniformity and noise presence leads to
misclassification.
Region Based Segmentation method overcomes the
drawback of threshold based segmentation. It segments
images based on the regions with similar characteristics. It
has 2 Methods: Region growing method, Region Splitting &
Merging. It is used in[6], on which neighboring pixels are
scanned and added to a region session. In [7][8],
Combination of spatial Fuzzy C-Means and morphological
techniques for segmenting the lung region from CT lung
images. They have divided their work into three main stages
namely
Background
removal,
Preprocessing,
and
Morphological based operations. Among them, a new
histogram background removal operator had been used to
remove the pixels around the lungs. Moreover, preprocessing
is performed by smoothing and removing the noise presents
in the segmented image. Finally, morphological operations
are used to separate the edges and to fill the small holes
present in the segmented lungs. In[9], ROI based region
segmentation, the initial step in computer aided diagnosis of
lung CT image is generally to segment the Region of Interest
(ROI) present in the image and then analyzes each area
separately in-order to find the presence of pathologies present
in ROI. Region growing has been combined with CC in this
work since it reduces the number of steps in segmentation for
the process of identifying a tissue in the CT lung image.
In[8], The lung lobes and nodules in CT image are segmented
using two stage approaches such as modified adaptive
fissure sweep and adaptive thresholding. Initially preprocessing is used to remove the noise present in CT image
using filter technique, then the fissure regions are located
using adaptive fissure sweep technique, then histogram
equalization and region growing is applied to refine the
oblique fissure.
Therefore the above techniques were only concentrated
on identifying particular diseases or disorders. It doesn’t
identify all pathological signs or diseases in a single system.
Even it failed to deal with attenuation variations & noise
presence in the image. Region growing segmentations covers
all the neighbor pixels which has same characteristics, this
mean that it covers almost all pixels with same intensity.
Whereas the new Fuzzy Connectedness Approach finds the
neighbor pixel which has similar spatial & textural property
by means of calculating affinity value. FC approach works
well even the images consists of noise.
B. Classification Based On Machine Learning
It focuses on extracting suitable features such as shape or
texture for a predefined classifier such as support vector
machines, random forests, neural networks etc. Support
Vector Machine(SVM)
performs classification by
constructing N dimensional hyper plane that optimally
separates the data into two categories. It is a linear classifier
and this is not suitable when giving large number of datasets
or continuous data. Decision trees are predictive models that
use a set of binary rules to calculate a target value. In[12],
addresses the challenging problem of segmenting the lungs in
CT scans. They proposed the context selective decision forest
(CSDF) as a new discriminative classifier which resulting in
higher prediction accuracy and greater use for the clinic.
Quantitative comparisons were made with state of art
algorithms and demonstrated comparable accuracy with
superior computational efficiency. Benefits of using decision
forest is it can able to handle continuous data and large
number of features. It provides more correctness and
accuracy comparatively. It is easy to implement and less time
consuming.
This Survey clearly shows the segmentation techniques
with its benefit and drawback. From this, we infer that, first
order segmentation such as Threshold, Region, Edge are not
suitable due to its flaw. So Fuzzy Connectedness(FC) will be
used to segment the lung and also to get volume information.
Whereas in Classification, Decision Forest is the most
suitable one to handle large number of features as well the
continuous data. So Random Forest would be the best for
Classification work since it combines many decision trees
and provides higher accuracy and performance.
III.
ARCHITECTURE
Fig.1 illustrates the entire system of pathological lung
identification approach. Input of the system is Lobe and Lung
Analysis (LOLA) images. Fuzzy Connectedness based
segmentation is performed on the input image to get
segmented image with its volume. Then for Pathological
testing, volume difference test will be performed. For that it
needs two Volumes. Volume one was got from FC
segmentation (ie., Lung Volume), In order to find second
volume of the image, Morphological Rib Cage analysis
technique is used. Then Rib Cage is determined using
Morphological
operations
by
removing
scapula.
2
Fig.1. Proposed Architecture
Because Scapula presence will results in getting increased
volume of the image. Then Convex hull technique is used to
fit polygon over the Rib cage identified. This process has to
be carried out for each and every image to find Volume of
Convex hull fitted Rib image. Finally Volume difference test
will be performed over 2 volumes, one got from FC
Segmentation and other from Rib Cage Convex Hull. From
this we infer that the image has presence of pathology or not.
The image is said to be Pathological image if Volume
difference occurs or else the images is said to be normal/
non-pathological image. Subcategories of pathological class
is given in fig.9. In-order to confirm and refine the results
obtained from Pathology Test, it moves to multi stage
refinement level where Random Forest Classifier is used. It
takes Training dataset which is collected from wide range of
pathological images & signs of pathological images such as
GGO(Ground Grass Opacity), Calcification, Cavity,
Spiculation, Lobulation, Air Bronchogram, Bronchial Mucus
Plugs, Plueral Identation(PI) and Obstructive Pneumonia[11].
The test dataset contains the feature extracted from LOLA
datasets. It models the classifier based on the Training set and
performs classification by means of passing the test sets over
it. Final output of the system gives 2 main classes of images,
class 0 which has non-pathological images and class 1 which
has pathological images with its sign.
IV.
MODULES DESCRIPTION
Description of modules such as Fuzzy Connectedness,
Convex Hull, Random Forest are described below.
3
b) Strength is determined by all possible
connecting paths between those pixels.
c) Each pair of nearby pixel has affinity value.
d) It traverses through the pixel with minimum
affinity value.
e) When some abnormality is seen, FC stops
region growing and outputs the segmented area. (Initial Seed/
pixel Selection can be done either manually or
automatically). Output of segmented lung is given in fig.6.
Step 1) FC Algorithm
Input: Image and Seed point.
Output: Segmented Image.
1. Read Image.
2. Convert to gray scale.
3. Select Seed Point /Initial pixels (say c,d).
4. Find Adjacency coordinates of pixels.
A. Segmentation Based On Fuzzy Connectedness
It is the representation of connectedness between the
pixels comprising an object in an image. It is developed
mainly for medical image segmentation. The seed / initial
pixel is determined by means of computation. Sometimes
multiple seeds also to be considered.
Using an initial seed pixel within an object in an image,
it computes affinity value for the adjacent pixels & traverse
through it. Here intensive computation may be required
because connectedness is defined based on an optimal path to
the seed pixel.
Dynamic programming is used to determine the optimal
paths. Sometimes operator-selected region of interest has also
been applied to reduce computation time. Fuzzy
connectedness is closely related to other graph-search
algorithms. Final object selection is performed using a
threshold on the resulting fuzzy connectedness map.
It has three components namely Homogeneity, Fuzzy &
Object Spel Adjacency.
Homogeneity based component indicates the
degree of local hanging togetherness of spels
due to their similarity of intensities.
Fuzzy Spel Adjacency component indicates the
degree of spatial adjacency of spels.
Object Spel Adjacency component indicates the
degree of local hanging togetherness of spels
with some given object feature.
(1)
5. Computes Affinity
Let fuzzy affinity Psi (c, d) quantify the
hanging-togetherness of two spels c and d.
(2)
6. Do Absolute Fuzzy Connectedness to get FC map
For any path π, the strength of the path is
minimum affinity along the path.
(3)
Find Maximum Path:
(4)
5. Return FC connectivity map
6. Adjust threshold if necessary
7. Get best Segmented Lung.
Step 2) Maximum Path Coverage Algorithm
1. Initialize array with fc=0:1.
2. Push all pixels μ (c, d) > 0 to Q.
3. While Q is not empty do
a. Find strongest fc from Q.
b. Find its adjacency & remove it.
c. Find fmax= path with minimum affinity.
If fmax > fc then
i. Set max f c =fmax.
ii. Push all pixels e such that
min [( fmax, μ(c,d) )] > f(e) to Q
End if.
4. End while
5. End.
Fig.2. Fuzzy Connectedness.
Working of Fuzzy Connectedness : FC working is
presented in fig.2, FC works based on strength of
connectedness between 2 pixels. That is hanging
togetherness – based on the similarity in spatial location
as well as in intensity.
a) Fuzzy Connectedness is the strength of
connectedness of all pixels.
4
obtained by means of fixing convex hull on lung which
presence inside the rib.
B. Proposed Approach on Rib Cage Convex Hull Fitting
This is used to find the Volume of the rib cage. Fig.3,
describes the working of Rib Cage Convex Hull Fitting.
Background subtraction & Morphological operations are
used to identify ribcage and remove scapula from it. The
identified rib will be fitted by convex hull to get its volume.
This volume will be used for performing Volume difference
Test. Input of this module is LOLA images. Then extract the
rib cage part alone from the image. In this Stage, Convex
Hull technique is used to fix polygon on the extracted Rib
cage of the image. So that Volume of the Convex Hull will
be generated.
Volume Difference Test
Volume difference test is performed on two volumes got
for the same image (V1 obtained from FC and V2 obtained
from Ribcage Convex hull for the same image). The image is
said to be pathological if the difference of the two volumes
exceeds the threshold. Fig.8 shows the volume difference test
performed on the images.
The same process repeats for all the images present in
the Lola set and declares that the image has pathology or not.
Then it enters into further refinement stage for pathology
conformance.
C. Random Forest
The Classifier is built according to the training datasets
and tested effectively. See Fig.4, Random Forest is a
collection of Decision Trees in which wide range of textural
features were given to train the classifier. Test datasets will
be passed on it to perform classification. The classified result
as two main classes such as Pathological and NonPathological class. And Pathological class is further divided
into 9 categories of common Imaging signs of lungs. GLCM
(Gray Level Co-occurrence Matrix), GLRLM (Gray Level
Run Length Matrix) & Histogram feature extraction
techniques are used. Lola dataset is used for testing and for
training various pathological images. Weka tool is used for
classification. Classification result of RF is given in table 1.
Fig.3. Ribcage Convex Hull Fitting.
Working of Rib Cage Convex Hull Fitting:
Usually background of the image has higher
intensity than the object present in the image.
Histogram is used to threshold the image.
Smoothing techniques are used for image
Enhancement.
Blobs present in the image is identified by
performing Blob detection technique. Here each and
every blobs present in the image are identified with
its details such as area, perimeter, centroid etc.
Using the Blobs information, Scapula blobs can be
removed.
By using morphological operations such as
connected components, scapula blob is removed and
rib cage alone can be obtained.
Convex Hull is a technique which is used to fix hull
or mask on the image which obtained from the
previous step.
o Ie., Convex Hull is fixed on the lung which
presents inside the Rib blobs.
Fig.7 shows Convex Hull fitted on lung image presence
inside the identified rib blobs. Second Volume information is
Fig.4. Random Forest.
Feature Extraction
This approach is used to extract/select the important
features from the images. Main aim is to avoid irrelevant
feature from the image because this may leads to
misclassification when using machine learning techniques.
For Feature Selection following techniques are used.
Gray Level Co-Occurrence Matrix (GLCM) - The
co-occurrence matrix is a statistical model that is useful in a
variety of image analysis applications, such as in biomedical,
remote sensing, industrial defect detection systems, etc. The
5
gray-level co-occurrence matrix (GLCM) is a frequency
matrix, indicates the frequency of a pair of pixels that are at
“exactly the same distance and direction of the displacement
vector”.
Gray Level Run Length Matrices (GLRLM) - A
gray level run is a set of consecutive, collinear picture points
having the same gray level value. The length of the run is the
number of picture points (pixels) in the run. The major reason
for use of these features is that the lengths of the runs reflect
the size of the texture elements. It also contains the
information about the alignment of the texture. For a given
picture, it computes a set of the gray level run length matrices
(GLRM’s) for runs with any given direction.
Histogram (HIST) - Histogram is performed to show
the number of pixels in an image at each different intensity
value found in an image. It also can be used to enhance
image.
WEKA - Waikato Environment for Knowledge
Analysis is a popular suite of machine learning software
written in Java, developed at the University of Waikato, New
Zealand. It is a collection of machine learning algorithms for
data mining tasks. The algorithms can either be applied
directly to a dataset or called from Matlab code.
Weka contains tools for data pre-processing, classification,
regression, clustering, association rules and visualization.
Fig.6. FC Segmented Lung.
Convex hull Fitted on Rib Lung Image
Fig.7.RibCage Fitted Lung Image
Weka (For Classification)
Fig.5. Common Signs of Lung diseases.
V.
EXPERIMENTAL SETUP & RESULTS
The input LOLA dataset[1] is collected from Lola
website. FC is used for segmenting the lung. Classification is
done by using WEKA tool. Train datasets are collected from
various lung databases[11] which includes signs of lung
diseases, pathological lung images & normal lung images.
LOLA dataset is used as test dataset in Classification.
FC Segmentation
Fig.8. Random Forest Classification.
Performance Measures
Comparision with Other Segmentation Technique:
FC and RF methods are considered to be best for
identifying pathological lung with signs. Table 1 holds
Comparision details of Segmentation Algorithm.
6
S.No
Segmentation
1
Threshold-based
Segmentation
Region-based
Segmentation
Fuzzy-Connectedness
Segmentation
2
3
images. Further, the robustness and the effectiveness of the
proposed method was tested on more lung scans acquired
through various sources containing a wide range of
abnormalities. And the classifier performance is compared
with other classifier. So, this confirm that the best accuracy
and effectiveness of the presented method.
Future work is to mainly concentrate on identifying
the particular type of lung diseases with its severity.
Accuracy
Percentage(%)
70%
82%
98%
Table 1 Segmentation Performance.
REFERENCES
Comparison With Other Classifiers:
Weka tool is used for Classification. Below Table.2
holds the information about various Classifiers, its time taken
and accuracy percentage obtained.
S.No
Classification
Time
Accuracy
(Sec)
Percentage(%)
1
REP
0.5
99.7809%
2
Decision
0.9
99.7809%
3
Adaboost
0.19
99.0142%
4
Bagging
0.16
99.7809%
5
RF (Proposed)
0.17
99.8905%
[1]
[2]
Table 2 Classifiers Performance.
[3]
Here the same training dataset and test dataset were applied
to various classifier to check their performance. From this
analysis, The RF(Random Forest) Classification is the one
which classifies the lung with greater accuracy with
reasonable time.(Given in fig.9).
[4]
[5]
[6]
[7]
[8]
Fig.9. Classifiers Performance Chart.
The above analysis shows that FC and RF are best to identify
signs of pathological images.
[9]
CONCLUSION AND FUTURE WORK
The proposed method is the first generic approach
covering a wide range of pathologies. The core of this work
is FC segmentation and rib cage volume estimation. This
method is equipped with multiple refinement stages including
Random Forest machine-learning classification to handle
pathologies. Also the classifier is trained with more than 900
[10]
7
Awais Mansoor, Member IEEE, Ulas Bagci, Member
IEEE, Ziyue Xu, Brent Foster, Kenneth N. Olivier,Jason
M. Elinoff, Anthony F. Suffredini, Jayaram K. Udupa,
Fellow IEEE and Daniel J. Mollura."A Generic
Approach to Pathological Lung Segmentation". IEEE
Transactions on Medical Imaging, Vol. 33, No. 12,
December 2014.
American Lung Association, Epidemiology and Statistics
Unit Research
and Program Services, “Estimated
prevalence and incidence of lung disease by Lung
Association Territory,” Lung Assoc, Washington, Tech,
2010.
G. J. Kemerink, R. J. Lamers, B. J. Pellis, H. H. Kruize,
and J. Van Engelshoven, “On segmentation of lung
parenchyma in quantitative computed tomography of
the lung,” Med. Phys., vol. 25, p. 2432, 1998.
S.G.Armato, III and W. F. Sensakovic, “Automated lung
segmentation for thoracic CT: Impact on computeraided diagnosis,” Acad. Radiol., vol. 11, no. 9, pp. 1011–
1021, 2004.
T.S. Chen, P.C. Chung, P.S. Liao, “A Fast Algorithm for
Multilevel Thresholding,” Journal of Information
Science and Engg, pp.713-727,2001.
M.N.Prasad, M.S. Brown, S. Ahmad, F. Abtin, J. Allen, I.
daCosta, H.J. Kim, M. F. McNitt-Gray and J. G. Goldin,
“Automatic segmentation of lung parenchyma in the
presence of diseases based on curvature of ribs,” Acad.
Radiol., vol. 15, no. 9, pp. 1173–1180, 2008.
Karthick S, Puraneeswari A and Sathiyasekar K, “A
Survey Based on Region Based Segmentation,”
International Journal of Engineering Trends and
Technology (IJETT) – Volume 7,2010.
M. Arfan Jaffar, Ayyaz Hussain, Anwar Majid Mirza,
“Fuzzy entropy based optimization of clusters
for the segmentation of lungs in CT scanned
images,” Knowledge Info Systems,Vol.24, pp. 91-111,
July 2010.
Timothy N.Jones, Dimitris N.Metaxas, “Image
Segmentation Based on the Integration of Pixel
Affinity and Deformable,” published in the Proceedings
of CVPR,1997.
A. Prabin, J. Veerappan, “Automatic Segmentation of
Lung CT images by CC based Region Growing,”
Journal of Theoretical and Applied Information
Technology. Vol. 68 No.1,October 2014.
[11]
[12]
Xiabi Liu, Ling Ma, Li Song, Yanfeng Zhao, Xinming
Zhao and Chunwu Zhou, “Recognizing Common CT
Imaging Signs of Lung Diseases Through a New
Feature Selection Method Based on Fisher Criterion
and Genetic Optimization,”IEEE Journal of BioMedical
and Health Informatics, Vol.19, no.2, March 2015.
Awais Mansoor, Brent Foster, Daniel J. Mollura, Kenneth
N. Olivier, Jason M. Elinoff, Jayaram K. Udupa,
Suffredini, Ulas Bagci and Ziyue Xu, “A Generic
Approach to Pathological Lung Segmentation,” IEEE
Transactions on Medical Imaging, Vol. 33, No. 12,2012.
8
Ensemble Approach on Maize Leaf Nutrition
Deficiency Identification Using Enhanced
Multivariate Partial Least Square
S. Sridevy
Dr. Anna Saro Vijendran
Assistant Professor
Department of PS and IT, AEC and RI
Tamilnadu Agricultural University
Tamilnadu, India
Director & Head
Department of MCA
SNR Sons College
Coimbatore, Tamil Nadu, India
Abstract—With the evolution of technology, people have
adopted their day today lives to utilize the benefits of
highly advanced technologies. Plants are among the
earth's most useful and beautiful products of nature.
The crucial need is that many plants are at the risk of
extinction. The proposed system implements a novel
approach to identify the lack of nutrients in the crop
image based on feature extraction. Images are collected
from the Agricultural University, Coimbatore. The
input images are processed in three steps:
preprocessing, feature extraction and regression. In the
preprocessing, HSV transformation and histogram
enhancement are processed. After the preprocessing,
features of images are extracted based on their unique
characteristics of images they are colour, shape and
texture. Based on the extracted features to identify the
nutrition deficiency in maize leaf images multivariate
partial least square is used as image analyzing
techniques. The experimental result shows the more
promising result on the proposed work while comparing
the existing approaches.
Feature extraction is a general term for methods of
constructing combinations of the variables to get
around these problems while still describing the data
with sufficient accuracy. Automatic detection of plant
diseases is an important research topic as it may
prove benefits in monitoring large fields of crops, and
thus automatically detect the diseases from the
symptoms that appear on the plant leaves. This
enables machine vision that is to provide image based
automatic inspection, process control and robot
guidance. Comparatively, visual identification is
labor intensive, less accurate. When crops suffered
disease, if the symptoms are not obvious or more
complex, it will be very difficult to distinguish the
characteristics so that delay the effective control of
crop diseases seriously. We can analyze the image of
disease leaves by using computer image processing
technology and extract the features of disease spot
according to color, texture and other characteristics
from a quantitative point of view. The caused and
extent of disease can be diagnosed timely and
effective, it could be prevented and control
comprehensive combined with the specific
circumstances of crop. It has a great significance in
automatic and intelligent management aspects of crop
growth and health with the research of crop diseases
using image feature extraction technology.
The quality inspection of leaves consists of two
main aspects, internal and external examinations.
Human sensory usually achieves the internal quality
inspection, smoking test or chemical analysis, while
the external quality inspection is mainly achieved
through human vision. It is costly and yet time
consuming to inspect internal quality since leaves
contain too many ingredients to be handled. As an
alternative, external quality examination is often used
Index Terms—image processing, agriculture, maize
crop, multivariate partial least square and feature
extraction.
I.
INTRODUCTION
Feature extraction involves simplifying the amount of
resources required to describe a large set of data
accurately. When performing analysis of complex
data one of the major problems stems from the
number of variables involved. Analysis with a large
number of variables generally requires a large
amount of memory and computation power or a
classification algorithm which over fits the training
sample and generalizes poorly to new samples.
9
instead in the examination of internal quality of
leaves, since external features are closely related to
internal quality. The external quality inspection of
leaves includes judgment of color, maturity, surface
texture, size and shape. Human vision, which is
inevitably limited by personal, physical and
environmental factors, has been the predominant
means of inspection.
Research has received considerable attention in
recent years [1, 2] on the utilization of moments for
object characterization in in-variant tasks and noninvariant tasks. The mathematical concept of
moments has been around for many years. It has been
used in various fields varying from mechanics and
statistics to pattern recognition and image
understanding. This paper focuses on automation
through computer vision. Computer vision [3] is
concerned with the theory behind artificial systems
that extract information from images. This work was
done to extracts different features from maize leaf
images based on color, texture and shape.
Furthermore, the system used the results of the
classification scheme in identifying the class of
nutrition deficiency of maize leaf images
Phosphorous deficiency using artificial neural
network segmentation and Otsu method.
Marian wiwart et al [7] adapted Euclidean
distances between the color of leaves at successive
nodes the change of color in leaves (Faba bean, pea,
yellow lupine plant) were analyzed. These color
variations in leaves was due to lack of Nitrogen,
Potassium, phosphorus and Inverted-V deficiency
respectively.
Jyotismita Chaki et al. [8] worked on a complete
plant leaf recognition system which had two stages.
The first stage was the extraction of leaf features.
Moment invariants technique was often used as
features for shape recognition and classification. The
second stage involved classification of plant leaf
images based on the derived feature obtained in the
previous stage. Comparing descriptors of the
unknown object performed classification. The
features were given as input to the neural network
which was based on supervised learning algorithm
having multilayer preceptor with feed-forward back
propagation architecture. The result of this ranged
from 90-94% accuracy having mean square error in
the range of 0-0.9.
Stephen Gang Wu et al. [9] worked using PNN
(Probabilistic Neural Network) for classification of
leaf images based on 12 leaf features. A principal
component analysis was used for reducing the 12
dimensions into five dimensions for faster
processing. The 12 features used were physiological
length, physiological width, leaf area, and leaf
perimeter, smooth factor, aspect ratio, form factor,
rectangularity, narrow factor, perimeter ratio of
diameter, perimeter ratio of physiological length and
physiological width, and vein features. If a Standard
size image was not used then the features of the same
leaf vary with different sizes images. Since, it was the
same leaf, the value of all the features were expected
to be same. The values for these features vary for
different scaled versions of the same leaf image. In
this research, features used were scaling, translation
and rotation invariant.
Y. Nam et al. [10] studied a shape-based leaf
image retrieval system using two novel approaches.
These were named as, an improved maximum power
point algorithm and a revised dynamic matching
method. In his approach, they used a region-based
shape representation technique to define the shapes of
the leaf. Certain points of interest were evaluated and
the distances between the points were taken as the
features for image matching.
Bivashrestha et al. [11] studied various features
related to the shape of the leaves for leaf image-based
II. LITERATURE SURVEY
This section provides some discussion about
several applications on feature extraction on image
processing techniques in agriculture available today.
LiliMa et al [4] in their work studied the Nutrient
deficiency found in soya bean plant by considering
six stages of growth in soya bean plant with the
application of nitrogen fertilizer that ranges between
0%-150% in each stages . The noises were removed
and the object of interest was separated between the
backgrounds using minimum error threshold method.
The color appearances of soya bean leaves were
evaluated using RGB and HIS color space models.
The extinct difference between the normal and the
leaf with variation was found using image processing
method.
R.A.DL.Pugoy et.al [5] presented in their paper
about rice, which is one of the domestic plant in
worldwide and analysis in this crop, is very
important. The various nutrients deficiency, nutrients
toxicity and the type of diseases from the leaves of
rice plant using image processing technique
comprising clustering and spatial analysis in order to
obtain the color variations on leaf were analyzed.
Maicon A. Sartin et al [6] designed image
processing system for precision agriculture to
improve the agriculture production system in cotton
plant to find macro nutrients deficiency especially
10
plant classification. The features selected in this
automated plant identification were isoperimetric
quotient, convex hull ratio, aspect ratio and
eccentricity. These features were vital in identifying
leaf shapes, number of lobes and margin types.
Leaves of nine classes were used and a classifier was
implemented and tested using 105 leaves of 38
different species. K-means clustering was used to
classify 105 leaf images into nine clusters.
R. Janani et al. [12] proposed a method for the
extraction of shape, color and texture features from
leaf images and training an artificial neural network
(ANN) classifier to identify the exact leaf class. The
key issue lies in the selection of proper image input
features to attain high efficiency with less
computational complexity. They tested the accuracy
of network with different combination of image
features. The tested result on 63 leaf images revealed
that their method gave the accuracy of 94.4% with a
minimum of eight input features. Image processing
and neural network toolboxes were used in
MATLAB to implement the work.
Zheng Niu et al [13] with four wavelengths and a
supercontinuum laser as a light source was designed
to monitor the fine structure and the biochemical
parameters of vegetation.
Pawan et al [14] in their proposal initially
preprocessing the input image using histogram
equalization is applied to increase the contrast in low
contrast image, K-means clustering algorithm is used
for segmentation which classifies objects based on a
set of features into K number of classes and finally
classification is performed using Neural-network.
Thus image processing technique is used for
detecting diseases on cotton leaves early and
accurately. It is used to analyze the cotton diseases
which will be useful to farmers.
In this paper, the salient features of leaves were
studied. Recognition was studied to develop an
approach that produced the better feature extraction
technique. This work was done on maize leaf and
extracting the colors, texture and shape features. It
was based on local region leaves and was tried to get
better accuracy results as compared to previous
works.
III. METHODOLOGY
Image
Acquisition of
Maize Leaf
Image
Preprocessing
Images
Feature Extraction
Color Feature
Extraction
Shape Feature
Extraction
Texture Feature
Extraction
Nutrient Deficiency
Identification using
Multivariate Least Partial
Square
Figure 1 Block Diagram of the Proposed Maize Leaf
nutrition deficiency Recognition System
The proposed architecture of this research work is
shown in the figure 1. The dataset used in this work
is maize leaf images according to the
recommendation given by AgriPortal of Tamilnadu
Agricultural University, Coimbatore [15]. The
images are preprocessed prior to the identification of
presence of deficiency. Initially the given input
images are converted to HSV Color Transformation
to determine the intensity and color variance near the
edges of the objects. Next Histogram equalization is
applied to produce gray map which increases the
intensities can be better distributed on the histogram.
Different features of the leaf are obtained and these
features are given as input to the Multivariate Least
Partial Square. Training and testing of the proposed
work is carried out and the result is obtained.
The features used in leaf recognition are described
as follows:
Block diagram of the proposed method is shown in
Figure1.
11
A.
Shape Features
Moment 3: It is called Skewness. It gives measure of
the degree of asymmetry in the distribution.
There are several kinds of geometric features
involved as shape features:
Aspect ratio: The ratio of maximum width to
major axis length.
Width ratio: The ratio of width at half of
major axis to maximum width.
Apex angle: The angle at apex between leaf
edges on width line at 3/4th of the major
axis.
Apex ratio: The ratio of the angle made at
the apex by width at 9/10th of major axis to
angle at 3/4th of major axis.
Base angle: The angle at base between leaf
edges on width line at 1/4th of the major
axis.
Moment ratio: The ratio between Y
deviation and X deviation from the centroid
of the leaf. It accounts for the mass
distributed around the leaf, whether it is in
longitudinal or lateral direction. If the value
is greater than one, the leaf is longitudinal in
nature; and vice versa. Whereas, a value
equal to one classifies it as circular.
Circularity: The ratio of 4*PI*Area of the
leaf to the square of perimeter.
Moment 4: It is called Kurtosis. It gives the measure
of peakedness in the distribution.
C. Texture Features
An image texture is a set of standard of
measurements computed in image processing
intended to enumerate the apparent texture of a leaf
image. Leaf image texture gives information
regarding the spatial arrangement of colour or
intensities in a leaf image or selected region of a leaf
image. Texture features are extracted from Gray
Level Co-occurrence Matrices (GLCMs). GLCM is
very useful to obtain valuable information about the
relative position of the neighbouring pixels in an
image [7]. The co-occurrence matrix GLCM (i,j)
counts the co-occurrence of pixels with gray value i
and j at given distance d. The direction of
neighbouring pixels to represents the distance can be
selected, for example 135o, 90o, 45o, or 0o, as
illustrated in Figure 2 .
B. Color Features
Color features are essential for colour based image
analysis. The information of colour distribution in an
image can be captured by the low order moments.
The first & second order moment has been proved to
be efficient and effective in representing colour
distribution of image.
 Moment 1: It is called Mean. It provides average
Color value in the image.
Where MN represents the total number of pixels in
the image
Figure 2: Directions in calculating GLCM
 Moment 2: It is called Standard Deviation. The
standard deviation is the square root of the variance
of the distribution
D. Multivariate Partial Least Square (MPLS)
Multivariate data analysis is the simultaneous
observation of more than one characteristic. In
contrast to the analysis of univariate data, in this
approach not only a single variable or the relation
between two variables can be investigated, but the
12
relations between many attributes can be considered.
In general consider an multivariate linear regression
model with an n× m matrix X of predictors and an n
× p matrix Y of dependent variables; both matrices
are mean-centered. Similar to Principal component
regression (PCR), approximate
or
In this equation T and U are the respective score
matrices that consist of linear combination of the x
and y variables. Then P and Q are the respective
loading matrices of X and Y. Note that, unlike in
PCR, in PLS in general not the loadings are
orthogonal, but the scores a ≤ min(n; m) is the
number of PLS components. Additionally, the x- and
y-scores are related by the so-called inner relation a
linear regression model maximizing the covariance
between the x- and y-scores.
The regression coefficients are stored in a diagonal
matrix and the residuals in the matrix H.
Figure 3: Flowchart of MLPS
The inner relation can destroy the uniqueness of the
decomposition of the data matrices X and Y.
Normalization constraints on the score vectors t and u
avoid this problem. To fulfill these restrictions we
need to introduce (orthogonal) weight vectors w and
c such that
The figure 3 shows the overall work flow of MLPS in
which it first generate the matrix and convert it into
linear regression, next it generate the diagonal matrix
for regression then to normalization process is
performed for finding the maximum covariance
which helps in prediction of deficiency part of the
input maize leaf image.
With
IV. EXPERIMENTAL RESULTS
Consequently, here the score vectors do not result
from projection of X on loading vectors but on
weights.
The experimental process of nutrient deficiency
identification using Multivariate Partial Least Square
(MPLS) using enhanced feature extraction has been
The objective function of PLS regression can then be
written as
deployed using MATLAB. Control and deficient leaf
images are collected from experimental fields of
Tamilnadu Agricultural University, Coimbatore [13].
Total 60 images are collected from open source, in
13
those 30 images for training purpose and 30 images
for testing purpose.
Table1: Performance Comparison of Proposed Work
with exisitng Approaches based on Precision and
Recall
Measures
Multivariate
Image
Analysis
Original Image
RGB to HSV
Multivariate
Least
Partial
Square
image
analysis
Enhanced
Multivariate
Least
Partial
Square
image
analysis
Precision
0.75
0.86
0.96
Recall
0.68
0.79
0.94
Deficiency Part
Histogram Plot
Figure 4: Shows the output the proposed work
A. Performance Evaluation Metrics for Enhanced
MLPS Systems
In EMLPS, the most commonly used performance
evaluation measures are Precision and Recall [10].
Precision is defined as the ratio of the number of
relevant images retrieved to the total number of
images retrieved.
Figure 5 :Comparison of Proposed Work with
exisitng Approaches
P = Number of relevant maize leaf images retrieved
total number of maize leaf images retrieved
The result shows that the proposed work performs
better due to the potential feature extraction
techniques based on color, shape and texture. Thus it
shows the importance and selection of appropriate
feature extraction technique for better analysis of
nutrition deficiency in maize leaf images.
R = Number of relevant maize leaf images retrieved
total number of maize leaf images in the database
The table1 shows the precision and recall value of
proposed enhanced Multivariate least partial square
image analysis with other two existing approaches
namely multivariate image analysis and traditional
multivariate least partial square image analysis.
14
With Artificial Neural Network For Nutrient
Deficiency In Cotton Crop‖. Journal Of
Computer Science 10 (6): 1084-1093.
V. CONCLUSION
Incorporating feature descriptors is a feasible
alternative for classifying structurally complex
images. They offer exceptional invariance features
and reveal enhanced performance than other moment
based solutions. They can be computed in parallel
and as the computational performance of computers
increase; the time necessary for their calculation
perhaps will not be a problem in the nearby future.
The database is prepared for the experimental use.
The database contains various leaves with various
shapes, colours and size. Experiment is carried out
with the different maize leaves of different classes
and tested. This paper investigated the use of a
number of different colours, texture and shape
features for maize leaf nutrition deficiency
identification system. The features obtained from
colour moments can be further used for indexing
images based on colour. It can be further used for
classification of leaves into respective nutrition
deficient using Multivariate Partial Least Square
image analysis.
[7] Marian
Wiwarta,
Gabriel
Fordon
ski,
KrystynaZuk-Gołaszewska.
2009.
―Early
diagnostics of macronutrient deficiencies in three
legume
species
by
color
image
analysis.Computers
and
Electronics
in
Agriculture‖ 65: 125–132.
[8]
Jyotismita Chaki, Ranjan Parekh, ―Plant Leaf
Recognition using Shape based Features and
Network classifiers‖ (IJACSA) International
Journal of Advanced Computer Science and
Applications, Vol. 2, Year 2011.
[9] Stephen Gang Wu, Forrest Sheng Bao, Eric You
Xu, Yu-Xuan Wang, Yi-Fan Chang and QiaoLiang Xiang, ―A Leaf Recognition Algorithm for
Plant Classification Using Probabilistic Neural
Network‖,arXiv:0707.4289v1 [cs.AI], Year 2007.
[10] Y. Nam, ―Elis: An efficient leaf image retrieval
system‖, in Proc. Advances in Pattern
Recognition Int. Conf., Kolkata, India. Year 2005
[11] Bivashrestha, ―Classification of plants using
images of their leaves‖ ICPR, vol. 2, pp. 25072510, Year 2000.
REFERENCES
[12] R. Janani, A. Gopal, ―Identification of selected
medicinal plant leaves using image features and
ANN‖, Advanced Electronic Systems (ICAES),
2013 International Conference, Page(s): 238 –
242, Year 2013.
[1] M K HU, ―Visual pattern recognition by moment
invariants‖, IRE Trans. Info Theory, vol. 8, pp.
179-187, Year. 1962.
[2] T. H. Reiss ,―The Revised Fundamental Theorem
of Moment Invariants‖, IEEE Transaction Pattern
Anal. Machine Intelligence, Vol. 13, No. 8, Page
830-834, Year. 1991
[13] Zheng Niu, Zhigang Xu, Gang Sun, Wenjiang
Huang, Li Wang, Mingbo Feng, Wang Li,
Wenbin He, and Shuai Gao, ―Design of a New
Multispectral Waveform LiDAR Instrument to
Monitor Vegetation‖, IEEE Geoscience And
Remote Sensing Letters, vol. 12, no. 7, july 2015
[3] J. Du, D. Huang, X. Wang, and X. Gu,
―Computer-aided plant species identification
(CAPSI) based on leaf shape matching
technique,‖ Transactions of the Institute of
Measurement and Control. Vol. 28, No.3 Page.
275-284, Year 2006.
[14] Pawan P. Warne, Dr. S. R. Ganorkar, ―Detection
of Diseases on Cotton Leaves Using K-Mean
Clustering,‖ International Research Journal of
Engineering and Technology (IRJET), Volume:
02 Issue: 04 , July-2015
[4] Lili Ma, Junlong Fang, Yuehua Chen, Shuipeng
Gong. 2010. ―Color Analysis of Leaf Images Of
Deficiencies And Excess Nitrogen Content In
Soybean Leaves‖. E-Product E-Service And EEntertainment (ICEEE), Vol., No., 1-9.
[15] URL: http://agritech.tnau.ac.in
[5] PugoyR.A.Dl.and
V.Y.Mariano,
2010.
―Automated Rice Nutrient Deficiency, Nutrient
Toxicity And Disease Detection Using Digital
Image Analysis‖. 40th Cssp Anniversary And
Annual Scientific Conference, Davao City, 1520.
[6] Maicon A. Sartin, Alexandre C.R. Da Silva and
ClaudineiKappes, 2014. ―Image Segmentation
15
Low Power Multiplier Using Algorithmic Noise
Tolerant Architecture
S.Fathima
Dr.S.Uma
PG scholar, Dept. of ECE
Coimbatore Institute of Technology
Coimbatore, India.
Fathimait2014@gmail.com
Associate Professor, Dept. of ECE
Coimbatore Institute of Technology
Coimbatore, India.
Uma@cit.edu.in
Abstract—In this paper, Fixed-width Reduced Precision
Replica Redundancy Block (RPR) design is adopted in
Algorithmic Noise Tolerant (ANT) architecture. The main
objective of this design is to achieve High Precision, Low power
and high speed multiplier. By using Fixed- width RPR
computation error can be identified and corrected. The ANT
architecture was implemented using Xilinx and ModelSim tool.
Performance metrics of power is compared with existing Fullwidth RPR designs. Average Power consumption of the proposed
design is found to be 74.42µW at 100Hz. Proposed design achieves
less power and area efficiency than the Full-width RPR.
II. DESIGN METHODOLOGY OF ANT ARCHITECTURE USING
FIXED-WIDTH RPR.
Unlike Full-width RPR which increases the bit width, Fixedwidth RPR avoids infinite growth of bit-width. ANT
architecture as shown in Fig 1, includes Main Digital Signal
Processor (MDSP) and Error Correction (EC) block [1]. To
reduce the power consumption in Main block VOS is used. But
in VOS whenever critical path delay Tcp is greater than
sampling period Tsamp, soft error occurs which leads to severe
degradation of incoming signals. EC block contains reduced
computation delay. Under VOS, there are number of inputs
x[n] and its corresponding output of Main block is ya[n] and
output of Fixed-width RPR is yr[n]. Error detection is carried
out by comparing the difference between |ya[n]-yr[n]| with the
threshold value Th.
Index Terms—ANT architecture, Fixed-width RPR, Fullwidth RPR.
I. INTRODUCTION
Rapid growth of wireless devices in recent years leads to
need for ultra low power systems. To reduce the power
consumption, low power technique such as supply voltage
scaling [1] can be used, However, Noise interference problem
arises. Hence the design technique is to be enhanced to avoid
noise interferences [2].
Final output [n] = ya[n], if |ya[n]-yr[n]| ≤ Th
yr[n], if |ya[n]-yr[n]| ≥ Th
[1]
Threshold is calculated using the formula
Th =
[2]
Where yo[n] is error free output signal. By using this technique
power can be minimized without any signal degradation.
Multipliers are key components of many high performance
systems such as FIR filters, microprocessors, digital signal
processors, etc. A system’s performance is generally
determined by the performance of the multiplier[3] because the
multiplier is generally the slowest element in the system.
Low power techniques such as voltage overscaling[4] has
been proposed to lower the supply voltage without affecting
the throughput. But it leads to more degradation in Signal to
Noise Ratio (SNR). ANT technique [2] uses VOS with RPR to
reduce the soft errors and achieves high energy. RPR design in
ANT architecture [5] is not easily adopted and repeated.
However adopting with RPR increases the area and power
consumption.
To overcome these problems, Full-width RPR is replaced
with Fixed-width RPR in the proposed method. [9] By using
Fixed-width RPR computation error can be identified and
corrected with low power consumption and less area.
Fig 1. Proposed ANT architecture with the Fixed-width
RPR[1].
This paper is presented as follows: Section II focuses on
design methodology of ANT architecture using Fixed-width
RPR. Section III elucidates the simulation of Fixed-width
RPR. Section IV narrates about the simulation results and the
comparisons and concluded in Section V.
Fixed-width RPR is applied in DSP applications to avoid the
size of growing bit-width. In Fixed-width DSP neglecting LSB
output is a popular solution. The hardware complexity and
16
power consumption of Fixed-width RPR is 50% lesser than the
Full-width RPR[6]. However, neglecting LSB bits results in
rounding error.
III. SIMULATION AND RESULTS OF FULL-WIDTH AND FIXEDWIDTH RPR
A. Error Compensation for Fixed-width RPR:
The Function of RPR is to identify and correct the error
which occurs in the output of MDSP and maintain the signal to
noise ratio by reducing the supply voltage[5]. By using Fixedwidth RPR, computation speed can be accelerated as compared
with Full-width RPR.
Fig 3. Output Wavefom of Full-width RPR.
Fig 3 explains about the simulation of Full-width RPR which
contain 12 bit with the threshold value of 10 bit. If the critical
path delay exceeds over 10 produces an error, otherwise
normal operation takes place without any error.
Fig 2. 12 x 12 bit ANT multiplier is implemented with the sixbit Fixed-width RPR.
In MDSP of 12 bit ANT Baugh-Wooley array multiplier.
That too unsigned n bit inputs of x and y is given by,
X=
xi . 2i ,
Y=
yj . 2j
[3]
The multiplied output P is the summation of partial
products of xi . yj which is given in the equation,
P=
P k . 2k =
xi yj . 2i+j
[4]
Fig 4. Gatecount of Full-width RPR.
Full-width Baugh-Wooley divided into four parts. They are
Most Significant Part (MSP), Input Correction Vector (ICV),
Minor Input Correction Vector (MICV), Least Significant Part
(LSP) as shown in Fig 3.
Fig 4 explains about the area occupied by the Full-width RPR
is 8,571 total equivalent gate count for design.
[5]
Pt =
yj 2j
xi 2i + f(EC)
[6]
Where f(EC) is the error compensation function which is equal
to the summation of f(ICV) and f(MICV).
Most of the Errors are produced at Least Significant Part, so
neglecting the LSP leads to reduction of area, which further
reduces power consumption and hardware complexity. If the
power consumption is lowered, the SNR[6] can be maintained
without severe degradation.
Fig 5. Power Analysis of Full-width RPR
17
Fig 5 explains about the power consumption of Full-width
RPR. The power achieved at 50 MHz clock frequency is 50.97
mW.
Fig 8. Power Analysis of Fixed-width RPR
Fig 8 explains about the power consumption of Fixed-width
RPR. The power achieved at 50 MHz clock frequency is 49.51
mW.
Table I
Comparison of Power Analysis of Full-width and Fixedwidth RPR
Clock
Frequency(in
MHz)
50
100
150
200
Fig 6 Output Wavefom of Fixed-width RPR.
Fig 6 explains about the simulation of Fixed-width RPR which
contain 12 bit with the threshold value of 10 bit. If the critical
path delay exceeds over 10 produces an error, otherwise
normal operation takes place without any error.
Full-width
RPR(Power in
mW)
50.97
77.35
103.72
130.1
Fixed-width
RPR(Power in
mW)
49.51
74.42
99.33
124.24
By analyzing the Table 1, Proposed method consumes less
power than existing for various frequencies.
Fig 7. Gatecount of Fixed-width RPR.
Fig 9. Bar chart of Full-width and Fixed-width RPR at various
frequencies.
Fig 7 explains about the area occupied by the Fixed-width RPR
is 7,716 total equivalent gate count for design.
Fig 9. Shows that for various clock frequencies, the power
achieved by Fixed-width RPR is Lesser than the Full-width
RPR.
At the clock frequency of 50 MHZ, the power consumed by
the full width RPR is 50.97mW whereas 49.51mW in Fixedwidth RPR for the same 50MHZ clock frequency. While
comparing with the various frequencies of clock, the power
also get reduces.
18
IV. CONCLUSION
The proposed 12 bit ANT multiplier circuit using Fixedwidth RPR is implemented using Xilinx ISE 8.1 under 1v
supply voltage and operating at the frequency of 50, 100, 150,
200 MHz respectively. The power consumption achieved at 50
MHz is 49.51 mW.
.In comparison with Full-width RPR, the circuit area and
the Power consumption is drastically reduced. This type of
ANT architecture has higher computation precision, higher
SNR and can also be operated at low supply voltage.
REFERENCES
[1] V. Gupta, D. Mohapatra, A. Raghunathan, and K. Roy,
“Low-power digital signal processing using approximate
adders,” IEEE Trans. Comput. Added Des. Integr. Circuits
Syst., vol. 32, no. 1, pp. 124–137, Jan. 2013.
[2] G. Karakonstantis, D. Mohapatra, and K. Roy, “Logic and
memory design based on unequal error protection for
voltage-scalable, robust and adaptive DSP systems,” J.
Signal Process. Syst., vol. 68, no. 3, pp. 415–431, 2012.
[3] P. N. Whatmough, S. Das, D. M. Bull, and I. Darwazeh,
“Circuit-level timing error tolerance for low-power DSP
filters and transforms,” IEEE Trans. Very Large Scale
Integr. (VLSI) Syst., vol. 21, no. 6, pp. 12–18, Feb. 2012.
[4] J. N. Chen and J. H. Hu, “Energy-efficient digital signal
processing via voltage-overscaling-based residue number
system,” IEEE Trans. Very Large Scale Integr. (VLSI)
Syst., vol. 21, no. 7, pp. 1322–1332, Jul. 2013.
[5] N. Petra, D. D. Caro, V. Garofalo, N. Napoli, and A. G. M.
Strollo, “Truncated binary multipliers with variable
correction and minimum mean square error,” IEEE Trans.
Circuits Syst., Vol. 57, No. 6, pp. 1312–1325, Jun. 2010.
[6] S. J. Jou and H. H. Wang, “Fixed-width multiplier for DSP
application,” in Proc. IEEE Int. Symp. Comput. Des., Vol.
23, No.5, pp. 318–322, Sep. 2010.
[7] I. C. Wey and C. C. Wang, “Low-error and area-efficient
fixedwidth multiplier by using minor input correction
vector,” in Proc. IEEE Int. Conf. Electron. Inf. Eng.,
Kyoto, Japan, Vol. 1, pp. 118–122, Aug. 2010.
[8] Y. Pu, J. P. de Gyvez, H. Corporaal, and Y. Ha, “An ultra
low energy/frame multi-standard JPEG co-processor in 65nm CMOS with sub/near threshold power supply,” IEEE J.
Solid State Circuits, Vol. 45, No. 3, pp. 668–680, Mar.
2010.
[9] Y. Liu, T. Zhang, and K. K. Parhi, “Computation error
analysis in digital signal processing systems with
overscaled supply voltage,” IEEE Trans. Very Large Scale
Integr. (VLSI) Syst., Vol. 18, No. 4, pp. 517–526, Apr.
2010.
19
A Survey on Compression in Encrypted Domain
T.Suguna
Dr.R.ShanmugaLakshmi
Department of Information Technology
Government College of Technology
Coimbatore, India
tsugunait@gct.ac.in
Department of Computer Science and Engineering
Government College of Technology
Coimbatore, India
drshanmi@gct.ac.in
First the sender Alice compresses the image I and
encrypts the compressed image using an encryption function
with secret key Ek(Ic). The resultant image is transferred to the
channel provider Charlie for transmitting this to receiver.
Channel provider Charlie act as a forwarder and forward the
compressed encrypted image to the receiver. Then the receiver
Bob performs decryption and decompression sequentially to
get back the original image send by the sender.
Eventhough this communication provides the secured
transmission of image, processing is fully done in the sender
side. No processing is required by the channel forwarder. In
some situation sender may be a resource restricted device. It
has limited processing capability. And also sender
concentrates only on the security of the image and not bothers
about the efficient transmission of the image. In that scenario
some compression processing is handed over to the channel
provider. For this the order of applying encryption and
compression should be changed. This system is called as
Encryption Then Compression (ETC) system. This is shown in
the following fig. 2.
Abstract—In Multimedia communication, there is a need of
large data to be transmitted through the communication channel.
To provide better efficiency of the channel, we need compression
scheme to reduce the quantity of information to be sent. In
addition to efficient transmission there is a need for secure
transmission also. To provide both security and data compaction
we need both compression and encryption. In traditional method
first compressions is applied and then apply the encryption
technique on the compressed domain. In some situation like
resource constraint environment, user needs more attention to
security of their data than compression. In this situation we need
a reverse system of the traditional compression then encryption
(CTE) system called as encryption then compression system
(ETC). In recent years a number of different methods were
proposed in this area. This paper discuss about the different
methods used for ETC system .
Index Terms—Compression,
encrypted images
encryption,
compression
of
I. INTRODUCTION
Consider a multimedia communication scenario, in which a
content owner Alice wants to securely and efficiently transmit
an image I to a recipient Bob through the untrusted
communication channel provider Charlie. Traditionally this
could be done by as shown in fig. 1
Fig. 2. Encryption Then Compression system
In this ETC system Alice only encrypt his image
using the secret key and it is transferred to the channel
provider Charlie. He was not aware of the key, so he
compresses the encrypted image. Compressed image is then
forwarded to the receiver Bob. He performs decompression
and decryption to retrieve the original image.
Fig. 1. Compression Then Encryption system
20
II. METHODS USED FOR ETC SYSTEM
Predictive encoding and encryption
For predictive encoding both non-linear and linear
predictors are available. Compared to linear predictor
suggested in JPEG spatial, non-linear predictor suggested in
JPEG-LS provides better performance for most of the natural
images. Hence in this method non-linear prediction is used.
For prediction four adjacent pixels are used. Encryption is
done in the prediction domain by performing the following
operation.
Ci=∆i+Ki
A. On Compressing Encrypted Data
Johnson et. al [2] focuses both lossless compression of
encrypted bi-level images and lossy compression of encrypted
real valued data’s separately. Distributed source coding theory
is used in this work.
Lossless Compression of Encrypted Bi-level Images
Image is considered as strings of 0’s and 1’s. The string is
encrypted by adding a unique Bernoulli string of the
appropriate length. The encrypted string is compressed by
finding its syndrome with respect to Low Density Parity
Check codes (LDPC). In the receiver side decoding and
decompression is done simultaneously using iterative
decoding algorithm specified by slepian-wolf method [1].
where, ∆i- prediction error
Ki- key
+ - XOR operation.
Decryption is performed by performing XOR operation with
cipher text and the key. Same key is maintained in both the
sides by sending random seed as side information to generate
the key for decryption.
Lossy Compression of Encrypted Real-Valued Data
The data was encrypted by a stream cipher. Encryption is
done by adding Gaussian key sequence with the data sequence
on a sample by sample basis. The encrypted data has been
compressed by the scalar quantizer and trellis code to the rate
of 1 bit/sample. Decoder has the knowledge of the key
sequence and trellis decoder. Using these two estimates
decoder compute the encrypted sequence. Form this key
sequence is subtracted to form the optimal sequence of the
original data. Encoder and decoder performance is purely
depends on the source not on the side information (key). It
achieves Distortion rate of 3-3.8 dB and Bit error rate of 10-3
to 10-4
Compression and decompression
DSC coding with irregular LDPC codes are used for
compression. This considers both gray scale and color images
for their experimental analysis. It provides lossless
compression of encrypted images. Color image is separated
into 3 color planes and apply the procedure which is applied
for grey scale image. Compression ratio achieved using this
scheme is comparatively higher than the compression of
encrypted image directly.
D. Lossless compression of encrypted grey-level and color
images
Spatial and cross plane correlation between pixels and
Correlation between color bands are used to compress the
encrypted grey-level and color images [5].
B. On compression of encrypted images
This work [3] follows the graphical model for compression
of encrypted images. This graphical model consists of three
components connected together. They are source model,
encryption model and compression model. Source model is a
2-D spatially correlated model. Which includes a image pixel
and its corresponding four neighboring pixels such as up and
down, left and right. Second model is the encryption model.
Stream cipher technique is used for encryption. Encrypted data
is obtained by performing exclusive or operation with key and
data. For compression it uses the code model, which includes
linear transformation matrix. This matrix is used for the
conversion of the encrypted data into compressed form.
Decoding is achieved by applying sum-product algorithm
on the graph model. This algorithm is an inference algorithm
and it is applied iteratively to updates an estimate of the
distribution for each of the variables. It compresses an
encrypted 10000 bits bi-level image into 4299 bits and
recovers the image exactly. It is better than the models uses
the 1-D sources.
Grey level image
Subdivide the image into bit-planes and consider each of
them as a separate black and white image. Before performing
encryption the image is stored, in such a way that spatial
correlation between bit-planes is removed. Two methods are
used to remove spatial correlation between bit-planes. First
method is a fast and simple method, in which each row is
scanned from left to right and perform XOR operation
between current bit and previous bit. Using this method
vertical edges are extracted.
Second method is very effective method applied on the
grey level values. In this method also each row is scanned
from left to right but instead of XOR operation, prediction
method is used with average of four adjacent bits to predict the
value. Then the pixel values are replaced by prediction error.
The image with the prediction error is then divided into bitplanes. To decide compression rate conditional entropy
between adjacent bit-planes is used. Probabilities are given as
the input to perform decoding and decryption.
C. Distributed Source Coding based Encryption and
lossslessCompression of Gray Scale and Color Images
Encryption is done on the prediction error domain instead
of the original image and compression is achieved by
distributed source coding [4].
Color Image
In color image each band is treated as gray scale
image. To get better compression rate, correlation between
21
reference band and other bands are used. Green color band is
selected as reference band and it is coded as grey level image.
In another approach RGB color space to YCbCr color space.
Reversible integer approximation of the correct transform is
used to attain lossless nature of the scheme while transforming
from RGB to YCbCr color space. Both XOR based and
prediction based spatial de-correlation can be applied to the
modified YCbCr and each band now treated separately as a
grey scale image for further processing.
Encoding, encryption and decoding follow the
method that is used by Johnson et al.[2] with small variation.
In encoding LDPC code is characterized by matrix having
number of rows is length of the bit planes and number
columns is length of the syndrome. Before to encryption
conditional probabilities of 0s and 1s are calculated with
respect to their reference plane and it is transmitted with the
encrypted bit-planes. Encrypted creates a file containing
header and description blocks to decide compression rate.
Decoding and decryption performed jointly. Decoder uses BP
algorithm. Each received bit plane is decoded with reference
plane and conditional probability. Each decoded bit-plane is
decrypted immediately to decode the subsequent bit-planes.
Compression
Compression of encrypted image is done with the
help of the measurement matrix Φ of size MxN. Where M<N
and that decides the compression ratio.
Y = ΦXenc
In addition to compression, compressive sensing is provide a
second level of security to the image. After compression Y
will be quantized using Lloyd quantizer to obtain Yq. This
will be send to the decoder.
Joint decoding and decryption:
Keys for encryption and compression are received via
secure channel and Yq is received via normal channel by the
decoder. Using these three values joint decoding and
decryption is done to obtain the original image.
F. Lossy Compression and Iterative Reconstruction for
Encrypted Image
In this paper pseudo random permutation based encryption
method is used to obtain the cipher text [7]. Cipher text is
compressed by removing the excessive rough and fine
information available in the transform domain. Then combined
decoding and decryption is performed by the spatial
correlation exist in the natural image. This is refined by the
iterative reconstruction process. The reconstruction process is
shown in fig.4
E. Lossy Compression of Encrypted Image by Compressive
Sensing Technique
Compressive sensing itself is provides better security for
user data. Encryption is performed in the spatial domain with
linear operations [6]. Entire process is as shown in the fig.3
Fig.3 ETC using compressive sensing
Encryption and compression is done by generating
two random matrix called as encryption matrix and
measurement matrix and the seed value for that is
communicated to the decoder through the secure channel.
Using this, the decoder decrypts and decodes the received
information.
Fig. 4 Reconstruction process
G. On Compression of Data Encrypted With Block Ciphers
Usually encryption is done in two ways. It may be stream
cipher or block cipher. In stream cipher sequence of bits are
considered for encryption using XOR operation. But in case of
block ciphers block of codes are used. AES (Advanced
Encryption Standard) and DES(Data Encryption Standard) are
examples for block cipher encryption. In block ciphers there is
practical limitation of mapping two separate plain text block to
a single cipher block. So in this method block cipher with
chaining mode like CBC(cipher block chaining) and Output
Feedback (OFB) or Cipher Feedback modes are used. Post
Encryption
Cipher text Xenc is generated by converting the given
image MxN into single dimensional matrix X of size N and
then apply the linear transformation. The process of
encryption is shown by the following relation.
Xenc = μX
Where μ is the matrix of size NxN.
22
Encryption Scheme (PES) based on the Slepian-Wolf coding
is used with block cipher chaining modes to perform
encryption.
Compression is done by exploiting the correlation
between the plain text and cipher text without having any
knowledge about the key. Initialization vector and cipher text
blocks are given as the input to the compressor. Compressor
applies Slepian-Wolf encoding procedure to compress the
input except the last cipher block. Block diagram of the
compresser is shown if fig.5
Compression
Compression of encrypted images is done by adaptive
Arithmetic Coding. The channel provider applies the adaptive
arithmetic coding for the received prediction error sequence
and generate binary bit streams. Length of the bit stream is
sent as side information for decryption and decompression.
Decompression and decryption:
Decompression and decryption is performed jointly at the
receiver side using the side information. Receiver divides the
received bit streams into L clusters then applies adaptive
arithmetic decoding and de permutation to get back the
original cluster. With all the cluster values, decoding of each
pixel is done in raster scan order. Bytes and bits per pixel are
used as performance measure. Very small (<0.2%) amount of
coding penalty is incurred.
J. Designing
an
Efficient Image
Encryption-ThenCompression System via Prediction Error Clustering and
Random Permutation
Image encryption is performed on prediction error domain
and random permutation [11]. Adaptive arithmetic coding is
used to losslessly compress the encrypted image.
Fig. 5 Block diagram
Receiver uses joint decompression/decryption
method to get the original information. First Slepian-Wolf
decoding and decryption procedure is applied from right side
block that is from last block to the left block that is
initialization vector to get the original information.
Image encryption
The schematic diagram of the image encryption is as shown
in the fig. 6
H. Compression of encrypted images with multilayer
decomposition
Multi layer decomposition [9] is used for lossy compression
of encrypted gray scale images.
Encryption operation is starts from dividing images into sub
images. Prediction errors are calculated for different layers of
the image. Then the sub image and prediction errors are
encrypted by exclusive-or operation and a pseudo random
permutation. Channel provider compresses the encrypted
images by performing quantization and optimization with rate
distortion criteria on various layers of the permuted prediction
errors. At the receiver, compressed encrypted image is
dequantized and decrypted with the help of the key known by
the receiver.
Fig.6 Image Encryption
Input image is given to the GAP predictor, which predicts
the error. Instead of encoding original pixels, here prediction
errors are encoded. To encode this, the prediction errors are
mapped to one of the values [0-255]. Then these mapped
values are divided into L number of clusters and apply the
random permutation to obtain the encrypted image.
I. On the design of an efficient encryption-then-compression
system
Encryption
In this paper image encryption is done by prediction error
clustering and random permutation [10]. Prediction of each
pixel is made by using image predictor GAP because of its
excellent de-correlation capability. Predicted value Īi,j of each
pixel Ii,j is refined by feedback mechanism to obtain correct
prediction value Ĩ i,j. prediction error associated with each pixel
is calculated by
e i,j= Ii,j - Ĩ i,j
Instead of applying encryption in the prediction domain as a
whole, it divides the prediction errors in to L clusters. Then
each cluster is reshaped and permuted by cyclic shifts. Finally
encrypted image is generated by concatenating all the
permuted clusters.
Lossless compression
Channel provider does not know the encryption key used by
the sender. So to compress the encrypted image cluster
information is passed as the side information. Using this
compression takes place adaptive arithmetic coding. First
predicted error sequence is converted into binary bit streams
parallel to improve the performance. Binary bit streams are
combined to form bit sequence. The length of the bit stream is
sent as side information to the receiver Bob.
23
Sequential Decompression and Decryption
Upon receiving compressed encrypted bit stream, for each
bit stream Bob applies adaptive arithmetic decoding to obtain
permuted predicted bit stream. Then this was de permuted by
the secret key known by the receiver. To obtain the original
reconstructed image raster scan method is used.
quantized and then are encoded by embedded block coding
with EBCOT followed by rate control operations. Rate control
is used to make the bit stream conform to a target size. Finally,
a JPEG 2000 compliant bit stream is generated by adding
packet headers, a main header and other control codes.
Perceptual encryption makes an image difficult to
understand visually. The feature similarity (FSIM) index
between original image I and encrypted image Ie was
calculated to confirm perceptual encryption. The average Peak
Signal to Noise Ratio (PSNR) comparison is done to ensure
efficiency in compression. Receiver performs decompression
and decryption to obtain the original image.
K. A game theory-based block image compression method in
encryption domain
This paper proposes block based compression technique in
encryption domain to provide better security as well as
compatibility to the service providers. Block based
compression is formulated as a game theoretical problem and
it is optimized by game theory [12].
Encryption of the image is done by two steps. First entire
image is divided into blocks and perform the block
permutation. Second step is within each block perform
position permutation. Key derivation function is used to
generate the keys requires to perform inter block and intra
block permutation. Same key generation mechanism is used
by both encoder and decoder. Only seed for key generation is
transmitted to the decoder to generate the required keys.
Decryption is also very similar to the encryption process.
Compression of encrypted image is done by forming game
theoretical problem in the game theory. Game theory is used
to investigate the trade-off between quality of the image and
desired number of bits used to represent it. Each encrypted
block has a player to compete in the game. The strategy of the
player is to control the number of bits used to represent the
compressed stream in such a way that to provide good quality
reconstructed image.
For decompression and decryption along with the
compressed encrypted data, seed of the key is also delivered.
Receiver uses this information to reconstruct the original
image with good quality.
M. An encryption-then-compression system for JPEG
standard
Encryption is done using block based perceptual
encryption that is compatible to use JPEG compression for
encrypted image [14].
Encryption
Divide each color component of a color image
I={IR,IG,IB} into Bx* By blocks respectively. Permute
randomly the divided blocks using a random integer generated
by a secret key K1 (Block Scrambling). Rotate and invert
randomly each block using a random integer generated by a
key K2 (Block Rotation and Inversion). Apply the negativepositive transformation to each block using a random integer
generated by a key K3 (Negative-Positive Transformation).
Shuffle three color components in each block using a random
integer generated by a key K4 (Color component Shuffling).
After applying these four steps, generate the encrypted image
by integrating the transformed block images. The entire
process is shown in the following fig. 7
L. An Encryption-then-Compression system for JPEG 2000
standard
In this paper encryption and compression is based on
the Discrete Wavelet Transformation co-efficient [13].
Encryption
For encryption, the JPEG 2000 compliant DWT is applied to
Image I. Then, the DWT coefficients, C, which are X *Y in
size, are obtained. DWT coefficients C are perceptually
encrypted into Ce with two types of encryption schemes
namely Sign Scrambling and Block Shuffling based on a
pseudo-random number generator (PRNG) with a secret key.
Encrypted DWT coefficients Ce are transformed to a spatial
image by using inverse DWT. The spatial image is linear
quantized to maintain the signal range of the input image, and
then perceptually encrypted image Ie is obtained.
Fig.7 Four steps of encryption
Compression
Compression is done for both gray scale as well as color
images. To encode gray scale image, four basic steps are
followed. They are:
Dividing an image into non-overlapped
consecutive 8* 8 blocks.
Applying 2-D DCT to each block.
Block-based quantizing.
Entropy coding using Huffman coding.
Compression
Encrypted image is then compressed by channel provider. For
performing compression, the subbands are divided into cx *cy
sized code-blocks. DWT coefficients in the code-blocks are
24
Quality of an encrypted and compressed image is almost as
same as that of compressed original image and an encrypted
image has almost the same image quality as another encrypted
image, although they have different visibility. Large Key
Space allows for the increased security of the images.
On the other hand, in the JPEG compression of a color image,
two processing steps are added before the steps of gray scale
image compression. They are
Performing a color- space transformation from RGB
to YCbCr.
Sub-sampling the Cb and Cr components to reduce
the spatial resolutions.
Performing the above steps to the brightness component Y and
the sub-sampled chroma components independently and
combine the result to obtain the compressed image. Image
III. COMPARISON OF DIFFERENT METHODS
Following table Tab.1 shows the comparison different
methods used for compression and encryption with test
images and performance metrics.
Table.1 Comparison Table
Sl.No
1.
2.
3.
4.
5.
6.
7.
Title
On compressing encrypted
data [2]
On compression of encrypted
images[3]
Distributed source coding
based encryption and lossless
compression of gray scale
and color images[4]
Lossless compression of
encrypted gray level and
color images[5]
Lossy
compression
of
encrypted
image
by
compressive
sensing
technique[6]
Lossy compression and
iterative reconstruction for
encrypted image[7]
Compression of encrypted
images with multi-layer
Decomposition[9]
Technique used for
Compression
Distributed source coding
Trellis code
LDPC codes
Distributed source coding
Technique used for
Encryption
Stream cipher
Performance Metrics
Encryption
model
with
stream cipher
Encryption on prediction
error (XOR)
Bi level image
Compression ratio
Gray scale
Color image
YUV color space
Gray scale
Color image
YUV color space
Lena,
Baboon, Peppers and
House, each of size
256×256
Bi level image
XOR with Bernoulli key
Bit /rate
Compressive sensing with
Gaussian
random
measurement
matrix
generated by random
matrix
Pseudo
random
permutation
Permuted
transformation
domain
PSNR
Compression ratio
Orthogonal transform
PSNR
Compression ratio
Lena
Quantization
optimization
exclusive-or operation and a
pseudorandom
permutation on permutation
errors
Prediction Error Clustering
and Random Permutation
PSNR
Compression rate
Lena, Man, Couple
and Lake
PSNR
Lena
Compression ratio
PSNR
Lena
PSNR
Bit rate
Gray scale and color
images
PSNR
Images
from
standard evaluation
material (StEM)
and
arithmetic
Linear
spatial
Designing an Efficient Image
Encryption-ThenCompression System via
Prediction Error Clustering
and
Random
Permutation[10]
Adaptive
coding
9.
A game theory-based block
image compression method
in encryption domain[12]
An
encryption-thencompression system for
JPEG standard[13]
An
Encryption-thenCompression system for
JPEG 2000 standard[14]
Game theory based bit
allocation on image blocks
Block based encryption
JPEG std
Block based
encryption
JPEG 2000
Sign scrambling and block
shuffling of DWT
11.
Bit rate
Entropy
Compression ratio
Multi level LDPC with BP
algorithm
8.
10.
Test image
25
four
step
at the cost of security. There is always a tradeoff between
security and compression efficiency in the compression of
encrypted images. Hence it is always an open issue to provide
better algorithm for compression of secured images.
Compression ratio, PSNR, Bit rate and Entropy are used as
performance metric for compression of images. Especially
compression ratio and entropy are used for lossless
compression. Using compression ratio as a performance metric
[3] and [4] were compared. Both methods considered here uses
lossless compression of encrypted images. Nearly both
methods achieved approximately same compression ratio in the
range of 2 to 2.5. Encryption used in [3] and [4] is simple
stream cipher with XOR operation. Even encryption used here
is simple method, it provides better security using enough key
space.
REFERENCES
[1] D. Slepian and J. K. Wolf, “Noiseless coding of correlated
information sources,” IEEE Trans. Inform. Theory, vol. IT-19,
pp. 471–480, July 1973.
[2] M. Johnson, P. Ishwar, V. M. Prabhakaran, D. Schonberg, and
K. Ramchandran, “On compressing encrypted data,” IEEE
Trans. Signal Process., vol. 52, no. 10, pp. 2992–3006, Oct.
2004.
[3] D. Schonberg, S. C. Draper, and K. Ramchandran, “On
compression of encrypted images,” in Proc. IEEE Int. Conf.
Image Process., Oct. 2006, pp. 269–272.
[4] A. Kumar and A. Makur, “Distributed source coding based
encryption and lossless compression of gray scale and color
images,” in Proc.MMSP, 2008, pp. 760–764.
[5] R. Lazzeretti and M. Barni, “Lossless compression of encrypted
greylevel and color images,” in Proc. 16th Eur. Signal Process.
Conf., Aug. 2008, pp. 1–5.
[6] Kumar and A. Makur, “Lossy compression of encrypted image
by compressive sensing technique,” in Proc. IEEE Region 10
Conf. TENCON, Jan. 2009, pp. 1–6.
[7] X. Zhang, “Lossy compression and iterative reconstruction for
encrypted image,” IEEE Trans. Inf. Forensics Security, vol. 6,
no. 1, pp. 53–58, Mar. 2011.
[8] D. Klinc, C. Hazay, A. Jagmohan, H. Krawczyk, and T. Rabin,
“On compression of data encrypted with block ciphers,” IEEE
Trans. Inf.Theory, vol. 58, no. 11, pp. 6989–7001, Nov. 2012
[9] X. Zhang, G. Sun, L. Shen, and C. Qin, “Compression of
encrypted images with multilayer decomposition,” Multimed.
Tools Appl., vol. 78, no. 3, pp. 1–13, Feb. 2013.
[10] Jiantao Zhou, Xianming Liu and and Oscar C. Au, “ On the
Design of an efficient Encryption then Compression system, ”
ICASSP, IEEE, pp. 2872-2876, 2013.
[11] Jiantao Zhou, Xianming Liu, and Yuan Yan Tang,” Designing
an Efficient Image Encryption-Then-Compression System via
Prediction Error Clustering and Random Permutation”, IEEE
Trans. Info Forensics, vol. 9, No.1, pp. 39-50, Jan 2014.
[12] Shaohui Liu, Anand Paul, Guochao Zhang and Gwanggil Jeon,
“A game theory-based block image compression method in
encryption domain,” in springer Science, Business media New
York, April 2015.
[13] Kenta Kurihara, Sayaka Shiota and Hitoshi Kiya, “An
encryption-then-compression system for JPEG standard” in
IEEE Picture Coding symposium (PCS) , pp. 119-123, 2015.
[14] Watanab. O, Uchida. A , Fukuhara. T and Kiya. H, “An
Encryption-then-Compression system for JPEG 2000 standard”
in IEEE Int. Conf on Acoustics, Speech and Signal Processing
(ICASSP) , pp. 1226-1230, 2015
In addition to compression ratio, PSNR is also used as a
performance metric for lossy compression of encrypted images.
Methods used in [6] , [7] , [9] , [12] and [13] uses PSNR value
as performance metric. The comparison of this is shown as
below in fig.9.
Fig. 8. Comparison graph for PSNR values
PSNR values are inversly propotional to the bit error rate.
So high PSNR values provide better result. By comparing these
methods the PSNR values ranges from 30 dB to 39 dB.
Compared to [6] and [12], methods used in [7], [9], and [14]
had high PSNR values, which implies the good quality of the
image in the reconstruction phase. This is because of the
compex processing like iteartive process for reconstruction of
images[7], multilayer decomposition[9] and wavelet
transformation[14] used in the compression of encrypted
images.
IV. CONCLUSION
This paper discussed various techniques available in ETC
system to compress encrypted image by both lossless and lossy
compression of bi-level, gray scale, binary and color images.
Using Classical compression techniques it is impossible to
compress the encrypted images. Methods discussed in [1-14]
provide solution to this by considering distributed source
coding, block based compression and prediction algorithms for
compression and simple encryption method using permutation
26
A Survey on Assistive Technology using Natural
User Interface(NUI) computing to support and
augment therapeutic rehabilitation
S.J. Syed Ali Fathima
Assistant Professor, Dept. of CSE
Kumaraguru College of Technology
Coimbatore, Tamil Nadu, India
syedalifathima.sj.cse@kct.ac.in
Dr. Shankar.S
Professor, Dept. of CSE
Sri Krishna College of Engineering
and Technology
Coimbatore, Tamil Nadu, India
shankars@skcet.ac.in
Professor,
The Oxford College of
Physiotherapy
Bangalore, Karnataka, India
ahamptsp@gmail.com
The aim of therapeutic rehabilitation is to restore patient’s
ability to independently perform basic physical activities and
their task by doing certain prescribed exercises regularly and
repeatedly. Physical therapy can help improve joint function
by focusing on range of motion exercises while decreasing
pain, swelling, and stiffness.
Abstract—Therapeutic rehabilitation is a specialty in medical
field that deals with diagnosis, evaluation, treatment and
management of people with all ages with physical or
psychosomatic disabilities due to congenital disorder, accidents
or aging problem. It deals with deduction and reduction in
disabilities by providing improvement and restoration of
movement and functional ability through regular and repetitive
physical therapy exercises continued after discharge from the
hospital. However, the efficient treatment sessions are not
guaranteed due to lack of therapists and facilities, patients were
alone for over 60% of the day, patients engaged in ‘activity’ for
only 13% of the day, undergoing some traditional therapies make
patients to lose their interest and motivating patients to continue
the exercises is lagging, which in turn makes longer time for
recovery. Thus, there is a need to find ways of cost effective,
engaging and motivated training to support and improve
recovery and rehabilitation. The focus is to use technology as a
solution involving various computing techniques as a
supplementary treatment to traditional rehabilitation and
continued assessment of disabled patients. Natural User Interface
(NUI) is the emerging technique with the ability to interact with
computers or smart devices using the human body. NUI
computing is powered by human touch, gesture, voice, thoughts
and senses. This paper is a survey on assistive technology using
emerging NUI computing techniques like touch computing,
gesture computing, surface computing, brain computing and
applications of virtual reality and augmented reality to support
and augment therapeutic rehabilitation.
The disabilities due to post-stroke, post-traumatic
paralysis, myopathy, aging etc. cannot be cured within few
days’ of hospitalization or by taking oral medicines. In some
cases, in order to make them functionally enable and improve
quality of the life, it needs continuous physiotherapy treatment
even after the discharge from hospital.
The critical issues with conventional therapy is that the
efficiency of the treatment sessions are not guaranteed due to
lack of therapists and facilities, patients were alone for over
60% of the day, patients engaged in ‘activity’ for only 13% of
the day[1] , undergoing some traditional therapies make
patients to lose their interest and motivating patients to
continue the exercises is lagging. Therefore, the recovery
takes longer period. It is been proved that 2-3 hours of arm
training a day for 6 weeks improved motor recovery when
started 1-2 months after stroke[2] .Thus, there is a need to find
ways of cost effective, engaging and motivated training to
support and improve recovery and rehabilitation.
The focus is to use technology as a solution involving
various computing techniques as a supplementary treatment to
traditional rehabilitation and continued assessment of disabled
patients. The integration of highly interactive and immersive
technologies into rehabilitation has the definite possibility to
benefit both patients and therapists.
Index Terms—Rehabilitation; CLI, GUI, NUI; Gesture
Computing; Touch Computing; Brain Computing; Surface
Computing; Virtual Reality; Augmented Reality; Processing;
Open NI
I.
Mr. A Ahamed Thajudeen
Natural User Interface (NUI) is the emerging technique
with the ability to interact with computers or smart devices
using the human body. NUI computing is powered by human
touch, gesture, voice and senses. The application of interactive
and entertaining aspects of games as exercises using NUI
facilitates rehabilitation by engaging users in a training
environment where programming can be customized or
personalized tailored to personal abilities and requirements.
This helps to motivate the patient and create a positive impact
INTRODUCTION
Rehabilitation is the process of aiding an individual to
achieve the maximum level of independence in all level such
as physically, emotionally, socially, and spiritually and
maintain the quality of life in the society. Rehabilitate is the
word from Latin “habilitas” which means “to make able
again”. Assistive technology means any system or equipment
that is commonly used to support, enhance, maintain, recover
or improve functional capabilities of person with disabilities.
27
The Comparison of computing interfaces is given in
Table1.
and enjoyable experience, which in turn to leads to improved
results and faster recovery.
The forth coming sections will review the assistive
technology using emerging NUI computing techniques like
touch computing, gesture computing, surface computing, brain
computing and NUI application technologies like virtual reality
and augmented reality to support and augment therapeutic
rehabilitation.
II.
Computing
Input data
Interfaces
Support for
Therapeutic
Rehabilitation
COMPUTING INTERFACES
CLI
Text Commands
Low
GUI
Graphical commands using
Medium
pointers and pointing
A. Command Line Interface(CLI)
CLI is also known as command language interpreter,
console user interface and character user interface (CUI) in
which user issues text commands to interact with computer
programs. CLI is mostly preferred by advanced programmers
as they are keen and strong in controlling a program or
operating systems using crisp commands.. Scripting languages
like Perl, python and ruby are CLI based programming
languages which are widely used in medical and bioinformatics to access and manage medical datasets, disease
diagnosis, drug design, evaluate treatment plan and medical
discovery[3]
devices
NUI
Speech, Body Gestures,
High
brain signals, touch,
object’s skeletal structure
Table 1: Comparison of Computing Interfaces
III.
NUI BASED COMPUTING TECHNIQUES
The major NUI based computing can be classified as
follows and the review of application of these computing
techniques in therapeutic rehabilitation is discussed.
B. Graphical User Interface (GUI)
GUI is the program interface that allows users to interact
with systems through graphical icons and visual markers or
indicators taking the advantage of computers graphics
capabilities.GUI can free the user from learning
complex command languages. GUI uses basic components
like pointers, pointing device, icons, desktop, windows and
menus. The systems using GUI are also known as WIMP
(windows Icons Menus and Pointers) systems. The invention
of various hardware devices like speech synthesis, tactile
displays, sensors etc. and appropriate software applications
enabled GUI systems to be useful in medical rehabilitation for
blind and disabled people for the improvements in orientation
and mobility. GUI has now become more significant for a
rehabilitation device. The existence of such interface improves
the user friendliness and the safety of the device as it
facilitates the control over the device and the communication
between the user and the robot[4].
a.
Touch Computing
b.
Surface Computing
c.
Gesture Computing
d.
Brain Computing
A. Touch Computing
Touch computing enables the users to interact with
controls and applications more intuitively using their touch
input. For example, touching a graphical icon to open an
application. Most popularly used touch devices are mobile
devices, tablets, touch tables and walls. The basic working
principle behind touch computing is these handheld
technology devices are built with the panel carrying an
electrical charge and when a finger touches the screen; the
panel's electrical field is disrupted by the touch input. This
disruption is recorded as a computer event and sent to the
software program for processing to initiate a response to the
touch event. Optical touch technology functions in the similar
way when a finger or an object touches the surface, it causes
the light to scatter and the reflection is caught with sensors or
cameras. These sensors send the data to software which in turn
produces the response to the touch based on the type of
reflection measured.
C. Natural User Interface (NUI)
NUI is an advancement and specialized form of GUI
which allows human-computer interaction through intuitive
actions which are common human behavior. NUI allows users
to interact with technologies in natural ways using their
gestures, senses, thoughts and speech. NUI can be operated in
many ways depending upon the purpose and user
requirements. NUI implementation can be done using
intermediary devices for interaction that are either visible or
invisible. This allows computers to be used in new areas of
applications like medical rehabilitation which has been
difficult previously. The introduction of the natural interaction
modalities has increased the attractiveness and intuitiveness of
the prototyped Serious Game for Rehabilitation area of
application [5]
Touch panel ad-hoc developed games are introduced in
cognitive exercise therapy, motivating the user to actively
participate in the rehabilitation and making the whole process
reliable and now a day’s touch screens are commonly used in
daily life, re-moving the need for an initial learning stage [6].
Multi-touch tabletop system, the AIR Touch, is developed
28
which combines existing multi-touch technologies with a suite
of new rehabilitation-centric applications with the objectives
to ensure ease of use by patients, record performance measure
and leverage therapist expertise and their knowledge of a
patient[7].. MTmini Package is useful for creating our own
touchpad using Front Diffused Illumination technique.
2D camera to identify hand gestures in 3D space is presented
as recuperation aid for younger children with impaired
movement of the upper limbs on a standard Android tablet
device [11]. A self-monitoring exercise for post-stroke sufferers
using Kinect is developed by a team at Roke Manor and
Southampton University [12].
B. Surface Computing
Surface computing is the specialized GUI that interacts
with the user through the surface of an ordinary object instead
of monitor, mouse and keyboard. Touch computing deals with
flat surface, where as surface computing concentrates on nonflat three-dimensional objects surfaces like spherical,
cylindrical and parabolic surfaces. It provides a means to
Ubiquitous Computing of making every surface as interactive
in our day to day environment. It can be implemented using
displays like LCD or Projection screens, projector and Infrared
cameras. Infrared cameras are main component that operate
independent of light and allows for gesture detection in all
lighting conditions. Microsoft research is on 3D surface
computing which deals with sensing of 3D input data above
2D surfaces. The research is being performed on the
application of surface computing for the rehabilitation of
patients with brain damage following accidents or strokes. The
computer-aided seating system (CASS) with contour sensing
and evaluation devices will facilitate the study of tissue
distortion at the loaded seat interface. This system assist in
clinical seat contour design [8].
D. Brain Computing
A brain–computer interface (BCI) is also known as mindmachine interface (MMI), direct neural interface (DNI),
or brain–machine interface (BMI), that allows direct
communication pathway between an wired brain signals and
an external device. BCI is useful in researching, assisting,
augmenting, restoring human cognitive or sensory-motor
functions. BCI read neural signals and send to computer
programs to translate input signals into action. BCI using the
thought processing offers a means for paralyzed to operate a
computer, or prosthetic limb, or motorized wheelchair.
The Comparison of NUI based computing techniques is
given in Table2.
NUI based
Input Data
Computing
Support for Therapeutic
Rehabilitation
Techniques
Touch
Touch / Multi
Simple Hand gesture
Computing
Touch on Flat
Exercises
surface
C. Gesture Computing
Gesture computing enables interface with computers using
gestures of the human body typically hands, legs, eyes, facial,
speech expressions and any body movements. Speech
computing is the part of gesture computing deals with Speech
or voice recognition that allows users to interact with a system
through spoken commands. The system recognizes the spoken
words and phrases and translates them to a machine-readable
format for interaction. In gesture recognition technology, the
movements of the human body is read by the camera and
sends the data to a computer program expecting the gestures
as input to control devices or applications. It can be 2D-based
or 3D-based, working with the help of a camera-enabled
device, which is placed in front of the individual. The cameraenabled device beams an invisible infrared light on the
individual, which is reflected back to the camera and onto
gesture recognition Integrated Chip (IC) [9]. For Example,
Camera Mouse is a program that enables to control the mouse
pointer on the computer screen by moving user's head [10].
Surface
Touch/Multi
Advanced Hand gesture
Computing
Touch on Flat or
Exercises and supports
Non-flat surface
design of surfaces for
assistive devices
Gesture
Speech, Body
Whole body gesture
Computing
Movements (or)
exercises
Gestures, Facial
expressions,
skeletal structure
Brain
Thoughts as brain
Exercises relating
Computing
signals
physical and metal
thoughts (or) controlling
Kinect is a popular motion sensing
input
devices by Microsoft for Xbox 360 and Windows PCs . The
device consists of an RGB camera, depth sensor and multiarray microphone along with a proprietary software,]which
provide
full-body
3D motion
capture, facial
recognition and voice recognition capabilities. Handpose is an
another new innovation by Microsoft Research, giving
computers the ability to accurately and completely track the
precise movement of hands including the finger wiggle
through a Microsoft Kinect. An efficient method for using a
physical gestures by
thoughts (brain signals)
Table 2: Comparison of NUI based Computing Techniques
BCI system applications can be either invasive or noninvasive. The invasive requires the direct implantation of
29
needs more assistance for training. AR technology supports
rehabilitation by providing patients with
electrodes in the user’s brain and in non-invasive, the system
captures brain signals from the electrodes attached to the
patient’s scalp through an electroencephalogram (EEG)
recording. It is harmless to use noninvasive systems to avoid
risk associated with any brain injuries. Hence, non-invasive
BCI systems are used in practical neurological rehabilitation.
Thought-based control of a neuroprosthesis is been developed
in which hand grasp with implanted stimulation electrodes [13].
Post stroke rehabilitation with feedback training using virtual
hands is been developed in which BCI use the classified brain
patterns to generate opening and closing movements of the
right or left (virtual) hand[14]. A BCI system converts the user
thoughts into actions without involving voluntary muscle
movement. Thus this system provides a new means of
communication for the patients with paralysis or severe
neuromuscular disorders [14].
IV.
an entertaining and natural environment for treatments[16] . It is
used to create more realistic and interesting exercises. It is
capable of adopting a tangible object into the framework and
enable patients to touch and interact with the real and virtual
environment all together. AR Exercises can be designed for
rehabilitating finer movements such as reach, grasp,
manipulation and release of objects. Using AR therapy with
simple assistive devices seems to be a cost effective
alternative to other forms of therapy like robotic therapy, and
also can be easily integrated with conventional physiotherapy.
V.
Processing is
an open-source programming
language and integrated development environment (IDE) to
develop programs for visual design or visual arts with images,
animations and interactions[17] . OpenNI or Open Natural
Interaction is an open source soft ware development kit (SDK)
used for the development of 3D sensing middleware libraries
and applications[18] . The OpenNI framework provides a set
of open source Application Interfaces (APIs). The APIs
provide support for speech recognition, Hand gestures and
Body Motion Tracking, OpenNI APIs can be used with
processing programming language to create NUI based games
or exercise to support therapeutic rehabilitation.
TECHNOLOGIES WITH APPLICATION OF NUI
This section reviews on two major technologies with the
application of NUI that widely supports medical rehabilitation.
a.
Virtual Reality (VR)
b.
Augmented Reality (AR)
PROGRAMMING SUPPORT FOR NUI
A. Virtual Reality(VR)
Virtual reality (VR) refers to the 3D environment generated
virtually by computers that can be explored, interacted and
experienced by the humans directly using their sight and sound
senses with the help of NUI. The environment is created using
the software and presented to the user which makes the user to
feel the real environment. The person in this environment
becomes a part of it and they are able to perform actions and
manipulate objects. The researches is being performed on
developing VR based exercises to support rehabilitation and
are proven by research to deliver significant therapeutic
benefits to patients suffering from cerebral palsy, traumatic
brain injury, stroke, autism and other physiotherapy exercise
conditions. VR can be used to develop customized exercises to
meet the target of specific therapy. Interactive Virtual Reality
Exercise System (IREX®™) is a stimulating computergenerated virtual reality therapy world developed to guide
patients that can target specific body parts through clinicianprescribed interactive rehabilitation exercises, games and
activities [15] .
CONCLUSION
In this paper we have discussed different types of
interfaces, NUI based computing techniques like touch
computing, surface computing, gesture computing and brain
computing, and the technologies like VR and AR with the
application of NUI, aiding the assistive technology to support
and enhance therapeutic rehabilitation. Numerous researches
are being carried out on these computing techniques to make it
practical in medical practice. The research should also focus
on developing suitable protocols for creating and delivery of
rehabilitation services and products with these computing
techniques to meet medical device standards.
REFERENCES
[1] Bernhardt J, Dewey H, Thrift A, Donnan G. “Inactive and alone:
physicalactivity within the first 14 days of acute stroke
unitcare.” Stroke 2004 Apr;35(4):1005–9.
[2] Lawrence ES, Coshall C, Dundas R, Stewart J, Rudd AG,
Howard R, et al. “Estimates of the prevalence of acute stroke
impairments and disability in a multiethnic population”. Stroke
2001 Jun;32(6):1279–84.
[3] Carter AB. Review of methods in medical informatics:
“Fundamentals of healthcare programming in Perl, Python and
Ruby” by Jules J. Berman. J Pathol Inform 2011;2:49
[4] Laura De Rijcke, “Development of a Graphical User Interface
for a Rehabilitation Exoskeleton”,Vrije Univeriteit Brussel,
2012-2013
[5] Rego, P.A.; Moreira, P.M.; Reis, L.P., "Natural user interfaces
in serious games for rehabilitation," in Information Systems and
Technologies (CISTI), 2011 6th Iberian Conference on , vol.,
no., pp.1-4, 15-18 June 2011
B. Augmented Reality(AR)
Augmented Reality (AR) is the technology that presents a
direct or indirect view of real world environment with virtual
objects augmented by computer generated sensory inputs like
graphics, voice, video or gestures using NUI. Unlike virtual
reality, augmented reality is very closer to the real world by
blending virtual objects to real world environment using
motion-tracking technologies, such as motion sensors or
marker recognition. The VR systems has few drawbacks in
practical and home use implementation such as use of bulky
hardware, patients unable to feel or touch virtual objects and it
30
[6] Fuyuki Matsushima, Roberto Gorriz Vilar, Keita Mitani,
Yukinobu Hoshino, “Touch Screen Rehabilitation System
Prototype Based on Cognitive Exercise Therapy,” , International
Conference, HCI International 2014, Heraklion, Crete, Greece,
June 22-27, 2014. Proceedings, Part II
[7] Michelle Annett, Fraser Anderson;Darrell Goertzen, Jonathan
Halton, Quentin Ranson;Walter F. Bischof, Pierre
Boulanger,”Using a Multi-touch Tabletop for Upper Extremity
Motor Rehabilitation”, OZCHI 2009 Proceedings ISBN: 978-160558-854-4
[8] D. M. Brienza, K.-C. Chung, C. E. Brubaker, and R. J.
Kwiatkowski,”Design of a Computer-Controlled Seating
Surface for Research Applications”, IEEE transactions on
Rehabilitation Engineering, vol. 1, no. 1, march 1993
[9] http://www.technavio.com/blog/top-18-gesture-recognitiontechnology-companies
[10] http://www.cameramouse.org/
[11] E. Ziogas, J. Eyre,G. Morgan, “An Efficient Application of
Gesture Recognition from a 2D Camera for Rehabilitation of
Patients with Impaired Dexterity”, Newcastle University,
TECHNICAL REPORT SERIES,No. CS-TR-1368 January,
2013
[12] Roke Manor Research Ltd, Microsoft Kinect® Gesture
Recognition Software For Stroke Patients,
http://www.roke.co.uk/resources/articles/2012-TTP-Kinect.pdf
[13] Aymeric Guillot, Christian Collet, “The Neurophysiological
Foundations of Mental and Motor Imagery”, Oxford University
press,2010
[14] Pfurtscheller, G.; Muller-Putz, G.R.; Scherer, R.; Neuper, C.,
"Rehabilitation with Brain-Computer Interface Systems," in
Computer , vol.41, no.10, pp.58-65, Oct. 2008
doi: 10.1109/MC.2008.432
[15] http://www.gesturetekhealth.com/products-rehab.php
[16] A. Merians, D. Jack, R. Boian, M. Tremaine, G. Burdea, S.
Adamovich, M. Recce, and H. Poizner, “Virtual realityaugmented
rehabilitation for patients following stroke,” Phys. Ther., vol. 82,
no. 9, pp. 898–915, Sep. 2002. [Online]. Available:
http://www.ptjournal. org/cgi/content/abstract/82/9/898
[17] https://processing.org/
[18] http://openni.ru/
31
Experimental Investigation on Weld
Strength of Copper Wire – Al8011 Sheet
Using Ultrasonic Metal Welding
R. Karthik
Dr. P. Ashoka Varthanan
Assistant Professor
Professor & Head
UG Student
Sri Krishna College of Engg & Tech
Coimbatore, Tamil nadu, India
karthikr@skcet.ac.in
Sri Krishna College of Engg & Tech
Coimbatore, Tamil nadu, India
ashokavarthanan@skcet.ac.in
Sri Krishna College of Engg & Tech
Coimbatore, Tamil nadu, India
12ba052@skcet.ac.in
Abstract—High frequency vibrations are produced to
weld the specimen by means of ultrasonic welding
with less amount of heat energy. The sheet metal used
in this trial analysis has normal and shear force which
acts on the parts to be welded and the weld interface.
The forces are the result of ultrasonic vibrations of
the tool, pressed onto the parts to be welded. This
paper made an attempt to the produce the good
quality weld strength using ultrasonic welding with a
comparison of experimental values and RSM
predicted values. Experiments are performed using
Al-8011 plate and copper wire. Ultrasonic welding is
completed with the response surface methodology to
study the effect of input parameter such as weld
pressure, weld time and amplitude. The welded
samples are subjected to tensile test to find out the
maximum weld strength. Finite element analysis is
carried out to analyse the stress distribution and
structural analysis.
V. S. Gowthaman
automotive industry. It is a quasi-solid-state
process that produces a weld by introducing high
frequency vibration to the weldment as it is held
under moderately high clamping forces.
Fig.1 Ultrasonic metal welding
Index Terms— ultrasonic welding, Al8011(Aluminium
alloy) weld strength, response surface methodology
(RSM).
II.LITERATURE SURVEY
S. Elangovan, et al [1] described the joining of
two materials together using high frequency
vibrations in ultrasonic welding without producing
significant amount of heat. During ultrasonic
welding of sheet metal, normal and shear forces act
on the parts to be welded and the weld interface.
These forces are the result of ultrasonic vibrations
of the tool, pressed onto the parts to be welded.
They carried out a study on model for the
temperature distribution during welding and stress
distribution in the horn and welded joints are
presented. The presented finite element model is
capable of predicting the interface temperature and
stress distribution during welding and their
influences in the work piece, sonotrode and anvil.
Ganesh M, et al [2] stated that Ultrasonic
welding is devoted to weld thin sheet metals of
similar or dissimilar couples of non-ferrous alloys
I. INTRODUCTION
A. Ultrasonic Welding
Ultrasonic Welding process, in which two work
pieces are bonded as a result of a pressure exerted
to the welded parts combined with application of
high frequency acoustic vibration (ultrasonic).It
causes friction between the parts, which results in a
closer contact between the two surfaces with
simultaneous local heating of the contact area.
Interatomic bonds, formed under these conditions,
provide strong joint. Thickness of the welded parts
is limited by these power of the ultrasonic
generator. It is used mainly for bonding small work
pieces in electronics, for
manufacturing
communication devices, medical tools, watches, in
32
like copper, aluminium and magnesium without
addition of filler material resulting in high quality
weld; it can count on a low energy consumption
and on a joining mechanism based on a solid state
plastic deformation which creates a very
homogeneous metallic structure between the base
materials, free from pores & characterized by
refined grains and confined inclusions. It can join
also painted or covered sheet metals. Thin sheets of
aluminium have been joined to thin copper sheets
by means of Ultrasonic spot Welding. Results are
particularly effective in order to evaluate the
relevance of various phenomena influencing the lap
joint technique obtained on thin aluminium and
copper by the application of Ultrasonic Metal Spot
Welding (USMSW).in this we also observed that
when ultrasonic welding done on a specimen in
three different parameters, the weld time and
amplitude is playing vital role in weld strength.
When amplitude is low with high weld time and
moderate pressure, weld strength is poor. When
pressure is low high amplitude and high weld time,
the weld strength is good. When the amplitude is
high, the weld strength also is high.
S. Elangovan, et al [3] focuses on the
development of an effective methodology to
determine the optimum welding conditions that
maximize the strength of joints produced by
ultrasonic welding using response surface
methodology. RSM is utilized to create an efficient
analytical model for welding strength in terms of
welding parameters namely pressure, weld time,
and amplitude. Experiments were conducted as per
central composite design of experiments for spot
and seam welding of 0.3- and 0.4-mm-thick Al
specimens. An effective second-order response
surface model is developed utilizing experimental
measurements. Further, it is concluded that weld
strength decreases with increase of pressure
because increase in clamping force (pressure)
reduces the relative motion between surfaces
leading to reduced area of contact and hence
reduced strength. Also, weld strength increases
with increase of amplitude because increase in
amplitude gives increased area for rubbing action
between the metallic surfaces that leads to better
bonding and increase of weld strength. In some
cases, weld strength increases considerably up to
2.5 s. Beyond 2.5 s, the weld strength again starts
decreasing for any value of pressure and amplitude
because excessive weld time affects the existing
molecular bond in the weld joint.
V. Gunaraj, et al[4] describes Response
surface methodology (RSM) is a technique to
determine and represent the cause and effect
relationship between true mean responses and input
control variables influencing the responses as a two
or three dimensional hyper surface. selection of the
optimum combination of input variables for
achieving the required qualities of weld. This
problem can be solved by the development of
mathematical models through effective and
strategic planning and the execution of experiments
by RSM. This paper highlights the use of RSM by
designing a four-factor five-level central composite
rotatable design matrix with full replication for
planning, conduction, execution and development
of mathematical models. These are useful not only
for predicting the weld bead quality but also for
selecting optimum process parameters for
achieving the desired quality and process
optimization. RSM can be used effectively in
analysing the cause and the effect of process
parameters on response. The RSM is also used to
draw contour graphs for various responses to show
the interaction effects of different process
parameters.
Marius Pop-Calimanu, et al [5] stated that
the Ultrasonic welding is a solid-state welding
process in which similar or dissimilar work pieces
are jointed by the application of high frequency
vibratory energy, where the work pieces are held
together under pressure without melting. The
problem faced by researchers and industry which
have deals with ultrasonic welding process is poor
strength of the weld, due to improper selection of
welding parameters. Therefore, we are usually
interested to determine which variables process
affects the response. A stage is to optimize, this
means that we can determine the region in the
important factors that lead to the best possible
response. In this paper, welding parameters, like
welding time, welding pressure and amplitude of
the vibration are taken into account during the
realization of ultrasonic welded joints of
Al/20%SiC composite material under disks form,
whose thickness are 1 mm. This work focuses on
the development of an effective methodology to
determine the optimum conditions for welding, that
maximize the strength of joints produced by
ultrasonic welding. It was observed that welding
pressure and welding time has a significant effect
on the response (joint strength). It was observed
that the interactions of influence factors are more
significant on the net area of the joint, than on the
energy input on joint formation.
J. Pradeep Kumar, et al [6] states that
ultrasonic welding is as solid state joining process
that produces joints by the application of high
frequency vibratory energy in the work pieces held
together under pressure without melting. In
electronic and automotive applications, copper
wires are connected to the equipment (alternator/
rectifier) by a solid state joining process. For such
an application ultrasonic metal welding is useful.
This paper presents a study of ultrasonic welding of
copper wire to thin copper sheet as many of the
industrial applications are in need of this kind of
contact technology. Quality of the welded joints is
evaluated based on mechanical tests and the quality
33
criterion is then applied to evaluate the weld
strength. A second order regression model equation
is developed to predict the weld strength of the
joint based on the experiments conducted using full
factorial design of experiments. Experimental study
conducts for ultrasonic metal spot welding and
defines the quality of the joint using maximum
weld strength and failure type in the tensile test.
For this study, copper sheet and copper wire were
selected because they are the materials and joints
effectively used for many of the industrial
applications. The experiment was conducted using
full factorial design according to varying weld
pressure, amplitude and welding time. The standard
tensile test was selected to extract quality response.
Using ANOVA the percentage contribution of each
factors are determined. It can be found that the
second order regression model equation derived
from present study can be readily applied to predict
the weld strength value of the joint made between a
copper sheet and a copper wire.
potentially impact one or more Response variables.
RS model, which is an analytical function in
predicting weld strength values, is developed using
RSM. The second-order mathematical models have
been developed to predict the weld strength. The
polynomial equation for the three factors
considered in the present case is
Where
represents pressure, weld time, and amplitude; ,
,
, and
represent the constant, linear,
quadratic, and interaction terms, respectively. The
weld strength obtained from experimental results
for different combinations of parameters is given as
input to the Minitab software, and a second-order
mathematical model for predicting weld strength is
developed.
III. MATERIAL PROPERTIES
The material used here for ultra sonic welding is Al
8011, due to its properties of wrought, low weight,
corrosion resistance, and easy maintenance of final
product, The material and its alloys has a long
duration. The other material used in ultrasonic
welding is copper. The chemical composition listed
in the Table I.
V. METHODOLOGY
A. Methodology for experimental work
The material and the range of weld
parameters selected.
The weld specimens are prepared
according to ASTM standards (D100201).
Table I. chemical composition of al 8011
Element
Al
Fe
Si
Mn
Zn
Cu
Ti
Cr
Mg
is the response, i.e., weld strength;
Content(%)
97.3-98.9
0.60-1
0.50-0.90
0.20
0.10
0.10
0.080
0.050
0.050
The experiments are conducted as per
Design of experiment (DOE) using
Response surface method (RSM).
The weld strengths for each case are
observed and compared with analyzed
results.
B. Methodology for analysis work
A 3D model of the weld specimen is
created.
IV. RESPONSE SURFACE METHODOLOGY
The element type and material properties
for stress analysis are selected.
Response surface methodology (RSM) is defined as
a collection of mathematical and statistical methods
that are used to develop, improve, or optimize a
product or process. The method was introduced by
Box and Wilson. The main idea of RSM is to use a
sequence of designed experiments to obtain an
optimal response. It comprises statistical
experimental designs, regression modelling
techniques and optimization methods. Most
application of RSM involves experimental
situations where several independent variables
The
initial
conditions,
conditions are selected.
boundary
Various loads are given at the appropriate
nodes of the model and analysis is carried
out.
The simulation results are compared with
the experimental results obtained.
34
Make: Metal weld 2500, National Indosonic,
Pune
VI. EXPERIMENTAL SETUP
The experiment analysis was done with a
conventional ultra sonic welding machine
involving different weld parameters. The ultrasonic
welding machine used has got a horn made of
hardened steel (spot type) and anvil made of steel
with serrations at the top surface. The horn area
which comes in contact has serrations similar to the
top surface of the anvil to prevent the work piece
from sliding during welding. The specimen (0.3
mm Al 8011 sheet and 1.2mm diameter copper
wire) were specimens. Welding experiments are
conducted by considering three welding parameters
namely pressure, weld time, and amplitude of horn
vibrations. A total of 20 welds with 20 different
combinations of different pressure, weld time and
amplitude are carried out. Low–middle–high level
of welding parameters in welding space for
response surface methodology design is shown in
Table II. Ranges of welding parameters are selected
based on trial and error method.
Table III. Ultrasonic Welding Machine Specification
Input power
Output power
Output frequency
Maximum pressure
Stroke length
Maximum Amplitude
230V, 50Hz single phase
2500 W
20 KHz
8 bar
25mm
60 μm
B. Tensile testing
The welded samples are subjected to tensile test in
universal testing machine. 20 welded samples are
tested then the maximum load that the weld
specimen can withstand are obtained.
Table II. Experimental parameters and their levels
S. No
Factors
Designation
Level
1
Level
2
Level
3
1
Pressure
(bar)
A
2.5
3
3.5
2
Time (sec)
B
2
2.5
3
3
Amplitude
(µm)
C
28
42.5
57
Fig. 3 universal testing mahine
Table IV. Specifications of universal testing machine
Model
Maximum capacity
Least count
displacement
Grip separation
TKG – EC – 10KN
10KN
0.1mm
25 – 750mm
Fig.2 Welded specimens
A.ULTRSONIC
SPECIFICATION:
WELDING
MACHINE
Fig.4 Welded specimens after testing
35
Table V. ANOVA result on effect of process parameter on
VII. RESULTS AND DISCUSSIONS
weld strength
A. Effect of process parameters on weld strength
The process parameters such as weld pressure,
weld time and amplitude has significant effect on
weld strength was shown in Figure 5. Table V
represents the ANOVA results of weld strength. It
could be found from the figure and table that
amplitude of horn vibration has greatest effect on
weld strength by contribution of 19.13%. Also the
P-value reports that this factor is really very
significant.
It could be found from the figure 5 and table V that
weld pressure has effect on weld strength which
increases the weld strength in accordance with
pressure and a proficient gain of 13.56% of weld
strength. Also table proves that weld pressure has
significant effect on weld strength and the weld
time has not significant as like amplitude of the
horn vibration and the weld strength.
Main Effects Plot for WS
Fitted Means
P
T
DF
9
Adj SS
21.6995
Adj MS
2.41105
F Value
Model
Linear
3
9.1117
3.03725
10.91
P
1
3.7758
3.77576
13.56
0.01071
0.04
8.66
T
1
0.01071
A
1
5.3253
5.32527
19.13
Square
3
12.2692
4.08973
14.69
P*P
1
1.0825
1.08248
3.89
T*T
1
0.0731
0.07314
0.26
A*A
1
2.6797
2.67972
9.63
2-Way
3
0.3186
0.10619
0.38
P*T
1
0.2106
0.21064
0.76
P*A
1
0.0801
0.08014
0.29
0.02779
0.10
T*A
1
0.0278
Error
10
2.7840
0.27840
Lack-of-Fit
5
1.9057
0.38115
2.17
Pure Error
5
0.8782
0.17565
2.17
Total
19
24.4835
A
21.5
The response surface is plotted to study the effect
of process variables on the weld strength and is
shown in Figures 6-8. From Figure 6 the weld
strength is found to be increasing steeply in
accordance with the increase of weld pressure and
increase of weld time respectively. It is observed
from the Figure 7 that the weld strength increases
with increase in weld pressure and amplitude of
horn vibration. The Figure 8 shows that the weld
strength increases with increase of amplitude of
horn vibration and slightly increases in weld time.
21.0
M ean of WS
Source
20.5
20.0
19.5
2.5
3.0
3.5 2.0
2.5
3.0 30
40
50
Fig.5 effect of process parameters on weld strength
Surface Plot of WS vs T, P
Hold Values
A 42.5
2 1 .0
WS 20.5
20.0
3.0
1 9 .5
2.5
2.5
2
.5
P
3 .0
3 .5
T
2.0
Fig.6 Combined effect of pressure and time on weld strength
36
Surface Plot of WS vs A, P
8
2.5
3.0
28.0
19.9299
19.83427
9
3.0
2.5
57.0
21.5909
21.28843
10
3.0
2.5
28.0
20.2513
19.82895
11
3.5
3.0
28.0
21.1400
21.58791
12
3.0
3.0
42.5
20.4120
19.76736
13
3.0
2.5
42.5
19.0069
19.57155
14
3.0
2.5
42.5
18.8123
19.57155
15
3.0
2.5
42.5
19.5883
19.57155
16
3.5
2.0
57.0
22.1804
22.45726
17
3.5
2.0
28.0
21.2163
21.08007
18
2.5
2.0
57.0
22.0197
21.75302
19
3.0
2.5
42.5
19.9767
19.57155
20
3.0
2.0
42.5
19.7821
19.70192
Hold Values
T 2.5
22
WS
21
60
20
50
19
A
40
2.5
2
.5
P
3 .0
30
3 .5
Fig.7 Combined effect of pressure and amplitude on weld
strength
Surface Plot of WS vs A, T
Hold Values
P 3
2 1 .6
WS
20.8
60
20.0
50
40
2.0
2
.0
T
2 .5
A
30
3 .0
Fig.8 Combined effect of time and amplitude on weld strength
A total of 20 experiments were conducted at
different levels of parameters to obtain ultrasonic
spot welded joints of 0.3 mm Aluminium 8011
sheet and 1.2mm diameter copper wire. The values
of weld strength obtained from experiments and
those predicted from response surface model along
with design matrix are tabulated in Table VI for
spot welding of 0.3mm-thick Al 8011 and 1.2 mm
dia copper wire specimens.
Figure 9 illustrates the comparison between the
experiment sets and RSM predication for weld
strength and it’s found that the experimental and
predicted values are equal this shows the reliability
of the RSM models.
While comparing the
predicted values of strengths with experimental
values, it was observed that the deviations are
minimum except for a few combination of
parameters.
Table.VI List of RSM prediction with experimental dataset
Avg. Strength (N/mm2)
S.
Pressure
Time
Amplitude
No
(bar)
(sec)
(µm)
1
3.0
2.5
42.5
19.3945
19.57155
2
2.5
3.0
57.0
21.0587
21.37605
3
2.5
2.0
28.0
19.7692
19.97548
4
3.5
2.5
42.5
21.3769
20.81342
5
3.5
3.0
57.0
22.7544
22.72935
6
2.5
2.5
42.5
19.7458
19.58448
7
3.0
2.5
42.5
19.2007
19.57155
Experimental Predicted
Fig.9 experiment sets vs. RSM prediction for weld
strength
37
B. Finite element analysis for stress distribution
The finite element analysis was used to
determine the stress distribution for al 8011 to
copper welding. The contact between the two work
pieces is established and transient analysis is done
in order to find out the stress distribution in the
weld. Figure 10 shows the finite element model
used in the welding of the Al8011-copper
specimen. It has quadratic displacement behaviour.
The element is defined by ten nodes having three
degrees of freedom at all nodes. The element has
plasticity, creep, stress stiffening, large deflection,
and large strain capabilities. The pressure is applied
on the top surface of the top part on the nodes of
the area only where the horn comes into contact
with the work piece and the resulting displacement
of the work piece on top due to the high-frequency
vibration of the horn are fed as inputs for the
structural analysis. The work piece at the bottom is
arrested in all degrees of freedom.
Fig.11 Stress distribution on spot welded work piece
VIII.CONCLUSION AND FUTURE WORKS
The process of ultrasonic metal welding
and the mechanism of bond formation
during welding are studied. Ultrasonic
welding of copper wire to al 8011 sheet
joints are performed successfully and the
parameters affecting ultrasonic metal
welding are studied.
The material properties considered were Young’s
modulus (E), Poisson’s ratio (μ) and density (ρ) for
performing structural analyses as shown in Table
VII. The stress distribution was obtained according
to the tensile load applied and that finite element
model after stress distribution was shown in the
Figure 10.
Based on the design matrix developed
using response surface methodology
(RSM) the optimum combination of weld
parameters for copper wire to al 8011
sheet joint was identified.
The difference between weld strengths
obtained from experiments and those
predicted by RSM is minimum was
observed.
Increasing usage of light weight composite
materials in aerospace and automotive
industries are preferring ultrasonic
welding.
Fig.10 finite element model for spot welded work piece
IX. REFERENCES
Table VII. Material properties of work piece
Material
Young’s
modulus
(GPa)
Poisson’s
ratio
Density
(Kg/m3)
Al 8011
70
0.35
2710
Copper
120
0.34
9100
[1] S. Elangovan, S. Semeer, K. Prakasan, "Temperature
and stress distribution in ultrasonic metal welding—An
FEA-based study", journal of materials processing
technology 209 (2009) 1143–1150
[2] Ganesh M, Praba Rajathi R, "Experimental study on
ultrasonic welding of Aluminium sheet to copper sheet",
International Journal of Research in Engineering and
Technology, 02 (2013) 161-166
[3] S. Elangovan, K. Anand, K. Prakasan, "Parametric
optimization of ultrasonic metal welding using response
surface
methodology and
genetic
algorithm",
38
International Journal of Advanced Manufacturing and
Technology 0.1007 s00170-012-3920
[4] V. Gunaraj , N. Murugan, "Application of response
surface methodology for predicting weld bead quality in
submerged arc welding of pipes", Journal of Materials
Processing Technology 88 (1999) 266–275
[5]Marius Pop-Calimanu, Traian Fleser, "The Increasing
of weld strength by parameters optimization of
Ultrasonic welding for composite material based on
Aluminium using design of experiments", nanocon, 2012
[6]J.Pradeep Kumar, K.Prakasan, "Experimental studies
on joining Copper wire - Copper sheet using Ultrasonic
Metal welding", International Journal of Mechanical and
Production Engineering Research and Development
(IJMPERD ) ISSN 2249-6890 Vol.2, Issue 3 Sep 2012
21-29
39
Effect of Vibratory treatment on Metallurgical and
Mechanical properties of Aluminum alloys
weldments
Dr.K.Balasubramanian
Dr.V.Balusamy
Department of Mechanical Engineering
Sri Krishna College of Engineering and Technology
Coimbatore-641 008, Tamil Nadu, India
skcetbalu@gmail.com
Department of Metallurgical Engineering
PSG College of Technology, Coimbatore-641 004
Tamil Nadu, India
v_b_samy@yahoo.co.in
2.5%Mg and proved that the hot cracking resistant got
improved. They used Houldcroft test [8] to evaluate the hot
cracking susceptibility of the welds. Yoshiki et al [9] gained
refined crystal grains of magnesium by means of
electromagnetic vibrations during solidification of pure
magnesium. Equiaxed grains were formed when the
electromagnetic vibrations had frequencies less than 1000Hz.
Cui et al [10] used ultrasonic vibration to refine the 316L weld
metal micro structure and found columnar dendritic micro
structure decreased from 95% to 10%.
Benefits of vibration on grain refinement and mechanical
properties such as Hardness, Yield strength, Ultimate Tensile
Strength and Breaking Strength have been discussed by
different authors [11, 12 and 13]. Lu Qinghua et al [14]
applied vibration during submerged arc multi pass welding for
improving the quality of full welded valve and found
significant reductions in the welding deformation and residual
stress. Pučko [15] gained positive effect on impact strength by
means of vibration during welding. The reason is vibration
stabilizes the micro structure to become more resistant to heat
affects that could minimize impact toughness. It was also
found that the type of fracture turns more ductile with
vibration during welding.
The present work aims the improvement of metallurgical
and mechanical properties of aluminum alloy weldment by
applying vibratory treatment during the arc welding.
Abstract— In this work, an attempt is made to improve the
metallurgical and mechanical properties of high strength of
aluminum alloys through vibratory treatment. The materials
used for the investigation are AA6061 alloy and AA2024 alloy.
An important metallurgical difficulty is the formation of hot
cracking during arc welding. Experiments were carried out in
the presence and absence of vibratory treatment. Weldments
made with and without vibratory treatment were compared
using hot cracking test and characterisation tests like X-ray
diffraction
techniques
and
hardness
measurements.
Experimental results show that the metallurgical and mechanical
properties are improved by vibratory treatment.
Index Terms— AA6061 alloy, AA2024 alloy, Hot cracking, Xray diffraction, hardness
I. INTRODUCTION
An important metallurgical difficulty in arc welding of high
strength aluminum alloy is the formation of hot cracking. Hot
cracking is a high temperature cracking mechanism and is a
function of solidification behaviour of alloy system. The
formation of fine grained structure in the weld metal is one of
the important methods to control hot cracking. Many methods
are used to control the grain structure in the weld metals
which have been reported in literature and discussed below.
Yunjia et al [1] and Dvornak et al [2] studied the weld metal
grain refined by adding Ti and Zr as inoculants. Mousavi et al
[3] achieved grain refinement in electromagnetically stirred
AA7020 welds and proved that the grain refinement was due
to the grain detachment. Kou and Le [4] achieved grain
refinement in an oscillated arc weld of commercial 6061
aluminum alloy. Sundaresan and Janakiram [5] achieved
considerable refinement of the fusion zone grain structure in
α-β titanium alloys welding using this technique.
Balasubramanian et al [6] showed finer and more equiaxed
grain structure in GTA and GMA Welds of High strength
aluminum alloys by pulsed current.
II. EXPERIMENTAL PROCEDURE
A. Materials
The materials used for this investigation were AA2024
alloy and AA6061 alloy and their chemical compositions are
given in table 1.
TABLE I CHEMICAL COMPOSITION, WT.% OF ALUMINUM
ALLOYS INVESTIGATED
Alloy 1
AA2024
Some of the works related to vibratory treatment have been
reported and discussed follows. Using torch vibration, Davies
and Garland [7] produced grain refined weldment during
autogeneous GTA welding of aluminum alloy containing
Alloy 2
AA6061
40
Mg
0.83
Cu
4.36
Mg
1.06
Si
0.093
Mn
0.53
Fe
0.15
Al
Bal
Zn
0.14
Mn
0.096
Fe
0.41
Si
0.62
Cu
0.21
Cr
0.18
Al
Bal
B. Test specimen
To evaluate the specimens for resistance against hot
cracking, Houldcroft hot cracking test was employed. In 1955,
Houldcroft [8] first developed the so called fish bone test to
evaluate the hot cracking susceptibility of welds. The
specimen used for this work is of size 76 mm x 44 mm, and its
thickness is 4 mm. The specimen contains groves in two rows
with nine grooves in each row, cut to different depths, as
shown in figure 1.
vibratory treatment. The Vickers hardness test was also
conducted to assess mechanical properties.
III. RESULTS AND DISCUSSIONS
A. Effect of vibratory treatment on hot cracking
Houldcroft tests were carried out to find the resistance of
weldment against hot cracking. Specimen welded with
vibratory treatment was compared with the specimen welded
without vibration. Vibratory treatment was carried out at the
frequency 700 Hz for AA2024 alloy and 600 Hz in case of
AA6061 alloy.
Photographs showing cracks of different lengths in the
Houldcroft specimens are shown in figure 3. The hot crack
length and its sensitivity factor are shown in table 2.
Slot 0.8 mm
wide
44
76
All the dimensions are in mm
Fig.1. Houldcroft cracking test specimen
C. Welding
To carry out the hot cracking test, Gas Tungsten Arc
Welding was used. The main parameters used for welding are
Welding Current
Travel speed
Electrode
Shielding gas
Polarity
-
(a)
(b)
Fig.3 [a] Houldcroft test results for AA2024 alloy (a) without vibration
(b) with vibratory treatment of frequency 700 Hz
150 A
6.25 mm/s
EWTh-2, 3 mm diameter
pure argon gas
DCEN
D. Vibratory treatment
The vibration generator cum analyser used in this work
consists of Piezo electric transducer capable of producing
mechanical vibration in the frequency range of 100 Hz to 3000
Hz. The transducer is made to transmit the vibratory energy to
the weld plate through a welding fixture. The welding fixture
and the specimen holding method are such that the energy loss
during transmission from the transducer to the weld plate is
kept to the lower value. The arrangement of this vibratory
treatment unit is shown in figure 2.
(a)
(b)
Fig.3 [b] Houldcroft test results for AA6061 alloy (a) without
vibration (b) with vibratory treatment of frequency 600 Hz
TABLE II. CRACK SENSITIVITY OF ALUMINUM ALLOYS WITH AND
WITHOUT VIBRATORY TREATMENT
Crack length [mm]
Sl.
No
1
2
Aluminum
alloys
AA2024
alloy
AA6061
alloy
without
vibratory
treatment
50.5
37.5
with
vibratory
treatment
34.5
[700 Hz]
32
[600 Hz]
Crack sensitivity [%]
without
vibratory
treatment
with
vibratory
treatment
67.3
46
50
42.7
In AA2024 alloy, the crack sensitivity was around 67.3%
when the specimen was welded without vibration. The crack
sensitivity decreased to 46% when the specimen was welded
with vibration of frequency 700 Hz. In AA6061 alloy, the
crack sensitivity was around 50% when the specimen was
welded without vibration. The crack sensitivity decreased to
42.7% when the specimen was welded with vibration of
frequency 600 Hz.
B. Effect of vibratory treatment on grain size
X-Ray Diffraction technique was used to measure the
grain size in order to quantify the level of grain refinement
due to vibratory treatment. The grain size was measured for
the weld metals AA2024 alloy and AA6061 alloy. The
average grain size was calculated by using Scherrer’s
equation.
Fig.2. Experimental set up
E. Characterisation tests
The extent of grain refinement due to vibratory treatment
is determined through X-Ray Diffraction techniques. X-Ray
Diffraction technique was applied to the welded specimen,
mainly to compare the grain size of welded specimens with
41

where,
K
 cos
TABLE III. FULL WIDTH HALF INTENSITY AND DIFFRACTED
ANGLE FOR AA2024 ALLOY
τ – Grain size, in micron
λ – Wavelength of cobalt (1.79x10-4 micron)
β – Full width of half intensity (radian)
θ – Diffracted angle
Sl.No
1
2
3
4
5
6
Figures 4 and 5 show the X-ray diffraction patterns for the
AA2024 alloy and AA6061 alloy respectively which exhibit
significant line broadening. The extent of broadening is the
full width at half maximum intensity of the peak. From the
results of XRD technique, the values of full width of half
intensity and diffracted angle for the strongest peaks with
different frequency are tabulated in Table 3 and 4 for AA2024
alloy and AA6061 alloy respectively.
0
700
Peak
No
2θ (Deg)
FWHD
(B) (Deg)
1
5
2
3
10
7
44.96
77.3
52.4
44.93
77.26
52.34
0.155
0.167
0.176
0.238
0.291
0.250
TABLE IV. FULL WIDTH HALF INTENSITY AND DIFFRACTED
ANGLE FOR AA6061 ALLOY
Sl.No
1
2
3
4
5
6
(a)
Frequency
of vibration
(Hz)
Frequency
of vibration
(Hz)
0
600
Peak
No
2θ
(Deg)
FWHD
(B)
(Deg)
10
16
13
3
2
6
44.47
76.78
52.31
52.39
44.92
77.23
0.078
0.104
0.150
0.125
0.195
0.204
Calculations [AA2024 alloy]
Case (i) without vibration
d = (0.9 x 1.79 x 0.0001) / (0.155 x cos22.48 x 0.0175) =
0.063 micron
d = (0.9 x 1.79 x 0.0001) / (0.167 x cos38.65 x 0.0175) =
0.067 micron
d = (0.9 x 1.79 x 0.0001) / (0.176 x cos26.20 x 0.0175) =
0.057 micron
The average grain size is 0.062 micron.
(b)
Case (ii) with vibration (700Hz)
d = (0.9 x 1.79 x 0.0001) / (0.238 x cos22.26 x 0.0175) =
0.041 micron
d = (0.9 x 1.79 x 0.0001) / (0.291 x cos38.63 x 0.0175) =
0.038 micron
d = (0.9 x 1.79 x 0.0001) / (0.250 x cos26.17 x 0.0175) =
0.040 micron
The average grain size is 0.039 micron.
Figure 4. Diffraction patterns for AA2024 alloy (a) without vibratory
treatment (b) with vibrated at 700 Hz showing
broadening because of particle size
Calculations [AA6061 alloy]
Case (i) without vibration
d = (0.9 x 1.79 x 0.0001) / (0.078 x cos22.24 x 0.0175) =
0.126 micron
d = (0.9 x 1.79 x 0.0001) / (0.104 x cos38.39 x 0.0175) =
0.107 micron
d = (0.9 x 1.79 x 0.0001) / (0.150 x cos26.16 x 0.0175) =
0.067 micron
The average grain size is 0.1 microns.
(a)
(b)
Case (ii) with vibration (600Hz)
d = (0.9 x 1.79 x 0.0001) / (0.125 x cos26.20 x 0.0175) =
0.0804 micron
d = (0.9 x 1.79 x 0.0001) / (0.195 x cos22.46 x 0.0175) =
0.0503 micron
d = (0.9 x 1.79 x 0.0001) / (0.204 x cos38.62 x 0.0175) =
0.0549 micron
The average grain size is 0.062 micron.
Figure 5. Diffraction patterns for AA6061 alloy (a) without vibratory
treatment (b) with vibrated at 600 Hz showing
broadening because of particle size
42
REFERENCES
In AA2024 alloy, the average grain size was found to be
0.062 micron for the non vibrated specimen and 0.039 micron
for the specimen vibrated with a frequency of 700 Hz.
Similarly, In AA6061alloy, the average grain size was found
to be 0.1 micron for the non vibrated specimen and 0.062
micron for the specimen vibrated with a frequency of 600 Hz.
The grains of the vibrated specimen are found to be finer.
[1]
[2]
[3]
C. Effect of vibratory treatment on hardness
From the results of aforesaid tests of grain size
measurement by XRD technique, it is found that the vibratory
treatment has helped in the formation of fine grains. Fine grain
structure is known to improve the mechanical properties
particular the hardness. Weld metal hardness values were
measured for AA2024 alloy and AA6061 alloy with and
without vibratory treatment. Hardness was measured using a
Vickers Hardness tester at a load of 250 grams. The Vickers
hardness values of the weld metals AA2024 alloy and
AA6061 alloy without vibratory treatment and with vibratory
treatment are given in Table 5.
[4]
[5]
[6]
TABLE V. HARDNESS VALUES OF ALUMINUM ALLOYS WITH AND
WITHOUT VIBRATORY TREATMENT
[7]
Vickers Hardness numbers
Sl.No
1
2
Aluminum
alloys
AA2024
alloy
AA6061
alloy
without
vibratory
treatment
109
86
[8]
with
vibratory
treatment
[9]
131
[700 Hz]
93
[600 Hz]
[10]
The results show that the hardness has increased from 86
VHN to 93 VHN due to vibratory treatment for AA2024 alloy
and from 109 VHN to 131 VHN in case of AA6061 alloy.
[11]
IV. CONCLUSION
The effect of vibratory treatment on metallurgical and
mechanical properties was investigated and found that:
 The vibratory treatment applied during Gas Tungsten
Arc Welding of AA2024 alloy and AA6061 alloy
resulted in reduction of hot cracking in the weld
metal.
 Grains were disturbed and grain refinement of
AA2024 alloy and AA6061 alloy was achieved by
imposition of vibratory treatment during Gas
Tungsten Arc Welding.

The Vickers hardness test has confirmed that the
hardness values for the vibrated samples have
improved.
[12]
[13]
[14]
[15]
43
H. Yunjia, R.H. Frost, D.L. Olson and G.R. Edwards,
“Grain Refinement of Aluminum Weld Metal”, Welding
Journal, pp.280s, 1989.
M.J. Dvornak, R.H. Frost and D.L. Olson, “The
Weldability and Grain Refinement of Al-2.2Li-2.7Cu”,
Welding Journal, pp.327s, 1989
M.G. Mousavi, M.J.M. Hermans, I.M. Richardson and G.
Den Ouden, “Grain refinement due to grain detachment in
electromagnetically stirred AA7020 welds”, Science and
Technology of Welding and Joining, Vol. 8, No. 4, pp.
309-312, 2003.
S. Kou and Y. Le, “Nucleation Mechanisms and Grain
Refining of Weld Metal”, Welding Journal, pp.305s,
December. 1986.
S. Sundaresan and G.D. Janakiram, “Use of Magnetic arc
oscillation for grain refinement of gas tungsten arc welds in
α-β titanium alloys”, Science and Technology of Welding
and Joining, vol. 4, pp.151, 1999
V. Balasubramanian, V. Ravisankar, and G. Madhusudhan
Reddy,“Effect of pulsed current and post weld aging
treatment on tensile properties of argon arc welded high
strength aluminum alloy”,
Materials Science and
Engineering A, vol. 459, pp.19-34, 2007
G.J. Davies and J.G. Garland, International Metals
Review, Vol. 20(196), pp.83, 1975.
P.T. Houldcroft, “A simple cracking test for use with argon
arc welding”, British Welding Journal, Vol. 2, pp.471,
1955.
Yoshiki Mizutani, Takuya Tamura and Kenji Miwa,
“Microstructural refinement process of pure magnesium by
electromagnetic vibrations”, Materials Science and
Engineering A, 413-414, pp 205-210, 2005.
Y. Cui, C. Xu, and Q. Han, “Micro structure Improvement
in Weld Metal Using Ultrasonic Vibrations”, Journal of
Advanced Engineering Materials, vol. 9, pp. 161-163,
2007.
S.P. Tewari and Anand Shanker, “Effects of longitudinal
vibration on hardness of the weldments”, ISIJ international,
vol. 33, pp.1265-1269, 1993.
S.P. Tewari, “Effect of transverse oscillation on tensile
properties of mild steel weldments”, ISIJ international, vol.
39, pp. 570-574, 1999.
W. Weite, “Influence of vibration frequency on
solidification of weldments”, Scripta Materialia, vol. 42 ,
pp.661-665, 2000.
Lu Qinghua, Chen Ligong, and Ni Chunzhen, “Improving
welded valve quality by vibratory weld conditioning”,
Materials Science and Engineering A, vol. 457, pp. 246253, 2007.
B. Pučko and V. Gliha, “Charpy toughness of vibrated
microstructures”, Metallurgija, vol. 44, pp. 103-106, 2005.
INTERNATIONAL JOURNAL ON ADVANCED RESEARCH IN
SCIENCE, ENGINEERING AND TECHNOLOGY
(ISSN :
)
(A Quarterly Refereed Research Journal on Science, Engineering and Technology)
Information for Authors
1. All papers should be addressed to The Editor-in-Chief, International Journal on Advanced
Research in Science, Engineering and Technology (IJSET), Sri Krishna College of
Engineering and Technology, Kuniamuthur Post, Coimbatore : 641 008, Tamil Nadu,
India. E-mail : principal@skcet.ac.in and librarian@skcet.ac.in
2. Two copies of manuscript along with soft copy are to be sent.
3. A CD-ROM containing the text, figures and tables should separately be sent along with
the hard copies.
4. Submission of a manuscript implies that : (i) The work described has not been published
before; (ii) It is not under consideration for publication elsewhere.
5. Manuscript will be reviewed by experts in the corresponding research area, and their
recommendations will be communicated to the authors.
6. The manuscript should be in IEEE format
For publication and subscription details, please contact
The Librarian
Vankatram Learning Centre
Sri Krishna College of Engineering and Technology,
Kuniamuthur Post, Coimbatore : 641 008,
Tamil Nadu,
India.
E-Mail : librarian@skcet.ac.in
drfrlib@gmail.com
Mobile : +91-9750007196
Download