Uploaded by Aditya

Image Classification Using TensorFlow GPU

advertisement
Image Classification Using TensorFlow GPU
2021 International Conference on ICT for Smart Society (ICISS) | 978-1-6654-1697-9/21/$31.00 ©2021 IEEE | DOI: 10.1109/ICISS53185.2021.9532500
Derrick Yeboah
School of Computer and Information
Engineering,
Zhejiang Gongshang University,
Hangzhou, China.
derrick.yeb@gmail.com
Mahamat Saleh Adoum Sanoussi
School of Information Engineering
Huzhou University
Huzhou, China
Https://Orcid.Org/0000-0002-31812138
Abstract— There are several image classification and a
complicated methods that are been overlooked with many articles.
This article reviews the latest practices, issues, and options for
billing classification. Emphasis is placed on synthesizing
important advanced category strategies and targeting strategies
that can be used to improve ranking accuracy. Billing sorting is a
classic problem in image processing, computer vision, and
machine learning. In this article, we study deep learning-based
image classification using the TensorFlow GPU. Because the
datasets were bridges; CIFAR-10 and MNIST FASHION for the
classification module. The results show the efficiency and
accuracy of deep learning-based image classification using the
TensorFlow GPU. Additionally, some critical issues are
mentioned that affect overall performance. However, simple
research is needed to identify and reduce uncertainties in the
image processing chain to improve classification accuracy.
Keywords— Tensorflow GPU, Image Classifications, CNN,
Feature Extraction
I.
INTRODUCTION
Image classifications is becoming the trend among
developers, mainly due to the growth of facts in various early
companies, such as e-commerce etc[1]. Normally, the system
itself might be set with loads or perhaps hundreds of input
information so that you can make the ‘training’ consultation
to be greater efficient and speedy. It starts by using giving
some form of ‘education’ with all of the enter statistics.
Machine learning is also the frequent structure that has been
carried out closer to the image category. However, there are
still parts that may be stepped forward inside machine
mastering. Therefore, picture classification is going to be keen
on deep getting to know gadgets. Machine Learning has its
context while it comes to Image Classification. This
technology can understand humans, items, locations, actions,
and writing in snapshots [2]. The aggregate of synthetic
intelligence software and machine imaginative and prescient
technology can acquire the superb result of phototype. The
fundamental project of picture type is to make certain all of
the snapshots are categorized in keeping with its unique
sectors or corporations.[3]. The various applications
consisting of vehicle navigation. The primary motivation
behind the work: 1. Is to contemplate or see how pictures or
images are been arranged dependent on deep learning and
Tensorflow or applying the idea of a deep studying set of
regulations in the image category. 2. Picture Classification
using computer vision and human vision with the goal that the
pictures can be perceived both by machine and human. 3.
Image classification in respect to Deep Neural Network, in
light of TensorFlow. 4. Precision of image classification of
every level of each picture.
George K. Agordzo
School of Computer and Information
Engineering,
Zhejiang Gongshang University,
Hangzhou, China
Https://Orcid.Org/0000-0002-68183718
II.
RELATED WORKS
A. Concept of Deep Learning
In fact, the technology was introduced into machine learning
in 1986 to introduce the company through Rina Dechter and
in 2000 to colleagues connected to Igor Aizenberg and
Booleovichon's synthetic neural network portals. This
approach did not become commonplace until 2012. All
technology enthusiasts should be aware of the development of
in-depth learning. Geoffrey Hinton has been promoting
machine learning for artificial intelligence since the early
1980s. But the 1980s were not so simple. These and other
factors have led to technological leaps, such as vision, speech
recognition, and more.
B. Drawbacks 0f Deep and Machine Learning
Deep neural networks are not easy to train due to the socalled gradient disappearance problem, which can reduce the
number of layers in the neural network. When adding levels,
the gradient loss problem can lead to too long a workout with
good precision because the improvement between training
cycles is minimal. The problem does not apply to all
multilayer neural networks, but to those that use a gradientbased learning method. This means that there are several ways
to solve this problem, either by choosing the correct boot
function or by training a system with a heavy GPU.
While engineers and professionals around the world agree
that we may have entered the golden age of artificial
intelligence, we have implemented a gadget to find out the
commands you want to overcome many more obstacles than
you think. Since commissioning, the usage must be
implemented in the application. Because of the use of learning
systems, it can be difficult to admit that this is not a first-class
way to solve problems [4]. The dominance of machine
intelligence over the last ten years has changed the world as
we know it. Today, the noise of improving knowledge and
intelligence is widespread. This is probably justifiably great
because of the ability of this discipline. The number of
computerized intelligence agencies has increased enormously
in recent years and, as the report indeed shows, the number of
agencies identified by computerized intelligence services
increased by 100 in 2015. As of December 2018,% of
companies were processing at least one artificial intelligence,
and the Deloitte report focuses on the basic pace of cloud
computing and enterprise-based computing to improve
intelligence development at 87 and 83 percent revenue,
respectively. These numbers are impressive if you are
planning a short-term career change. Artificial intelligence
seems to be a very correct assessment. It's awesome and I'm a
fan of the huge system and artificial intelligence. However,
there are situations where the use of a knowledge-gaining
device is really useless, which is not the case in some cases at
the moment, and there are some cases, even if you may have
trouble applying it [5]. We are now talking about language:
insufficient training data, unrepresentative training data, data
978-1-6654-1697-9/21/$31.00 ©2021 IEEE
Authorized licensed use limited to: Zhejiang University. Downloaded on September 24,2023 at 16:59:16 UTC from IEEE Xplore. Restrictions apply.
on incorrect, irrelevant functions, over-equipment of training
data, too good, upside down.
C. Image Classification
With a history of more than 40 years, it was written on screen,
with the name of the classification and the appearance of the
scenes, a central element of machine vision. Due to the large
number of image and video databases, it is necessary to be
able to sort and retrieve images and videos successfully. It can
be used as a source of information and categorized in / out of
context. This element is based on a description of the scan
algorithm [6] in the classification algorithm. Telesounding
searches focusing on image classification have long attracted
the attention of the telesounding community, as classification
results form the basis of many environmental and socioeconomic applications. [7] [8].
D. Use of Machine Learning
One might ask “For what reason should machines need to
learn? Why not configuration machines to proceed as wanted
in any case?" There are a lot of reasons why AI is significant.
Machine learning is a buzzword within the era worldwide
right now, and for appropriate reason: It represents a prime
breakthrough in how computers can look [9].
Very essentially, a device mastering set of rules is given
an “education set” of facts, then requested to apply that
information to reply a query. For example, you could offer a
laptop a coaching set of photos, some of which say, “this is a
cat” and a number of which say, “This isn't always a cat.”
Then you may variety show the pc a chain of recent pics and
it might begin to select out which photos have been of cats.
Machine analyzing then keeps to characteristic to its
training set. Every photo that it identifies — successfully or
incorrectly — receives delivered to the coaching set, and this
system correctly receives “smarter” and higher at finishing its
task over time. It is specifically used to construct algorithms
that can receive enter information and use statistical
evaluation to predict the output. Some of the uses of machine
learning are: Virtual Personal Assistants, Image Recognition,
Predictions, Videos Surveillance, Social Media platforms,
Spam and Malware, Customer support, Search Engine,
Applications/Companies. [10]
practically discover the similarities in education statistics. It is
often assumed that the clusters found will be reasonably
informative due to their intuitive size. For example, a group of
substantially exclusive people can turn demographics into
only a group of wealthy organizations and be very important.
However, the information set does not have to be the same for
mapping this cluster, they can be used and the cluster can be
customized to map new models to one of the cluster options.
Therefore, the access to the data is such that it is possible to
work well with enough data with the job. The second
automatic application algorithms [12]are designed to improve
the structure of data fields. The structure of the structure is
stimulated by the cost capacity, generally limited by the
construction of the ideal parameters that will ensure the
structure of the information system. Unsupervised analysis
has been very successful, including backgammon applications
from the world and Persian machines that can use vehicles! It
may be a powerful approach whilst there is a clean manner to
assign values to actions. Clustering may be useful even as
there may be enough information to form clusters (despite the
fact that this seems to be tough at times) and specifically while
extra statistics about contributors of a cluster may be used to
offer similar results because of dependencies inside the
information. Classification getting to know is powerful whilst
the classifications are recognized to be accurate (as an instance,
simultaneously as managing sicknesses, it's far generally
instantly-in advance to decide the layout after the truth with
the aid of a post-mortem), or whilst the classifications are
absolutely arbitrary subjects that we would love the computer
so that you can recognize for us [13]. Classification gaining
knowledge is often critical at the identical time as the
selections made with the useful resource of the set of
guidelines is probably required as input a few other vicinities.
Otherwise, it would not be easy for whoever needs that input
to decide out what technique. Both concepts are often worthful
and which one you choose out want to rely on the conditions-what form of trouble is being solved, how an awful lot of time
is allocated to fixing it (supervised gaining knowledge of or
clustering is frequently quicker than reinforcement analyzing
techniques), and whether or not supervised studying is even
viable.
III. METHODS AND IMPLEMENTATION
The initial step is gathering the dataset. On the off chance
that an essential master is accessible, at that point could
propose which fields (properties, highlights) are the most
enlightening. [11]
The first method is to train the agent no longer through
explicit categorization, but through a reward system to
indicate success. Note that this type of training usually fits the
picture of selection problems, because the goal is not always
to provide a type, but to create alternatives that maximize
rewards. This approach is well generalized to the current
international level, where companies are likely to be rewarded
for positive action and punished for gaining more. The drone
studio has produced many hits, including the satisfaction of
backgammon packages from international champions or even
cars that can use cars! It's a really robust technique and there's
a great way to assign values to movements. In the unattended
study you are likely to bet on, no training data is selected, the
tool tries to explore without a coach. The second approach is
called clustering, in this type of learning the goal is not always
to maximize the functionality of the software program, but to
Authorized licensed use limited to: Zhejiang University. Downloaded on September 24,2023 at 16:59:16 UTC from IEEE Xplore. Restrictions apply.
Also, measure and determine the speed and accuracy of image
classification by using the two datasets. To evaluate the two
models, we conduct experiments on two standard datasets. As
we use an unsupervised approach for image classification, we
make use of the whole corpus of each dataset by aggregating
training and test sets. In this, the python was the programming
language used in creating the algorithm which was hosted on
Jupyter notebook and Anaconda prompt (Anaconda3). We
evaluated the images by comparing with MNIST FASHION
and CIFAR-10, algorithms in the aspects of training accuracy,
loss of accuracy and validation of training, loss of validation
in MNIST FASHION and CIFAR-10.
Fig. 1. Data Flow/Processing Diagram
Fig. 3. Images from FASHION MNIST
Setting up the Environment (TensorFlow GPU)
TensorFlow is an open source learning platform with open
source code. It has a large and flexible ecosystem of
community-based tools, libraries, and resources to increase
the need to develop well-developed machines and
technologies to support and improve machines. All models
implemented in neural networks use TensorFlow and CUDA
to reduce the use of GPU recognition and expert optimization.
A document analysis tool must also be installed and
configured to use Python Widget Manager packages on a
Jupiter with laptop.
• Create a Python "virtual environment" for TensorFlow
using conda
• Install TensorFlow-GPU from Anaconda Repositories
• Simple check TensorFlow is working with your GPU
• Create Jupyter Notebook Kernel for TensorFlow
• Testing the environment
IV.
RESULTS AND DISCUSSIONS
A. CIFAR-10
In this chapter, a standardized CIFAR-10 dataset and MNIST
Fashion dataset will be employed which contains 60,000
images. The two datasets are going to be compared and
determine which one is the best to classify the images.
Fig. 4. Image of the dataset (CIFAR-10)
B. Training
During the training, we trained CNN Fashion MNIST and
CIFAR-10 models in fashion stores and each of us 20 times.
Of the 30,000 images in the MNIST database, we used 30,000
images to model the model and 5,000 images to test the
model. I also used CIFAR-10 data. CIFAR-10 is a subset of
80 million small actions and contains 60,000 32 * 32 color
photographs of one of ten object categories to test a model
with 6,000 images per image. Category.
Fig. 2. FASHION MNIST
Authorized licensed use limited to: Zhejiang University. Downloaded on September 24,2023 at 16:59:16 UTC from IEEE Xplore. Restrictions apply.
Fig. 5. Loss and accuracy evolution image
Fig. 9. Test Image, their Predicted and True Labels
C. Qualitative analysis and discussion
Fig. 6. Dataset image (MNIST FASHION)
This section talks about the above results which describes the
qualitative and quantitative analysis of all the results of the
both models. The MNIST FASHION dataset, this dataset
consumes the least time in the training process, when
classifying the images, it takes less time to process but the
accuracy is not good and the losses are also less. The test
accuracy is 87%. The second dataset is the CIFAR-10 dataset,
the dataset consumes more time in the training process when
classifying the images. The accuracy is good but the losses
are more. The test accuracy is 89.59%. By comparing the two
dataset, I will choose the CIFAR-10 because it has better test
accuracy.
D. Uses of CNN
Fig. 7. Trained image
The LeNet5 model you proposed is the oldest Convolutional
model in the neural network. According [14]. The rise of this
network architecture has made progress in image processing.
The traditional neural network convolution model consists of
two parts: function extraction and data classification. After
extracting the image features that meet the algorithm
requirements through a specific network structure design, the
classification module classifies the target image based on the
extracted features. These convolution layers can extract
recognition functions into an image using convolution
markers of different sizes. When multiple convolution layers
extract advanced image features, the classification layer can
use these features that give the image category. Convoluting
neural networks essentially reduce the dimensionality of the
image by increasing the depth of the hidden layer, allowing
the model to extract rare image functions in low-dimensional
space [15][16]The Deep Learning Algorithm selects
convoluted neural networks (CNN), which work well within
computer vision.
V. CONCLUSIONS
Fig. 8. Dataset image (MNIST FASHION)
These experimental data show that the neural network
standard is fantastic and is the perfect method for image
classification. The two networks have in common the fact that
they all evaluate the identification of the orientation, size and
other characteristics of their network structure and, therefore,
the establishment of internal connections. Among them is
CNN, a convolutional neural network with the characteristics
of various parameters, rapid formation, high scores, easy
migration, a great improvement over the conventional neural
network method. This convolutional layer is the key to
Authorized licensed use limited to: Zhejiang University. Downloaded on September 24,2023 at 16:59:16 UTC from IEEE Xplore. Restrictions apply.
achieving this growth, which can be described as the
lifeblood of the entire convolutional neural network. As
previously mentioned, CNN has lost a lot of knowledge about
the team. Therefore, it is not possible to recognize the
orientation of the objects to improve the extraction accuracy
of the image function. Most of these images were essentially
the same, although CNN was unable to identify them, which
is often the dominant output factor. Thanks to a full merge
network without a full link layer (FC), CNN can handle a
variety of input formats. The unwind layer allows you to scale
your data, contributing to more accurate results. The structure
of the jumps in the results of different deep layers is
combined to ensure the durability and precision of the results.
Due to a combination of the two factors mentioned above,
CNN's membership level. Meanwhile, when we selected
images from the test set, and thus the training images were
roughly the same, even higher recognition accuracy was
achieved for CNN in CIFAR-10. The database we use was
carefully designed to form or not perform a recognition task
purely, suggesting that an object is often recognized from
different perspectives. With such an attitude, CIFAR-10
defeated the revolutionary MNIST in a database. In additions,
it is often recommended to carry out various comparative
studies of classifications. Strategies or procedures related to
the implementation and capability of key classification
algorithms need to be developed for future research. Given
the future work in the field, it needs to be researched,
evaluated and further development: data collection and
preparation, integration with other applications.
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
REFERENCES
[1]
[2]
[3]
[4]
[5]
Z. H., “No Title,” [Online]. Available:
https://medium.com/@Petuum/intro-to-.
F. Faux and F. Luthon, “Theory of evidence for face
detection and Tracking,” Int. J. Approx. Reason.
Elsevier, vol. 53, pp. 728–746, 2012.
Toshisada Mariyama, Kunihiko ukushima and M.
Wataru, “Automatic Design of Neural Network
Structures Using AiS,” . ICONIP, vol. (2), pp. 280–
287, 2016.
Y. Bengio and Y. LeCun, “Scaling Learning
Algorithms towards AI,” in In Large-Scale Kernel
Machines, 2019.
M. D. Zeiler et 'al , “Adaptive de-convolutional
networks for mid and high-level feature learning,",”
[14]
[15]
[16]
Comput. Vis. 2011 IEEE Int. Conf., pp. 2018–2025,
2011.
Roberts and Jacob, “Thinking Machines: The Search
for Artificial Intelligence,” vol. 2, no. 2, pp. 14–23,
2016.
G. Foody, “Impacts of Sample Design for Validation
Data on the Accuracy of Feed forward Neural
Network Classification. Appl,” vol. 7, no. 888, 2017.
Friedl, M. A., and C. E. Brodley, “Decision Tree
Classification of Land Cover Form Remotely Sensed
Data,” Remote Sens. Environ., pp. 399–409, 1997,
doi: 10.1016/S0034-4257(97)00049-7.
S. J., M. R. G. Marques, S. Botti, and E. Al., “Recent
advances and applications of machine learning in
solid-state materials science,” Comput. Mater, vol. 5,
no. 83, 2019, doi: 10.1038/s41524-019-0221-0.
A. M. and K. R, “Machine Learning,” Effic. Learn.
Mach., vol. 4302, no. 5990, pp. 9–1, 2015, doi:
/10.1007/978-1.
M. Xie, N. Jean, M. Burke, D. Lobell, and S. Ermon,
“Transfer Learning from Deep Features for Remote
Sensing and Poverty Mapping,” Comput. Sci., pp.
3929–3935, 2015.
Z. Ghahramani, K. A. Heller, and S. Williamson,
“Statistical models for partial membership,” 2008.
T. Xiao, Y. Xu, and K. Yang, “The application of
two-level attention models in deep convolutional
neural
network
for
fine-grained
image
classification,” in Proceedings of the IEEE
Conference on Computer Vision and Pattern
Recognition,
2015,
pp.
842–850,
doi:
10.1109/CVPR.2015.7298685.
Y. Lv, C. Cheng, S. Sugianto, C. N. Muponda, and
D. E. K. Agordzo George Kofi, “Driver’s distracted
behavior recognition research based on image
Segmentation,” J. Zhejiang Univ. Sci. Technol., vol.
32, no. 03, 2020, doi: 10.3969/j.issn.16718798.2020.03.008.
Prabhu., “Understanding of Convolutional Neural
Network (CNN) Deep Learning,” Towar. data Sci.,
2018,
[Online].
Available:
https://medium.com/@RaghavPrabhu/understandig.
H. Nam and B. Han, “Learning multi-domain
Convolutional neural networks for visual tracking,”
in Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 4293–4302.
Authorized licensed use limited to: Zhejiang University. Downloaded on September 24,2023 at 16:59:16 UTC from IEEE Xplore. Restrictions apply.
Download