10.10. How the network can serve as a tool for transformation of an

advertisement
10.10. How the network can serve as a tool for transformation of an
input space dimension?
(Translation by Rafał Opiał, rafal.opial@op.pl)
After that short trip to blue seas and humming palms, lets now come back to next scientific problem.
An unique characteristic of Kohonen Network is that within its responses there exists a topological
representation of an input signal space. Of course it applies for neighborhood and its consequences.
In figures that you saw above, and on those you will see during your own usage of the described
program, there are blue dots representing neurons, connected to each other with red lines. These
lines, as you know, connect neurons considered as adjacent. At the beginning these lines – similarly
as dots itself – are distributed randomly. During teaching you will notice thou that a network during
learning will develop a regular net consisted of red lines. What, in fact, the net represents? Well, it is
an expression of rule consisting in that adjacent neurons will tend to signal and detect adjacent
points from an input signal space! This way the order coming out of the fact that some points
representing input signals are close together will be transposed to a network in which adjacent
points will be signaled only by adjacent neurons.
In all the figures seen so far, two notions: closeness (similarity) of input signals and adjacency
(neighborhood) of neurons in a network could be easily associated because input signal space (and of
course weights space) was two dimensional (as in Fig. 10.13 or 10.16) and during that time also a
network topology was two dimensional (as in Fig. 10.9 or 10.11). So, there existed natural
correspondence between such notions as “an input signal lying higher than the previous signal” (in
this sense it had a higher value of a second component) and “an input neuron lying higher than the
previous neuron” (in this sense that it lies, according to agreed numbering – in the previous row (Fig.
10.42).
Fig. 10.42. Example of same dimension of network topology and data space
Such situation is not the only one possible. One can easily imagine a one dimensional network which
will learn to recognize two dimensional signals (Fig. 10.43).
Fig. 10.43. Example of different dimension of network topology and data space
You may study how behaves a network that makes such a conversion of two dimensional input signal
space to one dimensional structure of a network, because my program allows you to create one
dimensional networks (chains of neurons).
Such untypical structure of a network you may obtain by giving one dimension of a network (the best
the first one) equal 1. The second dimension is then worth to define rather big, for example 100,
because then the network behaves in an interesting way. It learns though pretty fast, so there is no
need to limit yourself (in this case).
Fig. 10.44. Mapping of two-dimensional input space into one dimensional neural topology
Fig. 10.45. More complicated example of mapping of two-dimensional input space into one
dimensional neural topology
Fig. 10.46. Yet another example of mapping of two-dimensional input space into one dimensional
neural topology
Take a look at figures 10.44, 10.45 and 10.46. It looks from them that one dimensional network
pretty reasonably fits in a highlighted area of an input signal space – this guarantees two pretty
important advantages.

Firstly, a chain of neurons being a one dimensional network fills (better or worse) the whole
selected area of an input signal space. It means that for every point of two dimensional input
signal space there is a representative in a one dimensional neural network that will indicate
its occurrence. There are no points or areas in the multidimensional input space that are
„orphaned”.

Secondly, for the objects in an input signal space that lie close to each other (that is they are
in some way similar one to another) – correspond adjacent neurons in the one dimensional
chain of neurons. Unfortunately, it is likely so, but it is not always, so you should expect
errors (fig. 10.47), but still in most cases, the fact that some two states of an input signal
space are being represented by two adjacent neurons implies that they are similar states.
Points
representing
similar input
signals
Distant neurons recognizing
similar input signals
Neighboring neurons
recognizing similar
input signals
Points
representing
similar input
signals
Fig. 10.47. Example of proper and non proper representation of similarity of input signal
in neural network topology
Let's together consider what meaning may it have. Now in a great number of tasks related to
informatics we encounter a similar and a very difficult problem. For getting the knowledge related to
some phenomenon and some processes there are being gathered enormous quantities of data. E. g.
to figure out the current state of a nuclear reactor in a big power plant it is necessary to measure and
evaluate hundreds different parameters. The same refers to a blast furnace in a steelworks, a multiengine aircraft during takeoff and also to a big company which we would like to effectively manage.
An actual state of each of these big systems is described with hundreds and thousands of numerical
parameters, we may imagine it then as a point in a space of great number of dimensions. This
multidimensional space is necessary, because a result of each measurement, an effect of each
observation, a state of each signaling device – should be put on a separate axis. As certain processes
evolve, values of every parameter change and a point corresponding to a state of the considered
system changes its position. Therefore for full estimation of a plane safety, stability of a reactor,
effectiveness of a blast furnace or a company profitability – all the time it is needed to evaluate
position of a point in a multidimensional space. Specifically – the thing is about a position of such
point that represents an actual state relative to areas for which we may assign a particular practical
meaning: stable activity of a reactor – or a symptoms of overheating, good quality of produced pig
iron – or its defective composition, favorable growth of a company – or a specter of bankruptcy.
In order to control and inspect a considered process we must have a current and certain information
about its state, what is a trend of change and where it heads for. When trying to solve this kind of
problems, in contemporary engineering, generally we use solutions where all original data is
gathered and passed to a person which makes decisions. As an effect of that we can often find giant
control rooms with many indicators and blinking lights, as the ones from nuclear power plants, that
you may know from movies, or cabins of supersonic jets where every square centimeter is filled with
some devices, meters and indicators, as well as kilometers long printouts containing many numbers
and charts of economic rates on which businessmen rack their brains.
Such solutions though are fatally ineffective in practice. This mainly comes out of the fact that no
man is capable of effective inspect, control and analyze of thousands of input data. In addition an
operator of nuclear power plant, pilot of an aircraft or a chief executive of a company does not need
such detailed data. He needs a synthetic global information: everything goes right or – there is
happening something wrong. Such an information may be produced by a Kohonen Network.
Imagine that you have built a network where to its every neuron come few hundreds or even a few
thousands of signals related to every collected measurement data. Such network is not necessarily
more difficult to program than a network with two inputs, but it will require more space in
computer's memory and more time during simulation of its working. Imagine that such network
predicts two dimensional proximity of neurons, which you use so that an input signal of every neuron
you will display in some (predetermined) point on a screen, and signals of neighbors you will display
in neighboring rows and columns in order to visualize their relation. After teaching the network with
using a Kohonen method you will obtain a tool making a specific “mapping” of multidimensional,
difficult to evaluate, data, to a plane of one screen, which can be easily looked over, evaluated and
interpreted. Image on a screen will provide data that can be interpreted as following:

Every practically occurring combination of input signals will be represented by exactly one
neuron (“winner”), which will detect and signal occurrence of this exact situation. If you will
depict on a screen only an input signal of this particular neuron, you will obtain an image of a
luminous point moving on a screen. If, based on previous studies, you will remember, which
areas of screen correspond to good states of a supervised process and which to dangerous
states – then by observing where actually the luminous point lies and where is it going – you
can obtain a synthetic evaluation of situation from a perspective of the supervised process.

On a screen you can also display output signals of neurons without making a discrimination
following from usage of a principle of competition. You can make an agreement so that
different values of output signals will be displayed in different colors and then a change in a
supervised process will be displayed as a colorful mosaic. It carries much more information
about a state of a supervised process and after obtaining some practice lets deep and
accurate reasoning about observed trends, however at first sight is pretty illegible.
Ways for depicting results of Kohonen network working can be more. But yet, note on one common
feature of them: they are always two dimensional images, convenient of being looked over with a
glance and relatively easy to interpret. These images do not carry information about detailed values
of particular data (they can be easily reached out when needed), but they show synthetic, overall
image which is so very necessary to person that makes decisions and evaluates results. These images
lower a detail level of the data being analyzed, but thanks to averaging and generalization described
above, this way may be obtained very good results.
Download