„Futures Generation for Future Generations”

advertisement
Creating objective evaluations, or the possibilities of similarity
analysis
László PITLIK, István Pető, László BUNKÓCZI
pitlik@miau.gau.hu, ipeto@miau.gau.hu, blaszlo@miau.gau.hu
SZIE, Institute for Business Analysis and Methodology
„Futures Generation for Future Generations”
World Futures Studies Federation (WFSF) 19. World Conference
Budapest, 2005. VIII. 21-24., Corvinus University
Introduction
The aim of the conference: to give a chance for opinion exchange between generations, to
have constructive common thinking. The sections of the conference include a wide range of
research areas from demography, across standard and deviant self behaving forms, till the
future forming role of informatics, till the development of communities, till the gender issues
and till comparing cultures.
The facts and the drawable conclusions with given authenticity can be explored from all the
problem groups separately. Nevertheless, when the informational society is defined as the
effort towards the maximal partial efficiency (automation), than the question may be raised,
what is the common in the topics of the sections of the conference? Is it possible and do we
have to handle in a unified way (systematically and automated) the similar problems?
The Department of the Business Informatics introduced in national and in international levels
in 2004 the COCO method (Component-based Object Comparison for Objectivity – up to
now with 63 Hungarian and foreign language theoretical essays and case studies:
http://miau.gau.hu/sitemap/kapcsolat.php3?where[kapcsolat]=coco&where[focsop]=kut&mod=l2003).
The essence of the method: on the base matrixes of arbitrary space and time connected (time
series based, generational based) objects (countries, cultures, people, enterprises, products,
services, etc.) and arbitrary measurable and observable attributes (object-attribute), what kind
of expert attitude free (weighting, pointing), valid for present and future awarding (below and
higher estimation, object specified danger and choice positions, expectable directions of state
change, not valuable factors – like brand, culture, religion) can be deducted for the objects.
In the case of those questions, where the objective approach is shadowed by the missing
statistical data assets, the task of the research on the base of the discrete experiences
concluded from the point based data is to point what kind of data are required in international
level, and what are those automatically dissections (cf. SWOT) that are able to show,
monitoring and avoid the loosing balance situations (cf. sustainability, decision support).
The exigent science-fiction literature, as a kind of future (picture) factory shows clearly, that
the balance of social processes are assured in that way, when the human subjectivity in
operative level is crowded out from the definition of sustainability, from monitoring the
loosing balances, from making the necessary decisions. The democracy in the informational
society is not else than accepting on the base of majority, introducing, holding (operative) the
scientifically based strategically relevant automatisms (indicators, simulations, decision
support) which are formed after the relative unlimited access to data, to analysing methods
and to education.
In other approach the similarity analysis aiming for objectivity may give the common all in
one mathematics (polito-metrics) of the multi-cultural, of the positive discrimination and
racism. A good example for it is a project plan (http://miau.gau.hu/miau/74/pitlik.doc)
examining the EU countries in getting more and more European. The essence of politometrics in this case: whether a group can be formed to show what can be named as European
pattern on the base of arbitrary social indexes, than the distance of any other country from it
can be calculated objectively. Along this calculation it may turn out that the examined country
is much more European or if not then the factors can be defined why it is far from the others
(cf. SWOT). In a certain case the religion’s effect/connection can be numbered against a
certain value (e.g.: GDP, number of suicides, migration, unemployment). In that case when
the religion shows a significant negative connection with important attributes than the
mathematical equation is ready for describing racism, or against with it the equation for
positive discrimination.
The role of similarity analysis in futurology
Developing, changing of systems without goal may be arbitrary scaled and formed. And like
this – can’t be analysed in essence (cf. with occasion). Whether the goal of the human
societies is the happiness, and the happiness is defined as the recognised objective necessity
(cf. with Montesquieu), than for future generating the necessary strategic frames are already
defined. Though the „happiness” can’t be measured, but the recognised necessity may be
measured with enough simply The recognised necessity is not else than having correct
forecasts (for arbitrary timeline and content, as accurate as possible). Otherwise the
calculable, or with fine word sustainability (cf. with a legitim choose between projected
scenarios), and improving the automation of routine tasks.
In this society picture which seems to be so automatic, man doesn’t loose it`s creative,
intuitive and associative character. The base for achieving these special human capabilities is
that if the man is really Human, or it can`t be automated, until it moves on a knowledge level
which hasn`t been explored yet (cf. successful research engineer vs. stahanovist line worker).
The informational society, the knowledge management etc. modern phrases certifies the
recognition of this new direction.
Creating futures is not an ideological, not an unrealistic self objective theory, than a mix of
the basics of system theory, of the self justifying empirical approach and automation (cf.
cybernetics) which includes the principles of adaptation, as it should be, like it has been
defined by itself for the experimental science for centuries. Thus the futurology is an
experimental science, as the panning of future forecasts can be examined well, and can be
expected that with interpreting the logical errors behind misses the objective statutories will
be recognized with higher and higher safety.
The role of the similarity in this process is complex. The potential futures where the
specification of the potential changes of the arbitrary attributes of the arbitrary objects is the
goal, may be diverse. After the experiences the changes affecting commonly numerous
attributes and objects can’t be arbitrary eccentric and extreme. The common change of the
elements of one system has to lead to a consistent future. It’s not allowed to think in
forecasting partial phenomena, as in a concrete question to choose the best in advance (the
closer to the real in the future) is almost impossible (cf. with the probability of a discrete
estimation). Though it’s not self understandable which complex error structure has more
advantage in case of common movements of more factors, but that will have a serious
advantage who tries to see the future in it’s whole complexity. It’s important or outline that
similar consistent futures may exist in the same moment of an analysis. This recognition
shows the movement possibilities of the individuals (and of the chaos too). In case of an
individual the sovereignity is supposed, the statistical average reaction may affect a system
itself in case of a large human group (cf. with Asimov: Sheldon theory).
In consistency an important part aspect is, that for all object/attribute, which are in heavily
extreme position against the other objects, are interfered by a stronger and stronger force for
back alignment. The future changes can be described after the passing of time on one hand,
than the differencies in space, and thirdly on the base of field of forces described by diverging
similarities in that way these field of forces are not independent from each other, that’s why
they are capable for checking each other.
In terms of the principles of the conference: Our goal for the conference is to generate futures
“far enough away to create, but close enough to live in.”
Instead of the unique analysis in the abstract (having a few dozen case studies showing the
possibilities in COCO methodology) shall stand here a not complete but enough detailed list
about what type of analysis can be performed (automatically after collecting the base data) for
relevant but enough far time where/when the expectations are drafted.
-
Meteorology, geology, hydrology (etc.): measuring the environment continously and
more and more frequently the common changes can be foreseen with higher accuracy
for more and more far.
-
Stock market, price analyses: for calculating the group behaviour samples analysing
the stock market changes is a good example.
-
Headhunter corporations, suitability examinations for certain jobs, IQ and other
psychological tests, grafology!, astrology?, palmistry?: the expected utility and future
can be explored of one individual after the searched and the given profiles.
-
Project view, monitoring principles, chance equilibrium, precedent theory: The
efficient work organization and law application shows a good example for the
similarity and fact based process control.
-
Public use data, constitutional rights: The key of sustainable development is, if
anybody may know all facts, anyone may compare any object with each other if they
seems to be similar, and after this analysis may feel himself as retrograde, as winner
and as in equilibrium state (cf. with gender problem, state of minorities). …
The COCO methodology
http://miau.gau.hu/sitemap/kapcsolat.php3?where[kapcsolat]=coco&where[focsop]=kut&mod
=l2003
Consistency
Connection between generations and COCO similarity analysis
Firstly it’s worth to clarify how the „generation” phase can be interpreted:
-
in classical term: age groups following each other,
-
in globalisation approach: similar objects in different development phase,
-
universal: one after the other.
In a similar way the process COCO-based similarity analysis has to be given briefly:
-
comparing objects with the same type of indexes,
-
to rate the indexes without (pointing, weighting) experts,
-
to explain objectively the observed differences in case of given goal phenomena,
-
which analysis’s objective may be forecasting, distance identification, grouping.
What kind of connection may stay between the two concept/idea:
-
dependent/minor connection, when one of the idea can be interpreted as the part of the
other
-
adjunct connection, which case has to be unfolded to different formations:
o the two phenomena has a common set,
o the two phenomena doesn`t have a common set, only loose associative
connection can be discovered.
Whether the similarity analysis, is seen as the process of getting known of the real life, then
from those three definitions given above, into the common sets between the two concept, the
followings may get as the main parts of the main concepts:
-
on arbitrary one after the other objects (as the base elements of the case based
reasoning)
-
executing analysing processes.
The existence of generations (in space and in time) may serve as a base pattern (base data, as
benchmark) for the dissection. For the self evaluation of dissections, or in case of specifying
the condition set for quitting a learning process the generational style gives the basics for the
definition of the learning and test data, for the experienced older expert’s heuristics in error
search strategy for consistency principles.
In other approach the generation is not only generation than is in connection with the
generating phase. The connection of this word which means to generate an action and the
similarity analysis is much more dependent/minor, than before, as generating (creating one or
more element/object within possibility circle) is a concept giving a frame for automated
dissections. Generating futures supposes the generation of mechanisms calculating themselves
which process can be divided for the definition of data (inside in it for generating fictive data),
for specifying the operations, where on the base of them the models (source codes) can be
generated, or defining the evaluation mechanisms of the generated models (consistent exiting
condition from learning), which checking algorithm can be seen as the part of the model, and
this supposes source code generation.
Summary
Though the modelling’s (future generating) base question („between two models which is the
better?”) means a never solvable philosophical borderline, with building up more and more
sophisticated model (self-) checking frame systems, either with the consistent integration of
knowledge from the patterns of the known supposed past, more and more automatic (without
ideological basics) deduct able the potential (alternative) complex common changes. The
deduction’s methodology will be able to fine itself after the fact-estimation comparisons and
along the complex analysis of the ways toward the comparisons. The philosophy of the
recognized necessity in this way may be able to develop the futurology to an industry like
service. As the future is never knownable till the whole deepness, thus the social and
individual handling of the always existing estimation error and sustaining model factories
may move out the people living in the informational society from the recent development
orbit which is based on monopolisation/mystery (cf. data protection), hoax (cf. marketing),
corruption (cf. with the Hankiss common meadow model), in that case when it keeps
genetically the tension in liveable range between the instinct based (e.g. mimicry) and
conscious acting (social optimum)…
References
http://miau.gau.hu/sitemap/kapcsolat.php3?where[kapcsolat]=coco&where[focsop]=kut&mod
=l2003
Download