Co-evolution Learning in Organizational-Learning Classifier System

From: AAAI Technical Report SS-02-02. Compilation copyright © 2002, AAAI (www.aaai.org). All rights reserved.
Co-evolution Learningin Organizational-LearningClassifier System
Takao Terano, Yasushi Ishikawa
Graduate School of Systems Management,University of Tsukuba, Tokyo
3-29-1, Otsuka, Bunkyo-ku,Tokyo112-0012,Japan
terano@gssm.otsuka.tsukuba.ac.jp
Abstract
This paper proposes an agent-based system with
Organizational-Learning
OrientedClassifier System
(OCS),whichis an extension of LearningClassifier System~CS)into multiple agent environments.
In OCS,each agent is equippedwith a corresponding Michigantype LCSand acquires problemsolving knowledge
basedon the conceptsof organizational
learningin management
and organizationalsciences.
In the proposedsystem, we further extend OCSto
employthe followingcharacteristics: (1) eachagent
solvesmulti-objective
problems,there are sometradeoffs about given problems;(2) the agents compose
multi-classes,andin eachclass, theypursuedifferent
goals, whichmightcauseconflictsamong
the classesof
agents,and(3) the agentslearn bothindividuallyand
organizationally.Wehaveappliedthe systemto
logical Marketing"
domain
in orderto explaincompetitive andcooperativeagentbehaviorsin developing
and
purchasingbotheconomical
and ecologicalproducts.
Introduction
Recentdistributed information systems over the Internet
often reveal very complexphenomenain practical operation, althoughthey are constructed from very simple software components,because the behaviors of the users are
not predictable. To understand the phenomena,we must
develop explainable and executable modelsto analyze the
activities of organizations, whichconsist of both humanand
artifacts. Computational(and Mathematical)Organization
Theory (COT;CMOT)
[Carley 1999] utilizes agent-based
modelingtechniques [Axelrod1997], [Epstein 1996], [Axtel 2000]. By agent modeling, we mean that we develop
models with groups of simple and small software components or software agents and that they behave in a given
environmentand solve someglobal problems.
Theapproacheswe address in this paper also employthe
principles of agent-basedmodelingtechniques. To develop
the model, we introduce the concepts of Organizational
Learning in management
and organizational sciences [Argyris 78]. This paper proposesan agent-basedsocial simulation systemwith Organizational-LearningOriented Classifter System(OCS)[Takadama99a, 99b, 01a, 01b], which
Copyright(~) 2002,American
Associationfor Artificial Intelligence(www.aaai.org).
All rights reserved.
25
is an extension of LearningClassifier Systemin GeneticsBased MachineLearning [Goldberg89], [Lanzi et al., 01]
into multiple agent environments.The proposedsystem is
characterizedby the agents that (I) individuallysolve multiobjective problems,(2) pursue different conflict goals,
they belongto different classes, and (3) learn both individually and organizationally. Compared
with multiagentlearning systems so far (e.g., [AAAI96]), the proposedsystem
is so complexthat the agents are able to learn fromrandom
initial knowledgeand so powerfulthat they are utilized to
simulatepractical social activities. Toshowthe effectiveness, therefore, we have applied the systemto Ecological
Marketing~lomainin order to explain both competitiveand
cooperative agent behaviors in developingand purchasing
both economicaland ecological products.
This paper is organizedas follows: in section 2, we describe the background
and objectives of the research, in section 3, weproposea newOCSbasedarchitecture, then using
the proposedmodel,in section 4, we carry out experiments
on Ecological Marketingsimulation. Basedon the experiments,section 5 and 6 respectivelydiscuss the issues of organizationallearning and co-evolutionlearning. Finally, in
section 7, concludingremarkswill follow.
Background and Objectives
The literature in COTand/or CMOT
frequentiy reports that
small autonomousagents can generate global interesting
structures and behaviors.In moredetail, they describe:
(1) someinteresting organizational phenomenausually
occur in observingthe systems’behaviors, (2) very subtle
changes of control mechanismsor system parameters will
dramaticaily changethe characteristics of the systems, and
(3) there are common
face-similarities of behaviorsof complex systems, for example,computersystems, social systems, and economicalsystems.
However,these researches in the literature only discuss
the emergentproperties whichcan be identified fromthe observers’ standpoints, and tend to ignore the design problems
of such multiagent systems. Theseresearches have not yet
attained to describethe flexibility andpracticability of practical informationsystems.Bythe wordflexibility and practicability, we respectively meanthat ambiguousknowledge
sharing and ongoingadaptation in the context of activities
occur.
Tosolve the difficulties, the study onextendingthe architecture of learning classifier systemsinto multiagentenvironmentshas becomepopular in recent years. In such environments,agents should cooperatively and/or competitively
learn each other and solve a given problem.As stated above,
amongstthem, we have conductedthe research based on the
concepts of Organizational learning in management
and organizationalsciences.
Thoughthere are various definitions and discussions
about organizationlearning in the literature [Argyris78],
[Duncan79], [Kim93], by organizational learning, wemean
that it is the processfor the organizationto learn to solve
a given problem,which cannot be solved by each individual agent in it, becauseof the insufficient capability of the
agent. The approachesto learning multiagent systems are
classified in the followingtwo categories. Theone is that
all agents in the systempursue the samegoal, and the other
is that multipleclassesof agents pursuedifferent conflicting
goals and affect each other in the problemsolving processes.
In both approaches,agents learn individually, however,from
our organizational learning viewpoints, the formeraims at
acquiringthe ability to get better solutions of the problem
by exchangingor transferring adaptation behaviors among
the agents. Onthe other hand, the latter almsat acquiring
the ability to howto act fromthe feedbackfromthe environmentalchangesdueto the activities of the classes of agents.
Fromthe technical viewpoints, there has been a number
of studies in the co-evolutionof populationsin GAandmultiagent learning literature. For example,[Rosin 97] proposes
a methodof "competitive coevolution," in whichfitness
is based on direct competitionamongindividuals selected
from two independentlyevolvingpopulations of "hosts" and
"parasites." [Bull 96] proposesa multi plied learning approach because the agents do not influence each other in
their learning processes. [Grefenstette 96] describeaexperiments with co evolutionary approaches,that are similar to
ecological environments
wherespecies adapt in evolvein interaction with each other. [Haynes97] analyzes cross over
operators and fitness functions that allow to rapidly evolve
a team of agents with goodtask performance.[Sikora 94]
integrates a similarity-based learning programand GAoperations. Compared
with such researches, in this paper, we
emphasizethe concepts of organizational learning and the
application of learning classifier systemsin Genetics-Based
MachineLearning[Goldberg89] into multiagent learning.
Wehave developedthe architecture of Organizationallearning oriented Classifier System(OCS).In OCS,we have
focuseduponthe formertype organizational learning by the
rule or classifier exchangemechanism
[Takadama99a, 99b,
01a, 01b]. However,the techniques we have implementedin
OCSso far are only applicable to the agents with the same
internal modelsand the samegoals. Therefore,in this paper,
weaddress the development
of the latter type organizational
learning in the OCSarchitecture in whichmultiple classes
of agents mayhavedifferent internal models.
Whenwe develop a learning multiagent system, we often
meetthe followingfour difficulties:
¯ Agents maysolves Multi-objective problems:
26
Whenacting as individuals, the agents mayindividually
achieve morethan one objective at the sametime. This
meansthat agents must solve multi-objective problems.
For example, in the Ecological Marketing domaindescribed later, the agents meetthe tradeoffs betweeneconomicaland ecological goods.
¯ Agentsmaypursue different conflicting goals:
Whileagents act for their objectives, the agents mayhave
different weightson their multiple objectives. Furthermore, in the environmentwhereagents act, there maybe
multiple classes of agents. A group of agents with the
same objectives forms an agent class. The agents in a
sameclass maycooperativelyimitate, learn, and adapt to
the activities of the other agents in the sameclass. However, the agents whobelongto different classes maypursue different conflicting goals. In the EcologicalMarketing domain,one agent groupis correspondingto the supplier class and the other is correspondingto the consumer
class. Bothclass agents wouldlike to contribute the environmental problems, however,from the economicalview
point, the objectivesare conflicting.
¯ Agentsmaylive in multiple environments:
If there are multiple environments,each environmentrequires somespecific activities to the agents living in it.
Agentmust adopt the requirements, as they are ones of
the agents objectives.
Ourpurposeof this research is that weextend the architecture of the original OCSin order to developmoregeneral
social simulationmodelsfor analyzingdistributed information systems with multiple classes composed
of such agents
withmultiple objectives. In the followingsections, we will
describe the conceptsand validate the effectiveness.
Organlzational-learnlng oriented Classifier
Systemand its Extension
In this section, weproposea newarchitecture with extension
of the OCSto simulatethe agent withthe characteristics explainedsofar.
Brief Descriptionof Organizational-learning
Oriented Classifier System
Figure 1 showsthe OCSarchitecture extended from conventional Learning Classifier System(LCS)[Goldberg89]
with organizational learning methods. The system solves
a given problemwith multi-agents organizational learning
wherethe problemcannot be solved by the sumof individual
learning of each agent. In OCS,we introducethe four organizational learning mechanismproposed in [Kim93]: (a)
Individual single loop learning, CO)Individual doubleloop
learning, (c) organizationalsingle looplearning, and (d)
ganizational double loop learning. Eachlearning mechanism respectively correspondsto (a) reinforcementlearning
(or modificationof weights)of classifiers in each agent, Co)
newclassifier or rule generation via genetic operations in
each agent, (c) the exchangemechanism
of goodclassifiers
amongvarious agents, and (d) collective knowledgereuse
for new problems.
Agentsdivide given problemsby acquiring their ownappropriate functions throughinteraction amongagents Weassumethat a given problemcannot be solved at an individual
level.
The
agent
inOCS
contains
thefollowing components:
¯ Problem
Solver:
- Detector and Effecter: they translate someparts (sub
environments)of a total environmentstate into an internal state (workingmemory)of an agent, applying
classifiers and/or collective knowledge,and derive actions basedon the informationin the workingmemory.
¯ Memory:
- Collective knowledgememory:
It stores a set comprising each agent’s rule set as collective knowledge.In
OCS,all agents share this knowledge.
- Individual knowledge
memory:
It stores a set ofclassitiers (CFs). In OCS,agents individuallystore different
CFsthat are composed
of if-then rules with a strength
weightfactor. At first, a givennumber
of rules are generated in each agent at random,and the initial weights
are set to the samevalue.
- Workingmemory:It stores the recognition results of
both sub-environmental
states and the internal states of
an actionof fired rules.
¯ Learning in OCS:
- Reinforcement Learning: In OCS,the RLmechanism
enablesagents to acquire their ownactions that are required to achieve their owngoals by tuning the rule
weights. Weemploya profit sharing methodfor the
purpose.
- Rule Generation: The mechanismgenerates a newrule
whenno rules in the individual knowledgecan be applied to the workingmemory
conditions and a rule with
the least weightis deleted.
- Rule Exchange:In OCS,agents exchange rules with
other agentsat a particular timeinterval in orderto distribute moreeffective rules that cannotbe acquiredat
the individual level learning. Whenexchanged,rules
with lowerweightsare replaced.
- Collective KnowledgeReuse: Whenthe agents have
solved someset of problems, the individual knowledge is chunkedand commonly
distributed to the all
the agents. The knowledgeis stored in order to apply
the future problemsolving.
The basic idea of the proposedsystemin this paper comes
from OCS,however,the proposed system is extended from
the original OCSso that we deal with the agents with the
characteristics described above. Wewill omit the detailed
descriptions about the system common
with the original
OCSThe detailed discussion about the mechanismsof the
original OCSis found in [Takadama01a, 01b].
Coping with Multi-Objective Problems
First, wehavechangedthe formof the classifiers, whichare
stored within Individual Knowledge~module, in order
27
I[
" Im Bnttwnlam
!
iT
T.
__t’,~L~_
..r~..
g’ ....
t~111¢4N
w
~m
J~mt~
~N
~s
Figure 1: Architecture of Organizational-learningOriented
Classifier System
to copewith multi-objective problemsdealt with the agent.
Referring to VEGA
(Sehaffer 1985), which implements
wayto deal with multi-objective problemsby Genetic Algorithms (GAs), we have applied the principle of VEGA
to LearningClassifier System.Eachclassifier individually
holds multiple strengths correspondingto each objective.
Eachof the strength values meansthe adaptation score towardsthe correspondingobjectives. Thescore is calculated
by a methoddescribed later. Individual knowledgewill be
iteratively improvedby reinforcementlearning. If two or
moreof the agents havethe sameobjectives for their action,
then they are formedin the sameclass. Anagent decidesits
action by the followingprocedures:
(1) Select probabilistieallyone of the n objectives abouttheir
actions.
(2) Select someclassifiers that matchthe states of Detector"
of Learningclassifier system.
(3) Select one classifier fromthemaccordingto strengths by
"RouletteSelection"to decideits action.
(4) Applythe classifier to Effecter ito a real action.
Eachagent individually has n weight values according
to the correspondingn objectives. In each action, one of
the objectives is probabilistically selected accordingto the
weightvalues, and then correspondingactions will be taken.
Thus, by the difference of the weightfor each objective, we
represent the agent’s characteristics so that Agentpursues
different goals |
Classifiers are gradually trained via conventional LCS
mechanisms.The learning procedureis as follows:
(1) Select the elite classifiers. Select somefixed numberof
excellentclassifiers for eachn objective. Somefixed ratio
of the selected superiorclassifiers remainsfor eachobjectives accordingto their weightsfor each objective.
(2) Select classifiers for crossoveroperation.First, select two
classifiers fromall classifiers randomly.Next,select one
from the two by the tournamentselection. At the tournamentselection, select one objective, comparethe correspondingstrengths of the twoclassifiers, and select the
better one for the candidatefor the crossoveroperations.
(3) Applygenetic operations. Accordingto the general procedures of GAs,apply crossover, selection, and mutation
operatorsfor the next generation.
Extension for Multi-Class Agents
Wehave changed OCSto handle multi-class agents. As
shownin Figure2, multiple classes of agents share the same
environment.However,each agent class has each different
viewpointto interpret the status of the environment,
different
objectivesto act, or different actions. Toattain this, wedesign Effecter iand Detector for each individual class.
By this way, we extend OCSto multi-class agent system
without changingthe internal architecture of OCS-agent.
Evaluating Agents’ Actions
Wedescribe a wayto evaluate agents’ actions underthe requirementsfrommultiple environmentsand the aims to meet
their actions to the multiple objectives. Anagent in a same.
¯ agent class has the sameobjectives for its actions. These
objectives are implicit requirementsfromthe environments
that it lives. Asdescribedearlier, an agent probabilistically
selects oneobjective fromn of them,selects classifiers that
matchthe conditionof the detector, and select oneclassifier
to decide its action. To correspondto multiple environments
requirements,its action for oneobjective is reflected to all
the environments,
the action will changethe status of all environments.Eachagent comparesthe effect of its action to
the environmentthat the agent has selected basedon the objective, and averagesthe effects by all agents actions. That
is, the effects by all agents are averagedin the sameclass
to the environment.However,in this way, the other agents
do not alwaysact for the sameobjectives. Thus, the agents
evaluate their actions based on the resulting environment,
whichmighthave effects of the other agents with different
objectives.
The methodmayseem invalid, however,we believe this
evaluationis adequate,becausethe agents will see the other
agents actions subjectively by their ownobjectives. They
will judge their surroundingenvironmentsby their internal
models. While we evaluate the results of actions by one
agent, this relative evaluation is adaptive to the dynamic
changes of the environmentit lives. If one agent would
evaluateits action absolutely, it wouldevaluate its actions
as suitable after the environment
changesto the state the action wouldnot be suitable any more. This relative evaluation is a reasonablewayto adapt the dynamicchangeof the
environmentsunderwhichvarious agents act with different
objectivesthey aimat.
The proposedsystemhas no explicit objective functions
globally, but eachagent Individuallyhas the ones. Theenvironmentsthey live canbe changedto the state that they never
wantto be, becausethe environments
are shared and affected
by the other classes of agents that mayact in completelydifferent ways.Arealistic tactic to fit such environments
for
agents is not to aim at on one target state as an ideal one
for them. Anagent should compromiseto the environment
it live and it mayfollowa better result that the other agents
haveattained. In this point of view, the relative evaluation
methodis a flexible way.
28
[
EnvironmentA
}[
EnvironmentB
Figure 2: Extensionof OCSto Multi-Class Agents
Onescore calculated by this evaluation is reflected to
the strength for the correspondingobjective in the classitier; however,the score is not reflected to the other n - 1
strengths.
Methods for Organizational Learning
Weprepare the two organizational learning methodsfor
agents in a sameclass.
CopyingAction Parts of a Classifier Whenagent’s actions result in a bad status, the action part pf classifiers is
substituted by the other actions, whichhaveresulted in good
results, so far. Withthis method,an agent avoids makinga
mistakein the samesituation, again. The agent adopts the
actions those maynot be caused with the same objective.
So we call this method"Copyof Action (without normative
objective)".
CopyingClassifiers or Rules The objective of copying
rules is to share the success stories by all agents. While
an agent succeededin its actions, the agent offers classifier
sets that madethe success to collective knowledgeas the
shared place or blackboard. Whenapplying GAoperations,
all agents adoptthose classifiers as the elitist ones. Bythis
method,agents get rules as its individual knowledge
that will
be able to succeedin the samesituation. Wecall this method
"Copyof Rules ; Classifier sets to share are prepared for
each objective that the correspondingclass of agents owns
them.In that case, the ratio of the classifiers that the agent
adapts for each objective is accordingto the weightof objectives.
Ecological
Marketing Simulator
Ecological Marketingiis one of general marketingmethods that approachesconsumerswith a supplier or its products keen on corresponding ecological problems. Wehave
modeledEcologicalMarketingvery simple, and simulated it
in order to validate the effectivenessof the architecturediscussed above. Theobjective of the Eco-marketingSimulator
is to find out the stable conditionswhereboth supplier and
consumerclasses of agents live by proceeding coevolving
processesto acquire their knowledge
each other.
Our agent modelof Ecological Marketingis summarized
as follows.
¯ Agent classes: "Consumers"and "Suppliers"
¯ Environmentsthat agents live: "Economy-oriented"and
"Ecology-oriented"
¯ Objectives of Supplier class: "earn muchmoney"and
"correspondto ecology"on selling their products
¯ Objectives of Consumerclass: "buy a cheaper products"
and "correspondto ecology"on buyingthe products.
Betweenthe twoobjectives of supplier agents, there is a
constraint that the higher a supplier correspondsto ecology,
the higher the cost of the productsraises. Eachsupplier can
set the price of the product as it like. Suppliers and consumershold "cash position" and "ecological score" as their
internal variables. Withtheir actions, these variables will
changeby the rules as follows.
(1) Case of Suppliers
¯ CashPosition: Increases/decreases by sales. Asupplier
mustproduceat least 2 products, so while it could sell
less than 2, the supplier looses the cost of the remained
products.
¯ Ecologicalscore: Asupplier gets a score multiply"ecology correspondenceratio of a product" by "numberof
sold products". But while a supplier could sell less than
2, the supplier gets somepenalty scores.
(2) Case of Consumers
¯ Cashposition: A consumerpays someprices of a product
it bought,and gets a meanprice of all products.
¯ Ecological score: A consumergets the ecology correspondenceratio of a product.
To.implement
the agents, the conditionalparts of the classifter contain the followinginformation:Cashposition, Ecological score, Resultsof previoussales activities (for suppliers only), results of previous purchaseactivities (for consumersonly), the averageof moneyall the agents have, the
averageof ecologicalscores all the agents have, and so on.
Theaction parts of the classifier containecologicalscoreand
the marginof the products(for suppliers) and price of products and ecological score (for consumers).
Thelengthof one classifier is 50 bits for suppliersand 40
bits for consumers.
Eachagent has 5,000classifiers in its internal model. Parameters of GAoperations are summarized
as follows: GAoperations are applied every 500 iteration,
the ratio for the elite classifier selectionis 0.2, the crossover
ratio is 1.0, andthe mutationratio is 0.001.
Preliminary Experiments: Only Supplier Agents
Learn
For the first practice, we have implementedthe consumer
class with only heuristic followingtwo rules. That is, we
have not applied GAoperations to consumeragents. Onthe
other hand, makeragents are implementedby OCS,thus,
their actions will evolve.
29
10
...
L
..~- Grayd ~ns
I ~ With b~hmllhods
Figure 3: Numberof SucceededAgents in Pre-Experiments
(Averageof 5 experiments)
¯ Rule1: First, a consumerselects a product randomly,and
then, if there is a productfit moreto ecologyand the price
is not higherthan 110%
fromthe first one, select the product.
¯ Rule2: First, a consumerselects a product randomly,and
then if there is a product whoseprice is less than 90%
fromthe first one, selects the product.
The formerheuristic is used in order for the consumer’s
objectiveto correspondto ecology,and latter is usedin order
for the consumer’sobjective to buy a cheaper product. In
this experiment,the weightsfor both objectives are set to
0.5. Wehave removedthe variety of agents to measurethe
pure effect of organizationallearning methods.
Before the experiment, we have carried out the simulation without organizational learning mechanisms;however,
we havehad results that all suppliers go bankrupt.That is,
the prices of products go downto the least level samewith
their productioncost. Accordingly,we mustapply the organizational learning methodto a supplier’s objective to earn
much money.
To overcomethe results, the following two organizational learning methodsare designed. Whena supplier goes
bankrupt, the "Copyof actions" methodis executed. This
meansthat they copythe last 5 best actions to the supplier’s
bad correspondingclassifiers. Whena supplier succeeds,
that is, its moneyposition exceedsto the upperlimit, the
supplier providesthe last 5 classifiers applied, whichcause
the last 5 actions of it, into a shared place. Thesesuccessful rules are able to be acceptedas the elites by all suppliers
in the GAoperations. But the rules the samesupplier agent
has providedare excludedfor the acceptancefor itself. The
numbersof these success rules are limited to 40timesS.
Weshowthe effects of organizational learning methodin
Figures 3 and 4. ’Success’in this figure meansthe times of
moneyposition exceeds to the upper limit are morethan 9
times that the moneyposition falls the lower limit. ’Lose’
means the times of moneyposition goes under the lower
limit than that exceeds the upper limit. Wehave measured
themevery 2,000iterations of the simulation.
---_.....
!
....
~ ._,&s__ __,,~n)
I-.-~r~a~
I
....................... I
@
m
Figure 4: Numbersof Lost Agentsin Pre-Experiments(Average of 5 experimentsExtension of OCSto Multi-Class
Agents)
Figure 6: Results of EconomicalActionsTakenby Suppliers
in Pre-Experiments(Averageof 5 experiments)
..........................................................
.......................................
m
m
e
m
Figure 5: Results of EcologicalActions TakenbySuppliers
in Pre-Experiments(Averageof 5 experiments)
Thisresult showsthat "Copyof action" is effective to improvethe initial activities of agents, and "Copyof Rule"is
also effectiveto improve
their final states.
Weobservethe effects of these methodsby validating the
states of the environments
feed backedby the actions of the
agents. Weshowin Figures 5 and 6, which represent the
transformationof environmentswith suppliers’ "ecological
actions" and "economicalactions". Theseresults showthat
our methodsworkwell on this point of view.
Automated Modification of Actions in
Ecological Marketing Simulation
Wehave foundout the results of feedbacksof suppliers’ actions to the environmentsin Figures 5 and 6 that "ecological actions" increasesatisfactorily, but "economical
actions"
rather decreases.This is becausesuppliers are too muchcorrespondingto ecological issues, whichhave resulted in the
sacrifice of earningmuchmoney.That is, suppliers’ weights
for the twoobjectivesare not suitable for its multi-objective
problems.Therefore, to improvethe situation, we carry out
the secondseries of experimentsto modifytheir objectives
dynamicallyby using the results of feed backof suppliers’
30
Figure 7: Results of Agents’Actionswith "Modificationsof
Actions"
actions to the environment.
Here, we examinea method(1) to comparethe results
of both ecological and economicalactivities of agents, then
(2) to feed backthe results towardthe next simulationsteps.
Whilethe results are increasedin the one objective and decreased in the other one, we decreasethe weightfor the increased objective. Wecall this methodas "Modificationof
Actions". Wehave improvedthe suppliers’ multi-objective
solutions with this methodas Figure 7 shows. Weset the
unit step for the weighttransformationat 0.05, here.
Weobservethat the both results of the feedbacksare increasing, evenif the results of ecologicalactions are rather
worsethan the last result. That is, this methodcan gives
good organizational learning mechanismsfor the multiobjective problem. That is, supplier agents have learned
the best weights for objectives on-the-fly mannerby this
method.
Co-Evolving Multiple Agents in Ecological
Marketing Simulation
Co-evolutionis the evolutionbetweendifferent kinds of living things with their interactive affections [Mitchell96]. In
the framework of our research, Co-Evolution i means
the evolutionary acquisition of better knowledgebetween
Figure 8: Results of Agents’Actionswith "Interventionsfor
Actions"
the supplier agent class and the consumeragent class. In
the sameclass, they learn cooperatively,on the other hand,
against the class, they learn competitively.
The agents wouldlike to get the benefit each other by
cooperatingtogether towardtheir common
objectives correspondingto "ecology", and by competitive learning, which
meansto competethe conflict of interests betweensuppliers
to earn "money"and consumersto buy cheaper "products".
This co-evolutionallearningis attained in adaptation to the
changeof environments,whichis causedby their feedbacks
of actions to the environments.Wesimulate EcologicalMarketing with supplier agents and consumeragents, both of
whichare implementedby the classifier systemmechanisms
describedearlier.
In the preliminary experiments, suppliers go bankrupt
even if they use organizational learning methodsof "Copy
of actions iand Copy of Rules"; consumers win a runawayvictory in their competition. In the real world, this
result leads insufficiently supplyof products,and this affect
consumersa bad impact. So consumersshould makeconcessions to suppliers to earn moneyso far as they do not go
bankrupt.
To put such concessionsinto practice in our simulation,
we expand the Modification of Actions imethod to coevolutional learning betweenagent classes. Whilesuppliers cannot improvethe result of their economical
actions by
themselves, the suppliers intervene consumersto change
their weight for economicalobjective lower. Wecall this
methodas Intervention of Actions i Wesimulate the coevolution of multiclass agents in the EcologicalMarketing
by the parametersas follows: the weightfor each objective
is 0.5, the variety of rules (the marginof the weightfor objectives) is 0.2. Asa result, productssettle to highecological
and lowprice ones, and suppliers earn moneyso far as they
do not go bankrupt as Figure 8 shows.
Figure 8 Results of AgentsActions with "Interventions
for Actions"
Concluding
Remarks
This paper has proposeda newarchitecture to simulate complex agents behaviorssimilar to the ones in our real world
31
that mayhavemulti-objectives,varieties in characteristics,
multi-class, and/or multiple environments.Suchsituations
are quite popularin recent distributed informationsystems.
Wehave developedsuch an agent modelwith LearningClassifier Systemsin Genetic-BasedMachineLearningliterature
and havevalidated the effectiveness of the organizational
learning methods employed in the model. Our proposed
methods: "Copyof Actions", "Copyof Rules" and "Modification of Actions" have shownthe performanceon the multiagents’ organizationallearningabouttheir suitable actions
for their multi-objectives. Wehavealso shownthe capability of the "Interventionof Actions"methodto co-evolvethe
multi-classagents.
Future research includes (1) exploring further applications of the proposedagent modelingtechniquesto the other
informationsystemproblems,(2) extendingthe architecture
to examine
the effects of the varietyof agentsto the result of
the simulation, and (3) developingmuchmorenovel organizational learning methodsfor agent based modeling.
References
AAAISpring SymposiumSeries 1996: Adaptation, Coevolution and Learning in Multiagent Systems. March2527, 1996.
Argyris, C., D. and SchonA. Organizational Learning,
Addison-Wesley,1978.
Axelrod, R. Complexityof Cooperation: Agent BasedModels of Competitionand Collaboration, Princeton University
Press, 1997.
Axtel, R. L. WhyAgents? On the Varied Motivations for
AgentComputing
in the Social Sciences. BrookingsInstitution WorkingPaper No. 17, 2000.
Bull, L., Fogarty, T. C.. EvolutionaryComputing
in Cooperative Multi-AgentSystems.In [AAA196],pp. 22-27. 1996.
Carley, K. M., Gasser, L.: Computational Organization
Theory, in Weiss, G. (ed.). Multiagent Systems: A Modem
Approach
to Distributed Artificial Intelligence, MITPress,
1999, pp. 299-330.
Duncan,R. and WeissA. Organizational Learning: Implications for Organizational Design,iStaw, B. M. (Ed.) Research in OrganizationalBehavior,Vol. I, JAI Press, 1979,
pp.75 123.
Epstein, J.M.. Axtell, R.: GrowingArtificial Societies: Social Science from the BottomUp, MITPress, 1996.
Goldberg, D.E. Genetic Algorithms in Search, Optimization, and MachineLearning, Addison-Wesley,1989.
G-refenstette, J., Daley,R. Methodsfor Competitiveand Cooperative co-evolution, In [AAA196],
pp. 45 - 50, 1996.
Haynes,T., Sen, S. LearningCasesto ResolveConflicts and
ImproveGroupBehavior, International Journal of HumanComputerStudies, Vol. 48, No. 1, pp. 31-49, 1997.
Kim, D. The Link betweenindividual and organizational
learning, iSloan Management
Review, Fall, 1993, pp. 37
50.
Lanzi, P.L., Stolzmann,W.,
Wilson,S.W.(eds.): Advances
Learning Classifier Systems Third Int. Workshop,IWLCS
2000, Springer LNAI1996, 2001.
Mitchell, M. AnIntroduction to Genetic Algorithms, The
MITPress, 1996.
Rosin, C. D., Belew, R. K. NewMethodsfor Competitive
Coevolution,EvolutionaryComputation,Vol.5, yNo.1, pp.
1-29, 1997.
Schaffer, J.D. "Multiple Objective Optimizationwith Vector EvaluatedGeneticAlgorithms,"Proceedingsof the First
International Conferenceon Genetic Algorithmsand Their
Applications, LawrenceErlbaumAssociates, Inc., Publishers, 1985, pp.93 100.
Sikora, R.,Shaw. M.J. A Double-Layered Learning Approachto AcquiringRulesfor Classication: Integrating Genetic Algorithm with Similarity-Based Learning. ORSA
Journal on Computing,Vol. 6, No. 2, pp. 174-187,1994.
Takadama K., Terano T., Shimohara K. Agent-Based
Model TowardOrganizational Computing: From Organizational Learning to Genetics -Based MachineLearning, i
Proceedingsof IEEESMC’99
Vol. II, 1999-a, pp. 604-609.
TakadamaK., Terano T, ShimoharaK., Hori K., Nakasuka
N. MakingOrganizational Learning Operational: Implication from LearningClassifier System,iJ. Computational
and MathematicalOrganization Theory, (5:3), 1999-b, pp.
229-252.
TakadamaK., Terano T., ShimoharaK. Learning Classitiers MeetMultiagentEnvironments,fin Lanzi, P.L., Stolzmann,W,Wilson,S. W.(eds.): Advancesin LearningClassifier Systems Third Int. Workshop,IWLCS
2000, Springer
LNAI1996, pp. 192-210, 2001.
Takadama K., Terano T., Shimohara K. Nongovernanee
Rather Than Governancein a MultiagentEconomicSociety,
IIEEETrans. on Evolutionary ComputationVol. 5, No.5,
pp. 535-545, 2001.
39.