Professional Documents
Culture Documents
2011 Book AdvancesInPracticalMulti-Agent
2011 Book AdvancesInPracticalMulti-Agent
2011 Book AdvancesInPracticalMulti-Agent
)
Advances in Practical Multi-Agent Systems
Studies in Computational Intelligence, Volume 325
Editor-in-Chief
Prof. Janusz Kacprzyk
Systems Research Institute
Polish Academy of Sciences
ul. Newelska 6
01-447 Warsaw
Poland
E-mail: kacprzyk@ibspan.waw.pl
Vol. 309. Isabelle Bichindaritz, Lakhmi C. Jain, Sachin Vaidya, Vol. 320. xxx
and Ashlesha Jain (Eds.) Vol. 321. Dimitri Plemenos and Georgios Miaoulis (Eds.)
Computational Intelligence in Healthcare 4, 2010 Intelligent Computer Graphics 2010
ISBN 978-3-642-14463-9 ISBN 978-3-642-15689-2
Vol. 310. Dipti Srinivasan and Lakhmi C. Jain (Eds.) Vol. 322. Bruno Baruque and Emilio Corchado (Eds.)
Innovations in Multi-Agent Systems and Fusion Methods for Unsupervised Learning Ensembles, 2010
Applications – 1, 2010 ISBN 978-3-642-16204-6
ISBN 978-3-642-14434-9
Vol. 323. Yingxu Wang, Du Zhang, Witold Kinsner (Eds.)
Vol. 311. Juan D. Velásquez and Lakhmi C. Jain (Eds.) Advances in Cognitive Informatics, 2010
Advanced Techniques in Web Intelligence, 2010 ISBN 978-3-642-16082-0
ISBN 978-3-642-14460-8
Vol. 324. Alessandro Soro, Vargiu Eloisa,
Vol. 312. Patricia Melin, Janusz Kacprzyk, and Giuliano Armano, and Gavino Paddeu (Eds.)
Witold Pedrycz (Eds.) Information Retrieval and
Soft Computing for Recognition based on Biometrics, 2010 Mining in Distributed
ISBN 978-3-642-15110-1 Environments, 2010
ISBN 978-3-642-16088-2
Vol. 313. Imre J. Rudas, János Fodor, and
Janusz Kacprzyk (Eds.) Vol. 325. Quan Bai and Naoki Fukuta (Eds.)
Computational Intelligence in Engineering, 2010 Advances in Practical Multi-Agent Systems, 2010
ISBN 978-3-642-15219-1 ISBN 978-3-642-16097-4
Quan Bai and Naoki Fukuta (Eds.)
Advances in Practical
Multi-Agent Systems
123
Quan Bai
Tasmanian ICT Centre, CSIRO
GPO Box 1538, Hobart
TAS 7001, Australia
Email: quan.bai@csiro.au
Naoki Fukuta
Faculty of Informatics
Shizuoka University
351 Johoku
Hamamatsu Shizuoka 432-8011
Japan
E-mail: fukuta@cs.inf.shizuoka.ac.jp
This work is subject to copyright. All rights are reserved, whether the whole or part
of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilm or in any other
way, and storage in data banks. Duplication of this publication or parts thereof is
permitted only under the provisions of the German Copyright Law of September 9,
1965, in its current version, and permission for use must always be obtained from
Springer. Violations are liable to prosecution under the German Copyright Law.
The use of general descriptive names, registered names, trademarks, etc. in this
publication does not imply, even in the absence of a specific statement, that such
names are exempt from the relevant protective laws and regulations and therefore
free for general use.
Typeset & Cover Design: Scientific Publishing Services Pvt. Ltd., Chennai, India.
Printed on acid-free paper
987654321
springer.com
Preface
The output from those workshops has created this unique collection that is
presented in three parts. Part I reports the use of agent technologies in assist-
ing the collaboration among humans and software mechanisms, coordination
among software agents and systems, and decision support on knowledge dis-
covery; Part II covers agent-based simulation in finance, economics and social
science; and Part III discusses agent-based applications for environmental
monitoring and disaster management.
Each workshop paper was accepted after review by at least two experts.
Further improvements were included in many papers in preparation for this
collection. Readers will be able to explore a diverse range of topics and de-
tailed discussions related to the five important themes in our ever chang-
ing world. This collection plays an important role in bridging the gap be-
tween MAS theory and practice. It emphasizes the importance of MAS in
the research and development of smart power grid systems, decision sup-
port systems, optimization and analysis systems for road traffic and markets,
environmental monitoring and simulation, and in many other real-world ap-
plications and publicizes and extends MAS technology to many domains in
this fast moving information age.
The organizers of the workshops did an excellent job in bringing together
many MAS researchers and practitioners from the Pacific-Asia region and
from all over the world. The well-received workshops at PRIMA2009 and the
publications included in this collection convincingly demonstrate the signifi-
cance and practical impact of the latest work presented in MAS.
We are certain that this collection will stimulate increased MAS research
and applications, influence many graduate students to conduct research and
development in MAS, and have a positive impact towards making our future
better by creating an increasingly smarter world.
The editors would like to thank the organizers and program committee mem-
bers of the five PRIMA2009 workshops and all the workshop participants for
their support and contribution. We would also like to thank Dr Stephen
Giugni and Dr Greg Timms from the CSIRO Tasmanian ICT Centre for
their help in improving this book.
ACCDS 2009:
Fenghui Ren, University of Wollongong, Australia
Katsuhide Fujita, Nagoya Institute of Technology, Japan
MASFE 2009:
Takao Terano, Tokyo Institute of Technology, Japan
Tohgoroh Matsui, Tohgoroh Machine Learning Research Institute, Japan
Kiyoshi Izumi, National Institute of Advanced Industrial Science and
Technology, Japan
AAECS 2009:
Alexis Drogoul UMI 209 UMMISCO, IRD, UPMC, MSI-IFI, Vietnam
Benoit Gaudou UMI 209 UMMISCO, IRD, MSI-IFI, Vietnam
Nicolas Marilleau UMI 209 UMMISCO, IRD, UPMC, France
ATDM 2009:
Serge Stinckwich, UMI UMMISCO IRD/UPMC/MSI-IFI, Vietnam
Sarvapali D. Ramchurn, University of Southampton, UK
IASNW 2009:
Andrew Terhorst, CSIRO ICT Centre, Australia
Siddeswara Guru, CSIRO ICT Centre, Australia
Contents
2.1 Roadmaps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
2.2 Multi-Agent Environment Coverage . . . . . . . . . . . . . . . . 98
3 Weighted Multi-Agent RRT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
4 Agent Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
5 Exploration of WMA-RRT Roadmap . . . . . . . . . . . . . . . . . . . . 106
6 Implementation and Experimental Results . . . . . . . . . . . . . . . . 107
6.1 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
6.2 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
7 Future Works and Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 230
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231
4 Validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 319
4.1 Validation Scenarios Design and Setup . . . . . . . . . . . . . 319
4.2 Push vs. Pull and Optimizations . . . . . . . . . . . . . . . . . . 321
4.3 Impacts of Robots’ Density . . . . . . . . . . . . . . . . . . . . . . . 321
5 Discussion and Open Issues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 322
5.1 Toward a Generic Solution . . . . . . . . . . . . . . . . . . . . . . . 322
5.2 System Re-organization . . . . . . . . . . . . . . . . . . . . . . . . . . 323
6 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 324
7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325
Multi-agent System for Blackout Prevention by Means of
Computer Simulations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327
Miroslav Prýmek, Aleš Horák, Adam Rambousek
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327
2 The Rice Power Distribution Network Simulator . . . . . . . . . . . 328
2.1 Event-Driven Agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 329
3 Model Situation Implementation . . . . . . . . . . . . . . . . . . . . . . . . 330
3.1 The Network Topology . . . . . . . . . . . . . . . . . . . . . . . . . . . 331
3.2 The Source Node . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 331
3.3 The Power Line . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333
3.4 The Consumer Node . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334
4 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336
Asking for Help Through Adaptable Autonomy in Robotic
Search and Rescue . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339
Breelyn Kane, Prasanna Velagapudi, Paul Scerri
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339
2 Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 341
3 Concept . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 343
4 Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 345
4.1 Self-monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 345
4.2 Decision Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347
5 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349
5.1 USARSim . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 349
5.2 SimpleSim . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 352
6 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355
7 Conclusion and Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 356
Conceptual Framework for Design of Service Negotiation
in Disaster Management Applications . . . . . . . . . . . . . . . . . . . . . . . 359
Costin Bădică, Mihnea Scafeş
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 359
2 Service Negotiation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 362
XVIII Contents
The autonomy and intelligence of software agents have greatly enhanced au-
tomation in many operational domains. Benefits of using software agents are
the ability to assist in the collaboration among humans and software mecha-
nisms, coordination among software agents and systems, and decision support
on knowledge discovery. In general, agents may perform different behaviors
for different kinds of expectations. In cooperative systems, agents will per-
form tasks more cooperatively in order to achieve their common goals, while
in competitive systems, agents may compete with others in order to reach
individual aims. However, in most real world situations, both cooperation
and competition exist between agents. Agents need to be carefully organized
and managed in order to maximize the efficiency of the whole system, and to
balance benefits between others.
We solicited all topics related to the interdisciplinary aspects of agent-
based collaboration, coordination, and decision support as much as these
are in some way relevant for or applicable to software agents and multi-
agent systems. The following incomprehensive listing gives examples of such
potential topics:
- Agent-based collaboration
- Agent-based coordination
- Methodologies, languages and tools to support agent collaboration
- Protocols and approaches for agent-mediated collaboration and decision
support
- Teamwork, coalition formation and coordination in multi-agent systems
- Distributed problem solving
- Electronic markets and institutions
- Game theory (cooperative and non-cooperative)
- Auction and mechanism design
- Argumentation, negotiation and bargaining
- Trust and reputation
- Agent-based decision support
- Agent-based knowledge discovery
- Agent-based collective intelligence
- Agent-based situation awareness
These issues are being explored by researchers from different communities
in Autonomous Agents and Multi-Agent systems, distributed problem solv-
ing, complex systems as well as agent-based applications. This part collects
nine high quality papers, all involving some important aspect of agent-based
collaboration, coordination and decision support. The idea is to try to moti-
vate the researchers to explore cutting-edge challenges in the area and accel-
erate progress towards scaling up to larger and more realistic applications.
Katsuhide Fujita
Fenghui Ren
ACCDS’09 Workshop Chairs
DGF: Decentralized Group Formation
for Task Allocation in Complex
Adaptive Systems
1 Introduction
Allocating tasks to agents in complex adaptive systems is a significant re-
search issue. Task allocation problem can be simply described as that when an
agent has a task which it cannot complete by itself, the agent then attempts
to discover other agents which contain the appropriate resources and assigns
the task, or part of the task, to those agents. Recently, many approaches have
been proposed for task allocation. Some of them [6] [9] [14] [16] are based on
the centralized fashion which assumes that there is a central controller to
assign tasks to agents. An efficient agent group formation algorithm [9] was
proposed by Manisterski et al. for completing complex tasks. Their work took
both cooperative and non-cooperative settings into account and pointed out
Dayong Ye · Minjie Zhang
School of Computer Science and Software Engineering,
University of Wollongong, Australia
e-mail: {dy721,minjie}@uow.edu.au
Danny Sutanto
School of Electrical, Computer and Telecommunications Engineering,
University of Wollongong, Australia
e-mail: danny@elec.uow.edu.au
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 3–19.
springerlink.com c Springer-Verlag Berlin Heidelberg 2010
4 D. Ye, M. Zhang, and D. Sutanto
what can and cannot be achieved. However, their algorithm assumed there is
a central controller, which uses global information, to manage the group for-
mation. Zheng and Koenig [16] provided reaction functions for task allocation
to cooperative agents. The objective of their approach is to find a solution
with a small team cost and assign each task to the exact number of different
agents. Their work also assumed that there is a central planner to allocate
tasks to agents. Centralized fashion can make the allocation process effective
in a small scale network, since the central planner has a global view of the
system and is, therefore, aware of the capabilities of each agent. In that case,
communication overhead during allocation processes could be reduced. How-
ever, the centralized fashion has also several notable disadvantages. Firstly,
in some complex adaptive systems, it is difficult to have such a central con-
troller, such as social networks or sensor networks. In these networks, it is
hardly for a single agent to have the global view with regard to the network
but only the local prospect about directly linked neighbors. Secondly, when
the central planner is out of order or cracked by some attackers, task al-
location might fail accordingly. The last drawback is that the scalability of
a system with centralized control is limited because when too many agents
exist, the central controller has to maintain much information to hold the
global view and respond huge amount of messages from other agents. These
situations might drastically raise the CPU and memory usage of the central
controller and also consume much network bandwidth.
In order to overcome the shortcomings of centralized fashion, some dis-
tributed task allocation methods were developed based on decentralized style.
Compared with centralized fashion, decentralized style is more scalable and
robust. However, there are still a few limitations in current decentralized task
allocation methods. Some of these methods do not assume the existence of
any agent networks, but emphasize only on the behaviors of resource sup-
pliers and demanders, e.g. [13] [11] [8] [1] [2] and [3]. On the other hand,
several distributed task allocation mechanisms have to employ some global
information to achieve their goals, although the execution procedure of these
mechanisms are distributed. For example, in [10], the proposed resource allo-
cation method supposed that the resource load of resource suppliers is global
information which are known by all resource demanders. Furthermore, some
other approaches, e.g. Greedy Distributed Allocation Protocol (GDAP) pre-
sented in [15], only allow agents to request help for tasks to their directly
linked neighbors. In this way, the possibility of task allocation failure would
be increased due to the limited resources available from agents’ neighbors.
Some other group or coalition formation schemes have also been proposed,
although they were not devised for task allocation. Nonetheless, there are
more or less limitations of these schemes, and, therefore, they cannot be
directly used for task allocation in complex adaptive systems. A coalition
formation based self-organization technique was presented in [12]. The self-
organization process is through negotiation strategies, whose aim is maximiz-
ing the global utility of a sensor network. The disadvantage of this technique
DGF: Decentralized Group Formation for Task Allocation 5
For efficient task allocation, it is supposed that only an IDLE agent can
be assigned to a new task as an Initiator or a partial fulfilled task as a
Participant, or committed to a partial fulfilled task as a Mediator. A partial
fulfilled task is a task, for which a full group is in formation procedure and
has not yet formed. After formalizing the task allocation problem, in the next
subsection, the principle of DGF will be depicted.
attempts to discover the agents with available resources from its neighbors. If
any agents satisfy resource requirement, the Mediator aM will send a response
message, RespM ess, back to the Initiator aI . The Initiator aI then directly
makes contract with the agents which satisfy the resource requirement. If
the neighboring agents of the Mediator aM cannot satisfy the resource re-
quirement either, the Mediator aM will commit the partial fulfilled task tI to
one of its neighbors again. This process will continue until all of the resource
requirements of task tI are satisfied, or the T T L(tI ) reaches 0, or there is
no more IDLE agent among the neighbors. Both of the last two conditions,
i.e. T T L(tI ) = 0 and no more IDLE agent, demonstrate the failure of task
allocation. In these two conditions, the Initiator aI disables the assigned con-
tracts with the Participant s, and the states of these Participant s are reset to
IDLE.
When finishing an allocation for one task, the system is restored to its
original status and each agent’s state is reset to IDLE.
From the above description, it can be found that the proposed decentral-
ized group formation algorithm, DGF, enables Initiator s to request help not
only from neighbors but also other agents through commitment if needed.
Algorithm 2 gives the pseudocode of the decentralized group formation al-
gorithm employed by each Initiator during a task allocation process.
(14) break;
(15) else
(16) select ak as Mediator based on the number of ak ’s neighbours
: ak ∈ N eig(ai ) and State(ak ) = IDLE
(17) Pi ← Pi ◦ Pk ;
(18) State(ak ) ← M ediator;
(19) for ∀Pi , Pi : Pi ∈ P ∧ Pi ∈ P in sequential order
(20) Pi ← Pi ∪ Pi
end
After each agent adapts its neighbors linkage, it resets its counters of both
group joining attempts and group joining successes to 0, in an effort to es-
tablish an estimate of popularity with the new system structure.
a manager has received offers from all neighbors but still cannot satisfy its
task, the task is removed from its task list.
of different resources is 10, i.e. ε = 10. The TTL value of each task is set
to 4. The average degree is fixed at 6. In order to achieve the popularity
of each agent, the task allocation process will be run in 2000 rounds as an
initial period with no system adaption. Afterward, the system is allowed to be
evolved with the adaption strategy described in Section 3. For comparison,
we utilize another adaption strategy as a standard, i.e. Agent Organized
Network (AON) strategy, which was presented in [5]. The AON strategy is
briefly described as follows. If an agent, ai , decides to adapt its neighboring
linkage, the agent will remove the connection from the neighboring agent,
aj , which has the lowest popularity. The agent ai then requests a referral
from its neighbor with the highest popularity, say al . Finally, the agent ai
will establish a new direct link with the agent ak that is the most popular
neighbor of al . It is apparent that the AON strategy depends on neighboring
agents in most extent. An agent with the adaptation strategy proposed in
this paper, i.e. Novel System Adaptation Strategy (NSAS), can choose a
new neighbor from those candidate agents which cooperated with or were
requested by the agent (mentioned in Section 3). Thus, compared with AON
strategy, NSAS let an agent have more candidates when the agent decides to
adapt its neighbors linkage.
In this experiment, three criteria are used to evaluate the performance of
the two approaches.
CpR (Completion Rate of tasks): The proportion of the number of success-
fully allocated tasks to the total number of tasks in the experiment environ-
ment, namely:
# of successf ully completed tasks
CpR = (2)
|T |
Proof. In each iteration in the worst case (i.e. a fully connected system), for
each of the O(m) Initiators, O(n) resource query messages are sent. Then,
the O(n) available resource response messages are generated and sent back to
the Initiators. The messages generated during this phase are O(2mn). Next,
each of the O(m) Initiators forms O(m) groups with their O(n) neighbors.
14 D. Ye, M. Zhang, and D. Sutanto
Thus, the messages created at this stage are O(mε). After that, each of
the O(m) Initiators commits O(m) tasks to one of their neighboring agents,
and therefore sends O(m) messages. This process will be iterated in O(T T L)
hops. Hence, the messages generated during commitment process are O(T T L·
(2mn + mε + m)). Totally, during a task allocation round, the number of
communication messages created are O(2mn+mε+T T L·(2mn+mε+m)) =
O(T T L · m(n + ε)).
there are more neighbors, Initiator has higher probability to derive sufficient
resources to allocate its tasks successfully.
Fig. 1(b) displays the number of messages (CmC) of the two approaches
in different situations generated in a task allocation round. As DGF would
request other agents for help when resources from neighbors are insufficient,
the communication cost of DGF is higher than that of GDAP. Thus the
presentation of GDAP in this test is relatively good due to its considering
only neighbors which could decrease the number of messages created during
task allocation processes. It should also be noticed that with the increase of
average degree, the distinction between DGF and GDAP is reducing. This
situation could be explained that with more average number of neighbors,
Initiators have more opportunity to solve their tasks based on only their
neighbors in both approaches. Therefore, for DGF, some Initiator s do not
need to commit their tasks to other agents, and consequently save communi-
cation cost.
Fig. 1(c) shows the Coefficient of Performance (CoP ) of the two approachs
in different cases. With the average degree ascending, the CoP of GDAP
declines, while the CoP of DGF increases gradually. It should be recalled that
lower CoP means better performance. For GDAP, although communication
cost rises with the increase of average degree, the number of allocated tasks
also increases. Therefore, for each allocated task, the communication cost
remains constant in principle. On the other hand, with the increase of average
degree for DGF , the completion rate of tasks keeps steady displayed in
Fig. 1(a), while the number of messages, i.e. communication cost, mounts up
persistently demonstrated in Fig 1(b). Hence, for each successfully completed
task, the communication cost increases. However, DGF still has lower CoP ,
16 D. Ye, M. Zhang, and D. Sutanto
more agents and tasks in the network, and in order to allocate these tasks,
more communication steps cannot be avoided which results in communication
messages rising.
Fig. 2(c) demonstrates the Coefficient of Performance (CoP ) of the two
approaches. It can be found that the CoP of both DGF and GDAP keeps
almost steady with the increase of network scale. The reason of this result
is that even though the number of communication messages ascends, the
absolute number of successfully completed tasks also increases, and, thus,
the CoP can keep stable.
The conclusion can be achieved that DGF has better scalability than
GDAP, since both CpR and CoP of DGF can be preserved while GDAP
can maintain only CoP but CpR of GDAP decreases gradually.
CmC and CoP . This can be explained that an agent with AON strategy only
chooses one of its neighbors’ neighbors as its new neighboring agent, while an
agent with NSAS has more candidates when selecting a new neighbor, and,
hence, the agent is very likely to create a better neighboring linkage. It should
also be observed that, for both strategies, the ratio with which performance
is improving slows down over time. This implies that the system structure
converges to stability with the increase of task allocation rounds.
References
1. Abdallah, S., Lesser, V.: Modeling task allocation using a decision theoretic
model. In: Proceedings of AAMAS 2005, Utrecht, Netherlands, pp. 719–726
(2005)
DGF: Decentralized Group Formation for Task Allocation 19
2. Abdallah, S., Lesser, V.: Learning the task allocation game. In: Proceedings of
AAMAS 2006, Hakodate, Hokkaido, Japan, pp. 850–857 (2006)
3. Bachrach, Y., Rosenschein, J.S.: Distributed multiagent resource allocation in
diminishing marginal return domains. In: Proceedings of AAMAS, Estoril, Por-
tugal, pp. 1103–1110 (2008)
4. Barabasi, A.L., Albert, R.: Emergence of scaling in random networks. Sci-
ence 286(5439), 509–512 (1999)
5. Gaston, M.E., desJardins, M.: Agent-organized networks for dynamic team
formation. In: Proceedings of AAMAS 2005, Utrecht, Netherlands, pp. 230–
237 (2005)
6. Kraus, S., Shehory, O., Taase, G.: Coalition formation with uncertain hetero-
geneous information. In: Proceedings of AAMAS 2003, Melbourne, Australia,
pp. 1–8 (2003)
7. Kraus, S., Shehory, O., Taase, G.: The advantages of compromising in coalition
formation with incomplete information. In: Proceedings of AAMAS 2004, New
York, USA, pp. 588–595 (2004)
8. Lerman, K., Shehory, O.: Coalition formation for large-scale electronic markets.
In: Proceedings of ICMAS 2000, Boston, Massachusetts, USA, pp. 167–174
(2000)
9. Manisterski, E., David, E., Kraus, S., Jennings, N.R.: Forming efficient agent
groups for completing complex tasks. In: Proceedings of AAMAS 2006, Hako-
date, Hokkaido, Japan, pp. 834–841 (2006)
10. Schlegel, T., Kowalczyk, R.: Towards self-organising agent-based resource allo-
cation in a multi-server environment. In: Proceedings of AAMAS 2007, Hon-
olulu, Hawai’i, USA, pp. 74–81 (2007)
11. Shehory, O., Kraus, S.: Methods for task allocation via agent coalition forma-
tion. Artificial Intelligence 101(1-2), 165–200 (1998)
12. Sims, M., Goldman, C.V., Lesser, V.: Self-organization through bottom-up
coalition formation. In: Proceedings of AAMAS 2003, Melbourne, Australia,
pp. 867–874 (2003)
13. Smith, R.G.: The contract net protocol: High-level communication and control
in a distributed problem solver. IEEE Trans. Computers C-29(12), 1104–1113
(1980)
14. Theocharopoulou, C., Partsakoulakis, I., Vouros, G.A., Stergiou, K.: Overlay
network for task allocation and coordination in dynamic large-scale networks
of cooperative agents. In: Proceedings of AAMAS 2007, pp. 295–302 (2007)
15. Weerdt, M.D., Zhang, Y., Klos, T.: Distributed task allocation in social net-
works. In: Proceedings of AAMAS 2007, Honolulu, Hawaii, USA, pp. 500–507
(2007)
16. Zheng, X., Koenig, S.: Reaction functions for task allocation to cooperative
agents. In: Proceedings of AAMAS 2008, Estoril, Portugal, pp. 559–566 (2008)
Cellular Automata and Immunity
Amplified Stochastic Diffusion Search
Abstract. Nature has often provided the inspiration needed for new com-
putational paradigms and metaphors [1,16]. However natural systems do not
exist in isolation and so it is only natural that hybrid approaches be explored.
The article examines the interplay between three biologically inspired tech-
niques derived from a plethora of natural phenomena. Cellular automata with
their origins in crystalline lattice formation are coupled with the immune sys-
tem derived clonal selection principle in order to regulate the convergence of
the stochastic diffusion search algorithm. Stochastic diffusion search is itself
biologically inspired in that it is an inherently multi-agent oriented search
algorithm derived from the non-stigmergic tandem calling / running recruit-
ment behaviour of ant species such as Temnothorax albipennis. The paper
presents an invesitigation into the role cellular automata of differing com-
plexity classes can play in order to establish a balancing mechanism between
exploitation and exploration in the emergent behaviour of the system. . .
1 Background
Respect for the resourcefulness of the ant runs deep in humankind flowing
through a gambit of cultures crossing East and West. The Korean flood myth
tells how one of the last human survivors was assisted by ants in sifting
between grains of rice and of sand [18]. In the Middle East the Book of
Proverbs, Chapter 6, admonishes the reader to go to the ant who without
commander, overseer or ruler manages to store provisions in summer. While
in the West one of the the greek fables of Æsop tells of the industrious ant
and the lazy grasshopper.
Duncan Coulter · Elizabeth Ehlers
University of Johannesburg, Johannesburg, South Africa
e-mail: {dcoulter,emehlers}@uj.ac.za
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 21–32.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
22 D. Coulter and E. Ehlers
AgentPool.initialise()
while(AgentPool.activeAgents < Threshold)
AgentPool.foreach((x) => x.test())
AgentPool.foreach((x) => x.diffuse())
end while
SDS may be improved by allowing for the dynamic introduction of new agents
during the diffusion phase. When an agent applies the partial evaluation
function to its current hypothesis and receives a utility value higher than the
threshold then its recruitment behaviour may proceed in one of two different
ways based on the current available system resources.
Should sufficient system resources be available an active agent is allowed
to spawn a new copy of itself (with a slightly modified hypothesis) in lieu
of utilising the one-to-many communication pattern described earlier. This
allows the amplified SDS algorithm to embed a memory of the utility land-
scape which may prove useful in accommodating temporary perturbations in
dynamic environments. In living organisms cells have limited lifespans the
failure of this mechanism leading to cancer through unwarranted cell immor-
tality and a similar effect would be true in amplified SDS should the lifespan
of newly spawned agents not be strictly controlled. Three approaches may be
adopted in amplified SDS. . .
• Firstly, the lifespan of each agent in the system may be limited to a certain
number of diffusion phases. In this case it is very important that the prob-
ability of new agent instantiation be high enough to prevent the system
from losing all agents and hence entering into an irrecoverable state.
• Secondly, the total size of the agent pool may be expressed as a function
of the available system resources with agents being spawned and culled as
needed on a phase by phase basis.
• Lastly, an agent may be permitted the ability to perform only a limited
number of new agent instantiations before being removed.
the languages standard library supports a robust support for actors as a unit
of concurrency. Actors also provided a convenient mechanism with which
to implement the message passing required by SDS. Initially the JADE li-
brary [2] was considered in order to provide FIPA compatible communication
as the Scala language compiles directly to Java bytecode which runs on Java’s
Virtual Machine and hence supports a high degree of interoperability with
existing Java classes. However JADE imposes a layered behavioural decom-
position on the agent design. While such a design definitely possible for both
SDS and clonal expansion was considered not to be the most suitable choice
considering that interoperability with existing FIPA systems was not a re-
quirement for the prototype.
This doest not suggest that actors equate with agents by any means. Actors
and software agents, while sharing superficial similarities such as a use of
message passing are not equivalent. Actors are simply a concurrency control
mechanism and do not equate with agents in the same way that threads do
not equate with agents. An agent may well be implemented using one or more
actors or indeed one or more threads.
Similar to the technique employed in [3] the source of the neighbourhood
topology is obscured from the participating agents according to the Facade
design pattern. As an improvement to the previously utilised technique the
topology facade was this time realised using a trait rather than a conventional
Java interface. This allows the trait to maintain some state of its own instead
of simply defining a contact of interaction. The state maintained in this case is
the mapping between agents and cells in the underlying CA. Two approaches
to the maintenance of the CA in light of the dynamic addition and removal
of agents necessitated by clonal selection were considered.
1. Additional agents may be simply chained to the same cell as the agent
which instantiated them. This approach while converging some memory
and avoiding the question of how to initialise the new cells in an expanded
CA does suffer from a serious drawback. Since each newly instantiated
agent occupies the same cell as its originator they are eternally neighbours.
Over time in high churn scenarios (configurations where there is large
degree of agent addition and removal) the causal relationship between
originator agents and their offspring will come to dominate neighbourhood
determination to a degree which completely marginalises the underlying
complexity class of the automaton.
2. Alternatively, the CA can be expanded so that the one-to-one mapping
of agents to cells is maintained. The question then needs to be addressed
regarding how to initialise the newly created cells. The danger of not ini-
tialising them is that the newly added agents are immediately part of the
same contiguous cell block and hence neighbourhood. This means that in
class I or class II automata the danger of new agents immediately getting
stuck on the local maximum of the highest utility agent remains high. It
is the case that often a completely random configuration would disrupt
any persistent structures created by a class IV automaton. In the end a
30 D. Coulter and E. Ehlers
3 Results
The results presented here show the aggregate behaviour of the system in
terms of rate of convergence across several fitness landscapes defined by in-
stances of each of the four main classes of CA for solution space defined on a
power-law distribution as illustrated in Figure 4. For the purposes of the sim-
ulation the threshold value for terminating the system was kept unattainably
high in order to ensure an indefinitely long run.
4 Conclusion
The paper explored several incremental improvements to the SDS algorithm.
These improvements took the form of clarifying the potential shortfalls of
“fixing” the hypotheses of active agents in dynamic environments along the
steps to be taken when such changes occur. In this case the option of relaxing
the allowable recipients of recruitment messages to include other active agents
was broached. The paper also discussed the benefit of introducing minor
variations into the propagated hypotheses during the diffusion phase of the
algorithm in the interests of avoiding local maxima.
Clonal selection has an unambiguous amplification effect on the exploita-
tive nature of the system by way of increasing the population of active agents.
In biological immune systems the amplifying effect of clonal selection is tem-
pered by the lifetime of individual lymphocytes [4] . The agent topology’s
effect on convergence is directly tied to the underlying complexity class of its
defining CA.
1. Class I CA had a negative effect on the exploitative nature of the system
without a comparative increase in the explorative nature of the search.
2. Class II CA had neither a positive effect of exploration nor much effect on
exploitation
3. Class III CA had a positive effect on exploration and no major effect on
exploitation
4. Class IV CA appear to provide the best compromise between exploration
and exploitation although more research is needed in this regard.
To summarize the paper presents the use of several techniques to regulate
the emergent exploration versus exploitation properties of the stochastic dif-
fusion search algorithm. In keeping with biologically inspired metaphor of the
original algorithm the major approaches described in the paper draw their in-
spiration from the natural world. Clonal expansion provides for a mechanism
to enhance exploitation while cellular automata defined recruitment com-
munication channels provide for a mechanism to regulate exploration versus
exploitation, at run time, simply by specifying a single integer parameter.
References
1. Michalek, R., Tarantello, G.: Design Patterns from Biology for Distributed
Computing. ACM Transactions on Autonomous and Adaptive Systems 1, 26–
66 (2006)
2. Bellifemine, F., Caire, G., Greenwood, D.: Developing Multi-Agent Systems
with JADE. John Wiley & Sons Ltd., Chichester (2007)
3. Coulter, D., Ehlers, E.: Federated Patent Harmonisation: An evolvable agent-
wise approach. In: Proceedings of TMCE Symposium, vol. 2, pp. 1391–1393
(2008)
4. de Castro, L., Timmis, J.: Artificial Immune Systems A New Computational
Intelligence Approach. Springer, Heidelberg (2002)
32 D. Coulter and E. Ehlers
Abstract. When searching for information a user wants, search engines often
return lots of results unintended by the user. Query expansion is a promising
approach to solve this problem. In the query expansion research, one of the
biggest issues is to generate appropriate keywords representing the user’s
intention. The Related Word Extraction Algorithm (RWEA) we proposed
extracts such keywords for the query expansion. In this paper, we evaluate the
RWEA through several experiments considering the types of queries given by
the users. We compare the RWEA, Robertson’s Selection Value (RSV) which
is one of the famous relevance feedback methods, and the combination of
RWEA and RSV. The results show that as queries become more ambiguous,
the advantage of the RWEA becomes higher. From the points of view of query
types, the RWEA is appropriate for informational queries and the combined
method is for navigational queries. For both query types, RWEA helps to
find relevant information.
1 Introduction
The World Wide Web, a treasure house of knowledge, is widely used not only
by professionals for supporting their research or business but also by non
professionals in their daily life. In fact, it is very common for people these
days to use a search engine before looking into dictionaries or textbooks when
they want to know something.
However, search engines often fail to give results that users want. One
of the reasons is that the keyword (or query) submitted by the user is so
ambiguous for the search engine to grasp his/her intention. Accordingly the
Tetsuya Oishi · Tsunenori Mine · Ryuzo Hasegawa · Hiroshi Fujita ·
Miyuki Koshimura
Faculty of Information Science and Electrical Engineering, Kyushu University,
Motooka 744, Nishi-ku, Fukuoka-shi, Fukuoka-ken 819-0395 Japan
e-mail: {oishi,hasegawa,fujita,koshi}@ar.is.kyushu-u.ac.jp,
mine@al.is.kyushu-u.ac.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 33–48.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
34 T. Oishi et al.
2 Related Work
There have been a lot of studies on the query expansion. Wang and Davison
[13] built an auto-tagging system to assign tags to Web pages. The system
used a support vector machine (SVM) for classifying tags. After successfully
having constructed the auto-tagging system with good performance, they
took advantage of it for suggesting query expansion to the user. The result
pages of expanded queries created by the system were significantly better
than those of the Google.
Related Word Extraction Algorithm for Query Expansion 35
3 Experimental Condition
We evaluate the result of the query expansion in our experiment. In this
section, we explain the query expansion system. Additionally, we show the
types of the queries used for the query expansion.
36 T. Oishi et al.
Chirita et al. [2] [4] classified the queries into three types: the clear queries
(P @10 ≥ 70)2 , the semi-ambiguous queries (20 ≤ P @10 < 70) and the
ambiguous queries (P @10 < 20). From the points of view of ambiguity of the
queries, these are divided into 27 clear queries, 48 semi-ambiguous queries,
and 75 ambiguous queries.
Broader [1] classified the queries into three types (informational, naviga-
tional, and transactional query) based on the user’s needs. The intent of the
informational query is to acquire some information assumed to be present
on one or more web pages. The intent of the navigational query is to reach
a particular site that the user has in mind. The intent of the transactional
query is to perform some web-mediated activity, such as shopping, download-
ing various type of file, accessing certain database, finding services and so on.
From the user intent behind queries, these are divided into 99 informational
queries and 51 navigational queries.
4.1 RWEA
The algorithm extracts, from the text T , a group of words that are related
to a given set of keywords K. The extracted words are thus expected to be
relevant to K. Based on the textual distance from a keyword in K, we evaluate
each word in T , and output the words ranked according to the evaluation.
Importance of the distance between sentences
We regard the distance between sentences as important when considering
relevance of words. Moreover we pay attention to the order of the sentences
that contain some specific word (for example, the query word used for
Web retrieval, the word in the title of schedule, and so on) appearing in
a document. Our algorithm is based on the following idea: “If a certain
word A in a certain sentence is an important word for the user, the word
nearby this sentence is also important.”
Preparations
We define some symbols as follows:
• K: a set of keywords that give a basis for extracting related words.
• T : a text where some keywords related to K are extracted.
2
This formula means that the precision at 10 retrieved documents is larger than
70%.
38 T. Oishi et al.
keyword K A B
tj in text T A G B E G A F C F H D E
BVA (tj ) 5 4 3 2 1
BVB (tj ) 5 4 3 2 1
BVA (tj ) 3 4 5 4 3
BV (tj ) 13 12 11 8 5
EBV (j) 3 3.6 3.8 3.6 3
EBV (tj ) 4.33 3.33 2.89 2.22 1.67
keyword K A B
tj in text T A G B E G A F C F H D E
W EBV (wl ) 4.33 4.33 4.33 3.33 3.33 2.89 2.89 2.89 2.22 2.22 1.67 1.67
AveEBV (cp ) 3.61 3.83 4.33 2.50 3.83 3.61 2.56 2.89 2.56 2.22 1.67 2.50
tf (cp ) 2 2 1 2 2 2 2 1 2 1 1 2
V (cp ) 1.28 1.28 1 1.28 1.28 1.28 1.28 1 1.28 1 1 1.28
S(cp ) 4.62 4.90 4.33 3.19 4.90 4.62 3.27 2.89 3.27 2.22 1.67 3.19
cp A B C D E F G H
S(cp ) 4.62 4.33 2.89 1.67 3.19 3.27 4.90 2.22
Let EBV (tj ) be the value evaluated after flattening. It is calculated using
the following formula;
EBV (tj ) = BV (tj )/EBV (j).
Calculating S(cp )(see Table 2)
In addition to the evaluation based on the distance between sentences, we
take into consideration the concept of Term Frequency that is often used in
the TF/IDF method. First, we compute the average AveEBV (cp ) of the
expected evaluation values W EBV (wl ), for word wl that appears several
times. W EBV (wl ) = EBV (tj ) when wl in tj .
Moreover, we compute the weight V (cp ) using the tf value of word cp as
V (cp ) = 1 + {tf (cp )/o} log tf (cp ).
The parameters used in the above formula are as follows. o: the total
number of occurrences of words in T (mentioned above), and tf (cp ): the
number of occurrences of cp in T.
Using AveEBV (cp ) and V (cp ), we calculate the evaluation value S(cp ) of
cp in T as S(cp ) = AveEBV (cp ) × V (cp ). We assume that S(cp ) is the
evaluation value of word cp in text T in this algorithm.
4.2 RSV
RSV gives higher weight to the word w which only appears in the relevant
documents and does not appear in non-relevant documents than others.
RSVw is calculated as follows:
− −
df + +
+dfw +
){α log R+ +R −
dfw
RSVw = w
( R+ − R+ +R− dfw +dfw
+ +
(dfw +0.5)/(R+ −dfw +0.5)
+(1 − α) log − − } (1)
(dfw +0.5)/(R− −dfw +0.5)
5 Experiment
In this section, we explain our experiments. The setting of our experiments is
described in Section 5.1. In Section 3, we show the results of our experiments.
Fig. 1 Sorting the Web pages according to the order of the user’s evaluation
Even : when the pages include the information partially relevant to the
query.
Bad : when the pages do not include the information relevant to the
query.
Very Bad : when the pages include the information completely irrelevant
to the query.
2. Sort the Web pages
The system sorts the Web pages according to the order of the user’s eval-
uation by using the explicit feedback. The explicit feedback is the method
performed based on the user’s evaluation. For pages which have the same
evaluation, the order is preserved as in the original rank. Figure 1 shows
how the system sorts the evaluated Web pages according to the order of
evaluation value.
3. Decide the relevant documents and non-relevant documents
Here we decide the relevant and non-relevant documents by using the
pseudo or explicit feedback. The pseudo feedback means that we take
top-5 Web pages as relevant documents and bottom-5 Web pages as non-
relevant from the unsorted Web pages. The explicit feedback means that
42 T. Oishi et al.
we use the Web pages sorted based on the user’s evaluation to decide the
relevant and non-relevant Web pages.
5.2 Result
We compare the three methods for the query expansion by calculating MAP
(Mean Average Precision). We assume that the number of all correct answers
for calculating MAP is 20 because the number of all retrieved results for each
query is 20. In other words, the number of all correct answers is not bigger
than 20.
First, we describe a general trend of the methods for all the queries (in
5.2.1). Next, we evaluate the methods based on the types of the queries (in
5.2.2).
and the ratio between the MAP values. The ratio for the method using the
explicit feedback (“com-e”, “rsv-e”, and “rwe-e”) is larger than 1.0. In other
words, applying RF to the original rank makes MAP worse.
We re-ranked each retrieved result using Residual Collection (RC) [6]. RC
re-ranks the results retrieved with the expanded query except for the Web
pages retrieved with the query before the expansion. RC is the method that
the newly retrieved results are important. In other words, we calculate MAP
only for the Web pages newly retrieved with the expanded query. Figure 4
shows MAP with the results re-ranked by RC as well as Figure 2. We can
see that “rwe-e” always shows the highest MAP among the query expansion
methods. This shows that the expanded query created by RWEA can get new
useful Web pages.
(2-1)One-Word Query
Queries Initial Query Expanded Query
(# of queries) goo com-e com-p rsv-e rsv-p rwe-e rwe-p
clear query MAP 0.712 0.444 0.494 0.416 0.466 0.352 0.337
(10) Improvement - 0.623 0.694 0.584 0.654 0.495 0.474
semi-ambiguous MAP 0.197 0.091 0.130 0.094 0.094 0.154 0.049
query (16) Improvement - 0.464 0.660 0.478 0.476 0.781 0.248
ambiguous query MAP 0.012 0.030 0.011 0.020 0.008 0.032 0.016
(124) Improvement - 2.452 0.895 1.600 0.679 2.592 1.284
(2-2)Two-Word Query
Queries Initial Query Expanded Query
(# of queries) goo com-e com-p rsv-e rsv-p rwe-e rwe-p
clear query MAP 0.649 0.483 0.488 0.414 0.438 0.453 0.471
(34) Improvement - 0.744 0.752 0.638 0.675 0.698 0.725
semi-ambiguous MAP 0.232 0.251 0.163 0.225 0.135 0.217 0.197
query (65) Improvement - 1.082 0.702 0.970 0.583 0.935 0.848
ambiguous query MAP 0.054 0.099 0.052 0.082 0.048 0.092 0.048
(51) Improvement - 1.828 0.954 1.514 0.885 1.696 0.878
(2-3)Three-Word Query
Queries Initial Query Expanded Query
(# of queries) goo com-e com-p rsv-e rsv-p rwe-e rwe-p
clear query MAP 0.664 0.540 0.455 0.460 0.408 0.505 0.510
(37) Improvement - 0.812 0.684 0.692 0.615 0.761 0.768
semi-ambiguous MAP 0.234 0.257 0.184 0.242 0.179 0.191 0.146
query (63) Improvement - 1.099 0.788 1.034 0.765 0.815 0.625
ambiguous query MAP 0.031 0.040 0.030 0.033 0.030 0.073 0.053
(50) Improvement - 1.297 0.986 1.089 0.996 2.385 1.738
(3-1)Informational Query
Queries Initial Query Expanded Query
(# of queries) goo com-e com-p rsv-e rsv-p rwe-e rwe-p
clear query MAP 0.700 0.543 0.500 0.459 0.461 0.494 0.505
(56) Improvement - 0.776 0.715 0.655 0.658 0.705 0.721
semi-ambiguous MAP 0.220 0.216 0.143 0.210 0.151 0.181 0.150
query (96) Improvement - 0.982 0.649 0.954 0.687 0.822 0.681
ambiguous query MAP 0.026 0.045 0.025 0.036 0.026 0.060 0.033
(145) Improvement - 1.754 0.979 1.387 0.991 2.333 1.265
(3-2)Navigational Query
Queries Initial Query Expanded Query
(# of queries) goo com-e com-p rsv-e rsv-p rwe-e rwe-p
clear query MAP 0.584 0.416 0.413 0.383 0.354 0.400 0.400
(25) Improvement - 0.713 0.708 0.656 0.607 0.686 0.686
semi-ambiguous MAP 0.247 0.276 0.220 0.234 0.147 0.233 0.174
query (48) Improvement - 1.118 0.892 0.947 0.596 0.945 0.707
ambiguous query MAP 0.026 0.052 0.023 0.039 0.016 0.044 0.029
(80) Improvement - 2.026 0.889 1.493 0.624 1.711 1.111
6 Conclusion
In this paper, we have discussed three types of query expansion methods; the
Related Word Extraction Algorithm (RWEA), Robertson’s Selection Value
(RSV) and the method combining RWEA and RSV considering several types
of queries.
The explicit feedback worked better than the pseudo feedback. For all
queries, the combined method works best on the average.
According to Chirita’s classification, the MAP of RWEA for the ambiguous
queries was the highest. Based on the classification by the number of words in
a query, RWEA effectively worked for one-word and three-word queries, and
the combined method worked well for two-word queries. Finally, considering
the Broder’s classification, RWEA got the best score for the informational
queries, while the combined method worked best for the navigational queries.
The results show that the best query expansion method depends on the
type of queries. Therefore, we (or system) should select an appropriate ex-
pansion method according to the type of query.
Although we obtained better results for ambiguous or semi-ambiguous
queries with explicit feedback than those returned by search engine “goo”
whose main engine is “Google”, the explicit feedback imposes unpleasant
chore to the user. Therefore the future work is to reduce such chore for the
feedback.
References
1. Broder, A.: A taxonomy of web search. SIGIR Forum 36(2), 3–10 (2002),
http://doi.acm.org/10.1145/792550.792552
2. Chirita, P.A., Firan, C.S., Nejdl, W.: Summarizing local context to personalize
global web search. In: CIKM 2006: Proceedings of the 15th ACM international
conference on Information and knowledge management, pp. 287–296. ACM,
New York (2006), http://doi.acm.org/10.1145/1183614.1183658
3. Chirita, P.A., Firan, C.S., Nejdl, W.: Personalized query expansion for the web.
In: SIGIR 2007: Proceedings of the 30th annual international ACM SIGIR con-
ference on Research and development in information retrieval, pp. 7–14. ACM,
New York (2007), http://doi.acm.org/10.1145/1277741.1277746
4. Chirita, P.A., Nejdl, W., Paiu, R., Kohlschütter, C.: Using odp metadata to
personalize search. In: SIGIR 2005: Proceedings of the 28th annual international
ACM SIGIR conference on Research and development in information retrieval,
pp. 178–185. ACM, New York (2005),
http://doi.acm.org/10.1145/1076034.1076067
48 T. Oishi et al.
1 Introduction
Attention to “prediction markets”, which make predictions of the future us-
ing market mechanisms has been increasing. A prediction market applies
Takuya Yamamoto · Takayuki Ito
Nagoya Institute of Technology, Gokiso, Showa-ku, Nagoya 466-8555, Japan
e-mail: {yamamoto,ito}@itolab.mta.nitech.ac.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 49–62.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
50 T. Yamamoto and T. Ito
techniques of experimental markets that have been used in the field of exper-
imental economics to make predictions.
A prediction market is a market in which participants make trades of
securities predicting the result of a certain event that will be decided in
the future. A security provides a dividend based on the result of the event.
The price of the security serves to predict the event’s realization probability.
Prediction markets include markets of various types.
1. Type1
In a game of baseball, it is assumed that Stock A corresponds to “Win
by Team A” and Stock B corresponds to “Win by Team B”. As a result,
when Team A wins, there is a dividend of 100 cents per share of Stock A,
and Stock B has a dividend of 0 cents.
2. Type2
In an election, Stock A corresponds to “the number seats acquired by
political party A” and Stock B correspond to “the number of seats acquired
by political party B”, and Stock C corresponds to “the number of seats
acquired by political party C”. The price per share of stock X and the
dividend of the number of seats acquired by the political party X (cent)
are determined as the result of an election.
A participant predicts a result from various sources of information related to
the target phenomenon, and he trades to gain profits from his predictions.
The market price of the result unifies all the members’ predictions. When the
market price is Type 1, the price predicts the likelihood that Team A will
win. When the market price is Type 2, it predicts the likely number of seats
that will be acquired by political party X.
Some prediction markets in the U.S. are the Iowa Electronic Market
(IEM)1 and the Hollywood Stock Exchange (HSX)2 . Some of the predic-
tion markets in Japan are at sites such as Shuugi.in3 and Kounaru4 . The
IEM prediction market in the United States has been effective in predicting
election outcomes. It has correctly predicted 75% of the results of elections
traded on its exchange, a success rate that compares favorably with that of
opinion polls. An example of previous research in Japan on prediction mar-
kets was an experiment by Suginoo [10] as to “whether the viewership of the
Winter Sonata of televising would exceed 17% on August 13, 2005”. How-
ever, in this study the problem of reconciling completely different opinions
was not resolved through a prediction market. There were not enough trans-
actions conducted, and poor results in a market, such as described in the
above research, will dampen participants’ motivation.
Thus, in this research, we studied what kind of influence the use of an agent
had on a prediction market. For example, we studied how much influence
1
http://www.biz.uiowa.edu/iem/index.cfm
2
http://www.hsx.com/
3
http://shuugi.in/
4
http://kouna.ru/
Verification of the Effect of Introducing an Agent in a Prediction Market 51
2 Research Background
2.1 Prediction Market
A prediction market is a futures market for carrying out predictions in the
future. The economist Friedrich A. Hayek [4] asserted, “The price in a free
market is determined by the mechanism of communicating information about
the possibility created of a prospective phenomenon”. A prediction market
is something made based on this opinion, and various prediction realization
probabilities are reflected in its market transaction price. The opinion of the
crowd about a phenomenon will be collected in one number in the future
using the mechanism of a futures market and make the number a “predicted
value”. Since the structure of a futures market is used, the predicted value
may change in real time, and may change greatly, depending on external
information such as news events.
If this becomes the case, then, to buy a right for himself as well, Participant A
will to raise the proposed price and compete with participant B as a rival. On
the other hand, if there is a participant C who thinks that the probability of
the Nikkei of reaching 13,000 or more is lower, he will submit a lower bid and
52 T. Yamamoto and T. Ito
the price will fall (Table 1). Thus, according to the expectations (predictions)
of the participants based on various kinds of information they have, deals
will be made on prediction markets, and the price of bids submitted on these
markets will fluctuate (Fig. 1 and Fig. 2).
Verification of the Effect of Introducing an Agent in a Prediction Market 53
3. Decentralization
It is not necessary to abstract a problem, but each participant needs to
judge based on information he or she acquires directly.
Although it is expected that the kinds of information acquired by every
participant will differ, to maintain diversity, one should not judge based
only on attributes common to each participant.
4. Aggregation
All participants share the knowledge they have acquired, taking advantage
of the characteristics of the three above-mentioned points. Moreover, a
structure that carries out comparisons and examinations of participants’
various viewpoints and leads to a final conclusion is required.
Although it is expected that the kinds of information acquired by every
participant will differ, to maintain diversity, one should not judge based
only on attributes common to each participant.
5. Aggregation
All participants share the knowledge they have acquired, taking advantage
of the characteristics of the three above-mentioned points. Moreover, a
structure that carries out comparisons and examinations of participants’
various viewpoints and leads to a final conclusion is required.
A prediction market is equipped with the structure of the four above-
mentioned points. If seen from the viewpoint of utilizing collective intelli-
gence, a prediction market will have a well-made, elegantly simple structure.
actual vote tally percentage. At a stage in early 2004 a value below and above
[$0.51 $0.52] was already being shown, and this value was not very much
different from the end result. This tendency was mostly consistent through
one year, although there were inaccuracies at some points. Although opinion
poll results and analysis by specialists showed significant inaccuracies, most
of the prediction results that prediction markets made did not deviate much
from the actual final result. Of course, such excellent results by prediction
markets are not seen in all elections. However, it can be concluded that
predictions from prediction markets are more accurate than opinion polls
areC at least that was the case for the prediction markets for the November
2004 election.
6
http://zocalo.sourceforge.net/
60 T. Yamamoto and T. Ito
4.2 Experiment
The experiment in this research used Zocalo. The subjects (market partici-
pants) were students from a laboratory, and the number of subjects was 18.
The period of the experiment was one week, from September 7, 2009. An
experiment was conducted in regard to the following two prediction themes.
Theme 2. Will the “Chunichi Dragons” win in the game of “Chunichi Drag-
ons” VS the “Tokyo Yakult Swallows” held on September 13, 2009?
For Theme 1, if the weather is clear, one dollar will be paid out for a ‘yes’
answer to the theme question. For Theme 2, if “Chunichi Dragons” wins, one
dollar will be paid out for a ‘yes’ answer to the theme question.
For Theme 1, this experiment was conducted without an agent, and for
Theme 2, it was conducted with an agent. Table 3 shows an agent’s action.
Agent Action
Buy or Sell Random
Quantity Random in the range of 1-10
Value Random in the range of ±20 from
the latest value
with an agent. For a trade, two persons, a “buyer” and a “seller”, are re-
quired. Since there were few subjects in this experiment, we thought that
there would be few dealings for Theme 1. However, for Theme 2, there was
an increase in the number of dealings due to the introduction of am agent.
Therefore, it can be said that an agent is useful for the improvement of deal-
ings frequency. Moreover, the liquidity of a market can also be assured by an
increase in the number of dealings.
5 Conclusion
In this research, we demonstrated the effect of introducing an agent in a pre-
diction market. From the result of the experiment, we found that introducing
an agent improved the number of dealings. Although we conducted the ex-
periment on only two themes, in subsequent research we must increase the
number of experiments done and carry out comparative studies.
References
1. Berg, J., Forsythe, R., Nelson, F., Rietz, T.: Results from a Dozen Years of
Election Futures Markets Research, vol. 1, pp. 742–751. Elsevier, Amsterdam
(2008)
2. Berg, J., Rietz, T.: Prediction Markets as Decision Support Systems. Informa-
tion Systems Frontiers 5(1), 79–93 (2003)
3. Cowgill, B., Wolfers, J., Zitzewitz, E.: Using Prediction Markets to Track In-
formation Flows: Evidence from Google (2008),
www.bocowgill.com/GooglePredictionMarketPaper.pdf
4. Hayek, F.: The Use of Knowledge in Society. American Economic Re-
view XXXV(4), 519–530 (1945)
62 T. Yamamoto and T. Ito
5. Herek, G.M., Janis, I.L., Huth, P.: Quality of U.S. Decision Making during
the Cuban Missile Crisis: Major Errors in Welch’s Reassessment. Journal of
Conflict Resolution 33, 446–459 (1989)
6. Luckner, S.: Prediction Markets: How Do Incentive Schemes Affect Predic-
tion Accuracy? In: Jennings, N., Kersten, G., Ockenfels, A., Weinhardt, C.
(eds.) Negotiation and Market Engineering, no. 06461 in Dagstuhl Seminar Pro-
ceedings. Internationales Begegnungs- und Forschungszentrum für Informatik
(IBFI), Schloss Dagstuhl, Germany, Dagstuhl, Germany (2007)
7. Manski, C.: Interpreting the Predictions of Prediction Markets. Economics Let-
ters 91(3), 425–429 (2006)
8. Oomukai, K.: Web2.0 and Collective Intelligence. Information Processing Soci-
ety of Japan 47(11), 1214–1221 (2006)
9. Servan-Schreiber, E., Wolfers, J., Pennock, D.M., Galebach, B.: Prediction Mar-
kets: Does Money Matter? Electronic Markets 14(3), 243–251 (2004)
10. Suginoo, T.: A New Prediction Tool “Prediction Market”. Communication In-
quiries 8, 8–13 (2006)
11. Surowiecki, J.: The Wisdom of Crowds. Anchor Books (2005)
12. Wolfers, J., Zitzewitz, E.: Prediction Markets. Journal of Economic Perspec-
tives 18(2), 107–126 (2004)
13. Wolfers, J., Zitzewitz, E.: Interpreting Prediction Market Prices as Probabili-
ties. CEPR Discussion Paper No. 5676 (2006)
14. Yamaguchi, H.: It Looks Back Upon the U.S. Presidential Election (2004),
http://www.h-yamaguchi.net/2004/11/post_31.html
A Cognitive Map Network Model
Yuan Miao
Abstract. This paper presents a cognitive map network model which models
causal systems with interactive cognitive maps. Cognitive map is a family of
cognitive models that have been widely applied in modeling causal knowl-
edge of various systems like gaming and economic systems. Many causal
systems have multiple components which causally evolve concurrently, inter-
acting with each other. Modeling such a system as a whole (cognitive map)
is not an ideal solution. Sometimes it is also not possible as the individual
parties may not want to release their knowledge to other parties or the co-
ordinating component. The cognitive map network model proposed in this
paper represents a causal system as an ecosystem with individual components
modeled as cognitive agents. It is a cognitive map ecosystem whose evolution
is driven by the component cognitive agents. The cognitive ecosystem model
is applied in a toy economic system to illustrate its power in study of hidden
patterns.
1 Introduction
Cognitive map is a family cognitive models [14] representing causal knowl-
edge and facilitating inference accordingly, modeling human cognition and
reasoning. Cognitive Maps (CMs) [1], Fuzzy Cognitive Maps (FCMs) [9] and
Dynamical Cognitive Networks (DCNs) [12] are three related models for cog-
nitive modeling. Cognitive Maps are the first and fundamental models of the
family. CMs have been applied in various application domains, for example
analysis of electrical circuits [19], analysis and extension of graph-theoretic
behavior, and plant control modeling [7]. CMs are easy to use and straight-
forward models; however, they do not compare the strength of different rela-
tionships. Rather, each node simply makes its decision based on the number
Yuan Miao
School of Engineering and Science, Victoria University, Australia
e-mail: Yuan.Miao@vu.edu.au
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 63–72.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
64 Y. Miao
Fig. 8 Market falls in a severe oscillation with agents’ decision on delayed market
move
the active cognitive agents set a gap threshold. If the price is away from the
balance price farther than the gap, then trading decision is made. The rest
is the same as that in Fig. 6. It shows that the market can be pulled back to
the balance area (instead of a balance price) quickly. This pattern also proves
that the setting of a “spread” is important to avoid non-necessary sentiment
in the market, and reduce the sensitivity.
In some cases, a trading decision could not be made immediately following
the market move. Each trading strategy takes time to be implemented. Fig.
8 shows the pattern when the trading decision can only be altered after 2
time slots of the market move. The rest setting is the same as that in Fig. 7.
It shows that the existing market is in a big oscillation, and the pattern is
a composition of a few periodic oscillations. This is not a desirable pattern.
Large oscillation can be damaging to the economy. However, many large
A Cognitive Map Network Model 71
companies could not make swift decisions. Therefore, the speculation should
be allowed and encouraged in the market. So the move can be smoothed out.
The patterns in Fig. 7 and Fig. 8 indicate that speculation should be well
managed: not too less and not too much.
4 Conclusions
Cognitive map is a family of cognitive models that have been widely applied
in modeling causal knowledge. To date, all the applications of cognitive maps
model a whole causal system with a whole cognitive map. As many causal
systems have multiple components which causally evolve concurrently, inter-
acting with each other, this paper proposes a cognitive ecosystem model that
collectively models these components and allows them to interact within a
cognitive ecosystem. When such a system contains a large amount of compo-
nents, the traditional analysis approaches have difficulties to provide reliable
analysis. In the recent economic crisis, many AAA products from a large
number of major banks failed. Those AAA products are supposed to fail
only in a very small probability (e.g. 1 out of a million chances). Therefore,
these banks and products fail in a same crisis should be almost impossible.
That fact shows that the economic models were not reliable. Many economic
systems are so complex that traditional analysis tools have to omit too many
factors to come up with a mathematically manageable model, which becomes
unreliable. Cognitive ecosystem takes a different approach. Each component
in the model is not too complex thus can be modeled with cognitive agents
reasonably well. The whole ecosystem patterns can be revealed with simu-
lations, with affordable computation. Such analysis is also able to give the
probability of each result pattern, which is very useful for pricing and policy
making. Such probability is also likely to be more reliable than those derived
from the existing models. This paper uses a small number of cognitive agents
to illustrate different patterns. However, a large number of agents do not
have essential difference in term of the analysis but with a larger amount of
pattern data to process.
References
1. Axelrod, R.M.: Structure of Decision: The Cognitive Maps of Political Elites.
Princeton University Press, Princeton (1976)
2. Borrie, D., Isnandar, S., Ozveren, C.S.: The Use of Fuzzy Cognitive Agents to
Simulate Trading Patterns within the Liberalised UK Electricity Market. In:
Proceedings of Universities Power Engineering Conference, vol. 3, pp. 1077–
1081 (2006)
3. Cai, Y., Miao, C., Tan, A.H., Shen, Z.: Creating an Immersive Game World
with Evolutionary Fuzzy Cognitive Maps. IEEE Computer Graphics & Appli-
cations 29 (2009)
72 Y. Miao
4. Chen, M.E., Huang, Y.P.: Dynamic Fuzzy Reasoning Model with Fuzzy Cogni-
tive Map in Chinese Chess. In: Proceedings of SICE Joint Symposium of 15th
System Symposium and 10th Knowledge Engineering Symposium, vol. 3, pp.
1353–1357 (1995)
5. Craiger, P., Coovert, M.: Modeling Dynamic Social and Psychological Pro-
cesses with Fuzzy Cognitive Maps. In: Proceedings of IEEE World Congress on
Computational Intelligence, vol. 3, pp. 1873–1877 (1994)
6. Eccles, J., Dickerson, J., Shao, J.: Evolving a Virtual Ecosystem with Genetic
Algorithms. In: Proceedings of 2000 Congress on Evolutionary Computation,
vol. 1, pp. 753–760 (2000)
7. Gotoh, K., Murakami, J., Yamaguchi, T., Yamanaka, Y.: Application of Fuzzy
Cognitive Maps to Supporting for Plant Control. In: Proceedings of SICE Joint
Symposium of 15th System Symposium and 10th Knowledge Engineering Sym-
posium, pp. 99–104 (1989)
8. Hagiwara, M.: Extended Fuzzy Cognitive Maps. In: Proceedings of IEEE In-
ternational Conference on Fuzzy Systems (1992)
9. Kosko, B.: Fuzzy Cognitive Maps. International Journal of Man-Machine Stud-
ies (24), 65–75 (1986)
10. Miao, C.Y., Goh, A., Miao, Y.: A Multi-agent Framework for Collaborative
Reasoning. CRC Press, Boca Raton (2002)
11. Miao, C.Y., et al.: DCM: Dynamical Cognitive Multi-Agent Infrastructure for
Large Decision Support System. International Journal of Fuzzy Systems 5(3),
184–193 (2003)
12. Miao, Y., Liu, Z., Siew, C., Miao, C.: Dynamical Cognitive Network - an Exten-
sion of Fuzzy Cognitive Map. IEEE Transaction on Fuzzy Systems 10, 760–770
(2001)
13. Miao, Y., Liu, Z.Q.: On Causal Inference in Fuzzy Cognitive Maps. IEEE Trans-
actions on Fuzzy Systems 8(1), 107–119 (2000)
14. Miao, Y., Tao, X., Shen, Z., Liu, Z.Q., Miao, C.Y.: Transformation of Cognitive
Maps. IEEE Transaction on Fuzzy Systems (to appear)
15. Parenthoen, M., Tisseau, J., Morineau, T.: Believable Decision for Virtual Ac-
tors. In: Proceedings of 2002 IEEE International Conference on Systems, Man
and Cybernetics, vol. 3, p. 6 (2002)
16. Park, K.S., Kim, S.H.: Fuzzy Cognitive Maps Considering Time Relationships.
International Journal of Human-Computer Studies 42(2), 157–168 (1995)
17. Rashed, T., Weeks, J.: Assessing Vulnerability to Earthquake Hazards through
Spatial Multicriteria Analysis of Urban Areas. International Journal of Geo-
graphical Information Science 17(6), 547–576 (2003)
18. Rodin, V., Querrec, G., Ballet, P., Bataille, F., Desmeulles, G., Abgrall, J.,
Tisseau, J.: Multi-Agents System to Model Cell Signalling by Using Fuzzy
Cognitive Maps- Application to Computer Simulation of Multiple Myeloma.
In: Proceedings of the Ninth IEEE International Conference on Bioinformatics
and Bioengineering, pp. 236–241 (2009)
19. Styblinski, M., Meyer, B.: Fuzzy Cognitive Maps, Signal Flow Graphs, and
Qualitative Circuit Analysis. In: Proceedings of the 2nd IEEE International
Conference on Neural Networks (ICNN 1987), vol. 2, pp. 549–556 (1988)
20. Tsadiras, A.: Comparing the Inference Capabilities of Binary, Trivalent and
Sigmoid Fuzzy Cognitive Maps. Information Sciences 178, 3880–3894 (2008)
21. Zhang, W., Liu, L., Zhu, Y.C.: Using Fuzzy Cognitive Time Maps for Modeling
and Evaluating Trust Dynamics in the Virtual Enterprises. Expert Systems
with Applications 35(4), 1583–1592 (2008)
Coordination Strategies and
Techniques in Distributed Intelligent
Systems – Applications
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 73–93.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
74 A. Boukhtouta et al.
domain which requires the coordination among many entities across the civil-
military boundary, and can benefit from multi-agent coordination techniques.
1 Introduction
It is commonly recognized that many different disciplines can contribute to
a better understanding of the coordination function as a way to build and
provide appropriate tools and adequate approaches for designing complex
organizations and systems. Agents incorporated into these self-regulating en-
tities represent “communities of concurrent processes which, in their interac-
tions, strategies, and competition for resources, behave like whole ecologies”.
We describe in this paper applications requiring coordination or distributed
decision-making. A common characteristic of these applications is the highly
distributed nature of them. These applications can be also seen as challenging
problems for multi-agent coordination (See [20]).
The research community working in the area of Distributed Artificial In-
telligence (DAI) unanimously endorses the idea that coordination - a funda-
mental paradigm - represents a challenging research area. A clear consensus
has thus emerged around a forged, well-articulated, and strongly advocated
common vision that coordination is a central issue to agent-based systems
engineering research (See Bedrouni et al. [2] for more details).
This paper provides an overview of the application domains in which multi-
agent systems have been developed. The inherent distributed nature of these
application domains reveals that benefits in performance or efficiency can be
derived through the coordination between multiple agents. In other words,
the collective behaviour of the system is improved through the coordinated
interaction of their parts. In this paper, we will cover applications from de-
fence, transportation, health care, telecommunication, E-business, emergency
management, and ambient intelligence.
From the defence sector, we will describe how multi-agent systems have
been deployed to support course of action analysis through their ability to
monitor the battlefield environment as depicted in command and control sys-
tems, and additionally by tapping into simulations as ground truth are able
to reason about deviations occurring in the movement of entities represented
in the command and control systems. We also describe a large multinational
experiment in which hundreds of agents were federated in a simulated coali-
tion scenario. From the transportation industry, multi-agent coordination
techniques have been applied to manage and improve the overall efficiency
of incoming aircraft. From the health care industry, distributed multi-agent
system coordination concepts have been prototyped to support the monitor-
ing and treatment of diabetic patients. In the area of communication net-
works, coordination techniques have been applied to diagnose faults in such
networks through agents that are able to communicate and exchange local
information in order to understand non-local problems. In another setting,
Coordination Strategies and Techniques 75
2 Defence
2.1 Control of Swarms of Unmanned Aerial Vehicles
(UAVs)
The ability to coordinate military operations with the aid of information tech-
nology has become an imperative for command and control systems for the
past several years. One of the most important military problems (in vogue)
where coordination mechanisms should be used is in the control of swarms
of UAVs (unmanned aerial vehicles). The UAVs are considered in this case
as highly mobile agents that can be used for performing reconnaissance in
a combat environment. In this case the communication between the UAVs
is too costly because it would reveal their location to the enemy. The UAVs
are sent out to collect information from certain locations and then return.
The information sought by a UAV may depend on other information that
76 A. Boukhtouta et al.
it has collected by another UAV. There are many other cooperative control
problems of the UAVS, we enumerate: cooperative target assignment, coor-
dinated UAVs intercept, path planning, feasible trajectory generation, Etc
(See [1]). A recent study of performance prediction of an unmanned airborne
vehicle multi-agent system has been developed by Lian and Deshmukh [13].
The Markov Decision Processes (MDP) techniques and the Dynamic Pro-
gramming approach have been widely used to solve the problems cited above
(See Goldman and Zilberstein [11]).
3 Transportation
3.1 Air Traffic Flow Management
Air traffic congestion is a world-wide problem leading to delays as well as sub-
sequent monetary loss to the commercial flight industry. The ability to either
manage the flow of traffic more efficiently or improve the infrastructure (e.g.,
adding more runways) to increase capacity are two available mechanisms to
improve the efficiency of commercial aviation. Given the fact that increasing
the capacity through improvements to the infrastructure is a costly and very
time consuming solution, information technologies may be useful in helping
to efficiently manage the traffic flow.
The Optimal Aircraft Sequencing using Intelligent Scheduling (OASIS) [14]
is an intelligent agent-based system which is able to ingest data from Sydney
Airport at the rate of 65 aircraft arriving in 3.5 hours. It is designed to assist
the Flow Director (who is responsible for coordinating the landing sequences
for aircraft) in arranging the sequencing of incoming aircraft through intel-
ligent agents that represent the aircraft and supporting global agents. The
aircraft agents are responsible for predicting aircraft trajectories, monitor-
ing the actual movement versus predicted to understand discrepancies, and
engaging in planning activities. The global agents consist of a coordinator
agent, sequencer agent, trajectory manager agent, wind model agent and
user interface agent. The coordinator agent is the central node for interact-
ing with all of the other agents in the system. The sequencer agents uses
search techniques based on the A* search algorithm to determine the arrival
sequence which minimizes delay and overall cost. The trajectory manager
agent ensures that aircraft maneuvers do not violate statutory separation
requirements. The wind model agent is responsible for computing and pro-
viding the forecast winds based on the reported wind by the aircraft agents.
Lastly, the user interface agent is the mechanism by which the flow director
receives information from the system.
The planner agents accept a schedule time from the sequencing agent and
decide how to execute that plan to meet that schedule, and these agents
must monitor the actual movement against the plan and be able to predict
the trajectory. If there are discrepancies, the aircraft agent must notify the
global agents so that another sequence may be computed. The agents co-
ordination is facilitated through communicate of their activities using the
Procedural Reasoning System (PRS). The PRS system is an agent-based
framework for agents to coordinate their activities with each other based on
the Belief, Desires and Intention (BDI) model. The PRS framework is com-
prised of knowledge bases that contain agent beliefs as represented through
first order logic, agent’s goals or desires, a library of plans and a reasoning
engine. Depending on the agent’s beliefs and desires, the appropriate plans
are triggered for execution, which in turn affect the environment.
Coordination Strategies and Techniques 81
mother, who will take the appropriate action. The IMAS system was imple-
mented using the Java Agent Development Environment (JADE) and uses an
ontology and uses the Foundation for Intelligent Physical Agents (FIPA) com-
pliant messages for communication, and more specifically, the FIPA Agent
Communication Language (ACL). The ACL is based on speech act theory,
which is based on the fact that when one makes a statement, it is implied
that one is doing something.
Future areas for investigation include techniques to ensure the privacy of
the information so that only the right individuals or agents have access, and
also the necessity to properly deal with the security of the information that
is being transmitted so that it cannot be altered by third parties.
5 Communications Networks
capabilities based on information obtained from the data store. The expert
system in DIMAS is built on the AT&T C5 language, and its workflow pro-
cesses through a series of steps: monitor symptoms, reply to queries, do ac-
tions, analyze results, analyze evidence and decide next step.
The authors also differentiate the DIMAS system with other related work.
For example, they compare DIMAS to Distributed Big Brother (DBB). The
DBB uses the Contract Net Protocol to distribute the monitoring task to
Local Area network managers. However, within DIMAS, rather than offering
tasks to potential bidders and then awarding the task to one bidder, the DI-
MAS agents are asked whether they can suggest possible actions to diagnose
a problem. The suggested actions are then evaluated by the temporary owner
of a problem and then contracted out to the agent that suggested the action.
For example, each agent that detects a symptom of a problem may suggest
further diagnostic tests, such as checking for changes to base station param-
eters or monitoring the signal strength of call attempts. However, for each
action there is only one agent that can execute it. The decision that must be
made by the coordinating agent is which action to choose, rather than which
agent will execute that action. The subtle difference is that in DIMAS, the
contracting is action-centric and not agent-centric, as a series of actions may
be spread over multiple agents.
The Large-internetwork Observation and Diagnosis Expert System
(LODES) is an expert system for diagnosing faults in a LAN that consist
of multiple networks connected via routers. Each LAN has its own LODES
system, which is used to detect remote faults based on local data. The pri-
mary difference between LODES and DIMAS is that the latter is a passive
approach while the former is both passive as well as active.
In DIMAS, the agents can only observe local symptoms and/or faults. Co-
ordination between the agents is facilitated through the use of communication
using KQML performatives such as ask, tell and achieve to gather non-local
information. These performatives are used in various phases of the agent’s
process such as assign-responsibility, gather-evidence and allocate-blame. In
the assign-responsibility phase, the problem is detected using the ask and
tell performatives. The base station at which the symptom is detected uses
the ask performative to query the other base station agents about possible
causes for increase in calls with poor transmission quality, while the other
base stations use the tell performative to report back as to the fact that the
cause may be due to a misaligned antenna or a frequency change. The affected
base station agent then uses the achieve performative to ask the other base
station agents to check for changes in frequency within their logs (gather-
evidence). When this is ruled out, then more expensive requests are made
such as checking for conditions which are indicative of a misaligned antenna.
The allocate-blame phase assigns responsibility to the offending base station,
and communicates its findings to other base stations.
Coordination Strategies and Techniques 85
6 E-Business
6.1 Supply Chain Management
A supply chain is described in [5] as “a network of suppliers, factories, ware-
houses, distribution centers and retailers through which raw materials are
acquired, transformed, reproduced and delivered to the customer”. A Sup-
ply Chain Management System (SCMS) manages the interaction of these
processes in order to achieve some end goals.
In [5] the authors describe a SCMS based on the coordination of multi-
agent systems, as represented through information agents and functional
agents. The information agents are a source of information to the functional
agents, while the functional agents are responsible for specific roles in the
dynamic supply chain. The key notion behind these agents is that there is no
assumption that a certain number of agents be available in the supply chain.
The agents are assumed to enter and leave the system through a negotiation
process based on the negotiation performatives that have been developed
based on the FIPA ACL. For example, accept-proposal, CFP, proposal, reject-
proposal and terminate are used for pairwise negotiation between agents while
a third performative, bid, has also been included to capture third party nego-
tiations, for example, through an auctioneer. The paper [5], further describes
the use of Colored Petri Nets (CPN) as a modeling tool for managing the
negotiation process between agents. A simple example of the interaction be-
tween buyer and seller agents is described through the use of CPNs. The
buyer agent initiates the negotiation process by sending a message indicating
a call-for-proposal from the seller agent. The buyer agent then waits, while
the seller agent receives the message for consideration. When the seller agent
sends the reply (proposal, accept-proposal, reject-proposal or terminate) to
the buyer agent, the buyer agent then enters the waiting or thinking/inactive
state. If the thinking transition is triggered, the cycle repeats and it is up to
the buyer agent to send the next round of message. The paper also describes
the internal logic of the functional agents in making decisions by providing
an example of an agent that represents a company’s need for certain quantity
Coordination Strategies and Techniques 87
in the ability to coordinate such a large group across the civil-military bound-
ary during SSTR operations. Usually, in the civil sector, coordination is the
result of voluntary effort; hence coordination by “directing” is rarely effective.
Generally, relief agencies partly function within a framework of self-interest.
In other words, they assist their beneficiaries in such a way that their good
works are seen and valued by the donor community and the “profile” of their
agency is enhanced. In many cases, farther down on the list is the goal of rec-
ognizing the contribution of others or admitting someone else can do the job
better and therefore coordination is not necessarily an agencies first priority.
With regard to coordination across the civil-military boundary there are also
a number of challenges. The military tends to be a highly structured orga-
nization with a hierarchical command and control structure, while the civil
sector in many cases is more loosely structured and tend to be less formal.
Understanding emerging social networks and using the information that
is provided through such networks is important for effective coordination.
Social networks provide a visual cue as to the relationships between people
or groups and can be used to identify “mutual” friends. This is particularly
important when there is a requirement by one or more individual or groups
to locate other individuals or groups that might be able to provide support
in order to achieve a desired end result. Research and tools from the Social
Network Analysis (SNA) community may allow users to understand who the
experts are in the SSTR community and to whom and how they are linked.
As a simple example, social maps that depict connectivity between users in
the context of their discussion threads, and ability to filter the content within
a social map based on specific keywords are likely to provide the foundation
to enable the community to identify service providers or those that may of-
fer similar services or capabilities. The ability to rate individuals within the
social network may also be an important aspect in building trust within the
community of users. This is particularly important during pre-disaster situa-
tions so that some level of trust and common understanding can be achieved.
Furthermore, pre-disaster interactions can help in the development of con-
cept of operations or doctrine to provide guidance during real life situations
by helping people or organizations form the bonds of working together. One
of the challenges, however, will be to effectively visualize such a network or
efficiently filter through the various dimensions of information contained in
the social network. There is also a lack of automated coordination tools to
support the ability of people or groups to actively coordinate their activities.
There are processes in place but most coordination is manual. The ability
to manage tasks across the diverse actors involved in these operations has
the opportunity to improve the overall efficiency of these operations. Due to
the complexity of SSTR operations, multi-agent coordination techniques may
provide benefits.
There is already ongoing research into developing information sharing en-
vironments to be used during SSTR operations within which automated co-
ordination tools could be developed. These environments leverage content
Coordination Strategies and Techniques 89
dĂƐŬEĞƚǁŽƌŬ
dĂƐŬEĂŵĞ
• >ŽĐĂƚĞ'ƌŽƵŶĚdƌĂŶƐƉŽƌƚĂƚŝŽŶ
WƌŽǀŝĚĞĨŽŽĚƚŽ'ƌŽƵƉͲŝŶƌĞŐŝŽŶͲdž ;ƚƌƵĐŬͿŝŶZĞŐŝŽŶͲdž
>ŽĐĂƚĞ WƌĞĐŽŶĚŝƚŝŽŶƐ
'ƌŽƵŶĚ • ƌŝǀĞƌŬŶŽǁƐƌŽƵƚĞ͍
dƌĂŶƐƉŽƌƚĂƚŝŽŶ • ŽƐƚƐĂƌĞďĞůŽǁΨ
^ƵďƚĂƐŬƐ
;ƚƌƵĐŬͿŝŶ
•ƌŝǀĞƚƌƵĐŬƚŽ&ŽŽĚĚŝƐƚƌŝďƵƚŝŽŶ
ZĞŐŝŽŶͲdž
/ĚĞŶƚŝĨLJƚLJƉĞŽĨĨŽŽĚ ĞŶƚĞƌ
ƚĚŝƐƚƌŝďƵƚŝŽŶ • >ŽĂĚ&ŽŽĚŝŶƚŽdƌƵĐŬ
ĞŶƚĞƌ •ĞůŝǀĞƌĨŽŽĚƚŽŐƌŽƵƉͲ
8 Ambient Intelligence
The early developments in Ambient Intelligence took place at Philips in 1998.
The Philips vision of Ambient Intelligence is: “people living easily in digital
environments in which the electronics are sensitive to people’s needs, person-
alized to their requirements, anticipatory of their behavior and responsive
to their presence”. The main purpose of Ambient Intelligence applications is
to coordinate the services offered by small smart devices spread in a phys-
ical environment in order to have a global coherent intelligent behaviour of
this environment. From a computing perspective, ambient intelligence refers
to electronic environment that is sensitive and responsive to the presence of
people. Ambient intelligence paradigm builds upon several computing areas
and human-centric computer interaction design. The first area is ubiquitous
or pervasive computing. Pervasive computing devices are very tiny (can be in-
visible) devices, either mobile or embedded in almost any type of object imag-
inable, including cars, tools, clothing and various consumer goods. All these
devices communicate through increasingly interconnected networks (ad hoc
networking capabilities). The second key area is intelligent systems research.
Learning algorithms and pattern matchers, speech recognition and language
translators, and gesture classification belong to this area of research. A third
area is context awareness; research on this area lets us track and position ob-
jects of all types and represent objects’ interactions with their environments.
See [15] and [23] for more details about Ambient Intelligence applications.
9 Conclusion
Information technologies and information systems have become the corner-
stone that shapes the present and the future of our society and its underlying
affairs. They have penetrated almost every aspect of our lives. This omnipres-
ence is stemming from significant advancements in computer systems, soft-
ware, middleware, telecommunications infrastructure, etc. Also, enterprises,
organizations and governments usually use a large variety of networked in-
formation systems, interconnected software, etc. From other side, military
92 A. Boukhtouta et al.
References
1. Beard, R.W., McLain, T.W., Goodrich, M.A., Anderson, E.P.: Coordinated
Target Assignment and Intercept for Unmanned Air Vehicles. IEEE Transac-
tions on Robotics and Automation 18(6), 911–922 (2002)
2. Bedrouni, A., Mittu, R., Boukhtouta, A., Berger, J.: Distributed Intelligent
Systems: A coordination Perspective. Springer, New York (2009)
3. Carey, S., Kleiner, M., Heib, M., Brown, R.: Standardizing Battle Management
Language- Facilitating Coalition Interoperability. In: Proceedings of the Euro
Simulation Interoperability Workshop, SIW (2002)
4. Chang, Y.H., Ho, T., Kaelbling, L.: Multi-agent learning in mobilized ad-hoc
networks. In: Proccedings of AAAI Fall Symposium, Technical Report FS-04-02
(2004)
5. Chen, Y., Peng, Y., Finin, T., Labrou, Y., Cost, S., Chu, B., Yao, J., Sun,
R., Wilhelm, B.: A Negotiation-based Multi-agent System for Supply Chain
Management. In: Workshop on Agents for Electronic Commerce and Managing
the Internet-Enabled Supply Chain, Seattle, WA (1999)
6. Dahlburg, J.: Developing a Viable Approach for Effective Tiered Systems. NRL
Memorandum Report 1001-07-9024. USA (2007)
7. Dresner, K., Stone, P.: A Multiagent Approach to Autonomous Intersection
Management. Journal of Artificial Intelligence Research 31, 591–656 (2008)
8. Feltovich, P., Bradshaw, J.M., Jeffers, R., Uszok, A.: Order and KAoS: Us-
ing policy to represent agent cultures. In: Proceedings of the AAMAS 2003
Workshop on Humans and Multi-Agent Systems, Melbourne, Australia (2003)
9. Fischer: Cooperative Transportation Scheduling: An application Domain for
DAI. Applied Artificial Intelligence 10(1), 1–34 (1996)
10. Gibney, M.A., Jennings, N.R., Vriend, N.J., Griffiths, J.M.: Intelligent Agents
for Telecommunication Applications. In: Proceedings of the 3rd International
Workshop IATA (1999)
11. Goldman, C.V., Zilberstein, S.: Decentralized Control of Cooperative Systems:
Categorization and Complexity Analysis. Journal of Artificial Intelligence Re-
search 22, 143–174 (2004)
Coordination Strategies and Techniques 93
12. Leckie, C., Senjen, R., Ward, B., Zhao, M.: Communication and Coordination
for Intelligent Fault Diagnosis Agents. In: Proceedings of the Eighth IFIP/IEEE
International Workshop DSOM 1997, Sydney, Australia, pp. 280–291 (1997)
13. Lian, Z., Deshmukh, M.A.: Performance prediction of an unmanned airborne
vehicle multi-agent system. European Journal of Operational Research 172,
680–695 (2006)
14. Ljungberg, M., Lucas, A.: The OASIS Air Traffic Management System. In:
Proceedings of the Second Pacific Rim International Conference on Artificial
Intelligence, PRICAI 1992 (1992)
15. McCullough, M.: Digital Ground: Architecture, Pervasive Computing, and En-
vironmental Knowing. MIT Press, Cambridge (2005)
16. Min, J.U., Bjornsson, H.: Construction Supply Chain Visualization through
Web Services Integration, Technical Report, Civil & Environmental Engineer-
ing Dept., Stanford University (2004)
17. Mittu, R., Abramson, M.: Walters. J.: Improving Simulation Analysis through
C4I Systems and Use of Intelligent Agents. In: Proceedings of the Spring Sim-
ulation Interoperability Workshop, SIW (2004)
18. Mittu, R., Segaria, F., Guleyupoglu, S., Barber, S., Graser, T., Ross, R.: Sup-
porting the Coalition Agents Experiment (CoAX) through the Technology In-
tegration Experiment (TIE) Process. In: Proceedings of the 8th International
Command and Control Research and Technology Symposium (ICCRTS), July
17-19, National Defense University, Washington (2003)
19. Nunes, L., Oliveira, E.: Learning from multiple sources. In: AAMAS Proceed-
ings of the Third International Joint Conference on Autonomous Agents and
Multiagent Systems, New York, NY, USA, pp. 1106–1113 (2004)
20. Panait, L., Luke, S.: Cooperative Multi-Agent Learning: The State of the Art.
Autonomous Agents and Multi-Agent Systems 11, 387–434 (2005)
21. Peng, Z., Guohua, B., Bengt, C., Stefan, J.J.: Applying Multi-agent Systems
Coordination to the Diabetic Healthcare Collaboration. In: Proc. The Int.
Conference on Information & Communication Technologies: from Theory to
Applications- ICTTA, Damascus, Syria (2008)
22. Steeb, R., Cammarata, S., Hayes-Roth, F., Thorndyke, P., Wesson, R.: Dis-
tributed intelligence for air traffic control. In: Bond, A., Gasser, L. (eds.)
Readings in Distributed Artificial Intelligence, pp. 90–101. Morgan Kaufmann
Publishers, San Francisco (1988)
23. Wang, X., Dong, J.S., Chin, C., Hettiarachchi, S.R., Zhang, D.: Semantic Space,
A Semantic Web Infrastructure for Smart Spaces. IEEE Pervasive Comput-
ing 3(3), 32–39 (2004)
Multi-Agent Area Coverage Using a
Single Query Roadmap: A Swarm
Intelligence Approach
1 Introduction
The area coverage problem in robotics deals with the use of one or more
robots to physically sweep [8] or visually sense [5]the free space of an area.
The aim of this research study is to propose a robust mechanism to cope with
the problem of multi-agent visual environment coverage.
Ali Nasri Nazif · Alireza Davoodi · Philippe Pasquier
Metacreation, Agents and Multiagent Systems (MAMAS) Group, School of
Interactive Art and Technology, Faculty of Communication, Arts and Technology,
250 - 13450 102nd Avenue, Surrey, BC, Canada, V3T 0A3
e-mail: anasri@aut.ac.ir,{alireza_davoodi,pasquier}@sfu.ca
http://www.siat.sfu.ca
∗
The first and second authors have equally contributed in the paper.
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 95–112.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
96 A.N. Nazif, A. Davoodi, and P. Pasquier
Fig. 1 This figure illustrates how a RRT tree is expanded given the current state
of the tree.
qrand , will be placed randomly in the free space and its nearest neighbor node,
qnear , among previously added nodes will be selected for further expansion.
A new node, qnew , is produced as a result of moving qnear by , a predefined
value called step size, toward qrand and it will be added to the roadmap
if collision-free. Figure. 1 illustrates a snapshot of the expansion process.
assign exploration targets between the agents. Furthermore, Choset [4] sur-
veys different approaches such as approximation, grid-based and decomposi-
tion approaches. In one of the recent attempt Packer [12] applies a compu-
tational geometry approach in order to compute multiple Watchman routes
which can cover inside a given polygon. In this paper, she first solves the art
gallery problem and finds a set of guards which can collectively cover the
entire area. Then, the visibility graph of the obstacles and static guards is
constructed and decomposed into several routes. Finally, each route is allo-
cated to one agent to traverse.
Grid-based approaches are also used in environment coverage. Hazon and
Kaminka [8] propose a grid-based mechanism in which the environment is
decomposed into several same size cells on the basis of the size of the robots.
Considering the initial positions of the robots, they find a spanning tree of
the free cells which can be decomposed to some balanced sub-trees. Finally
each sub-tree is assigned to a robot.
Fig. 2 Illustrates a situation in which eight agents (small white polygons) are
available. The figure on the left represents the polygonal boundary enclosing the
agents, the diameters and the Main Root (the black circle). The figure on the right
illustrates the corresponding Secondary Roots
edges, which connect the agents initial locations to the Main Root. A set of
candidate nodes is produced as a result of moving out from the Main Root
by the value of towards the agents’ locations along the edges. The value of
is specified by the user and must be equal or less than the value of the step
size as defined in the previous section. If a candidate node is collision-free,
it is called a Secondary Node and assigned to the corresponding agent
whose location point is an end-point of the edge. If the candidate node is not
collision-free, the algorithm finds the nearest agent which has been assigned a
Secondary Node and assign it to the agent. Figure. 2 illustrates the Main
Root and Secondary Roots of a particular configuration.
From this point on, the algorithm is similar to the RRT planner algo-
rithm with two differences. The first one is in the number of initial points,
qinit . While RRT there is only one qinit node, which is randomly located
in the free space, the WMA-RRT includes as many qinit nodes as there are
Secondary Roots. The other difference is observed in the initial expan-
sion of the roadmap. In RRT the roadmap expands through the root of the
roadmap, qinit , while in WMA-RRT the Main Root is exempted from expan-
sion and instead, the Secondary Roots are considered for expansion. The
initial roadmap expands very rapidly throughout the free space of the envi-
ronment and yields a tree with one Main Root, which has as many sub-trees
as there are Secondary Roots. Each sub tree is called a Branch.
The edges’ weight of the roadmap is initialized to 0 so the roadmap is con-
sidered as a weighted tree. During the covering operation, at-least one agent is
assigned to each Branch. The weights of the edges are updated while travers-
ing the roadmap. The agents use WMA-RRT roadmap to move throughout the
free space, interact with the environment and indirectly communicate with
one another. In the other words, while traversing the WMA-RRT roadmap, the
agents perceive information available through the roadmap and update their
knowledge about the environment. Further, the agents change the state of
the environment by depositing information into the nodes of the WMA-RRT.
Multi-Agent Area Coverage Using a Single Query Roadmap 101
Table 1 Represents all the information available in the nodes of the WMA-RRT
roadmap.
Fig. 3 A cluttered environment with narrow passages. Two agents have been as-
signed to each branch.
102 A.N. Nazif, A. Davoodi, and P. Pasquier
Fig. 4 The figures on the left and right represent the results of running PRM and
WMA-RRT algorithms respectively on an environment.
4 Agent Architecture
As mentioned in previous sections, the agents in our system are simple
autonomous individuals capable of following some explicit condition-action
rules and performing the corresponding actions. To this end, our agents are
modeled based on reactive architecture and are implemented in such a way
that supports beliefs, actions and Plans of the agents. Furthermore,
they are utility based entities, which try to maximize their own utilities
independently. The utility function of an agent is defined as the average
over the weight of all the edges, the agent traverses. The utility function of
an agent motivates it to select the edges with minimum weight.
Additionally, the agents perceptions are represented as beliefs in the
agents’ belief bases and each agent contains some pre-defined plans stored
in its plan library. Each plan is an instruction which tells the agent what to
do in a particular circumstance. Each agent has a goal and uses a plan to
approach the goal. In the rest of this section, we discuss some basic beliefs,
actions and plans used by the agents during the covering operation. We use
Agent Speak, a multi-agent programming languages standard syntax and for-
mat in order to explain the agents beliefs, actions and plans. As discussed
in previous section, the agents update their knowledge base about the world
when they reach a node and perceive the information available at that partic-
ular node. In the other words, while traversing the WMA-RRT tree, an agent
detects the information available in the nodes of the roadmap and updates
its belief base accordingly. The followings are some beliefs of the agents:
Multi-Agent Area Coverage Using a Single Query Roadmap 103
that the maximum length of each edge of the WMA-RRT equals to the step
size, ,which has been already defined in WMA-RRT construction algorithm.
7. detect: The agents use their sensors in order to percept all the informa-
tion available in the current node. Standing at a node, an agent also can
detect the outgoing edges of the current node.
8. scan: Standing at a node, each agent can access a local roadmap con-
taining the current node, incoming edge and outgoing edges of the current
node. The agents cannot store these local roadmaps in their limited mem-
ories but an agent can temporarily employ the local roadmap in order to
enumerate the outgoing edges and assign them temporary labels.
9. enumerate: Standing at a node, an agent changes the state of the current
node by updating information relevant to the incoming and outgoing edges
of the current node. To differentiate the outgoing edges, the agents locally
enumerate them and assign each of them a label. To this end, all the agents
assume that, the labels of all the edges going out of the Main Root, are
zero. As mentioned in previous sections, the roadmap is locally available
to the agents so they can detect all the outgoing edges and label them by
enumrating all the outgoing edges in a clock-wise order starting from the
incoming edge and incrementally assigning each of them a label.To this
end, the algorithm assums that the label of all the Secondary Roots
are zero. In other words, suppose that the weight of the incoming edge of
the current node is w and there are two outgoing edges. Then, by doing
a clock-wise search starting from the incoming edge, the label of the first
and second edges are w + 1 and w + 2 respectively. Figure 5, illustrates
how the edges of WMA-RRT are locally enumerated by the agents.
Fig. 5 This figure represents how the edges of the WMA-RRT are locally enumer-
ated and labeled by the agents. Here, the biggest black circle is the Main Root and
two second biggest are the Secondary Roots.
Multi-Agent Area Coverage Using a Single Query Roadmap 105
3. traversing: This plan is invoked when the agent has made a decision
on which outgoing edge to traverse. This plan makes the agent lock the
selected edge, depart from the current node, traverse the edge and reach
the other endpoint of the selected edge.
+!traversing(Edge) : true <-
departNode;
lock(Edge);
traverse(Edge);
reachNode.
4. return: This plan is used when an agent reaches a leaf or blind node,
a node whose sub-trees have been traversed and set as completed, so the
agent cannot move forward and has to return one level back to the parent
of the current node.
+!return(Node) : leaf(Node) | blind(Node) <-
internalAction.nextEdge(Edge);
comeback(Node);
traverse(Edge);
detect.
Secondary Root. The agents then update the state of the environment
accordingly by changing the states of the nodes.
At each node, the agent executes the nextDestination plan to find the
next edge to traverse. Basically, The outgoing edge with minimum weight
is selected for exploration since edges with higher weights have been already
visited more so it is reasonable to select those edges with fewer weights. Then
the agent performs the trversing plan to update the state of the start node
of the edge, traverse the edge and update the state of the end node of the
edge. If an agent reaches a leaf or blind node of the current branch, it applies
the return plan in order to return to the parent node and the leaf or blind
node is marked as completed. An agent sets a node as completed if and
only if the node’s entire sub-trees have been marked as completed. When an
agent is going to perfom the return plan, it should know which edge is the
incomming edge. Since the states of the incoming edge and outgoing edges
of the current node are available, the agent can figure out which edge is the
incoming edge since the incoming edge is the only edge which has not been
marked as completed.
Eventually, when an agent returns to its corresponding Secondary Root,
it marks the Secondary Root as completed and returns one level back to
the Main Root to contribute in exploration of other branches if there is
any remaining. As soon as an agent realizes that all the branches have been
set as completed, it marks the Main Root as completed to indicate that the
exploration process is done.
It is very likely that one agent completes the navigation of its correspond-
ing branch and returns to the origin while some other agents are still work-
ing in other branches. In such situations, an agent manages to maximize its
contribution by helping some other agents in their exploration missions. As
an agent returns to its corresponding Secondary Root and consequently,
reaches the Main Root, it applied its findingBranch plan to find the branch
in which the minimum number of agents are working in and commit to explor-
ing it. Consequently, the agent could help other agents working in the same
branch to finish the exploration task. When an agent arrives in a node, which
has been marked as completed, it does not go further and therefore, returns
to the parent node by executing the return plan. This process guarantees
the exploration of the environment even in the case of occurring massive
failures in the agents. This feature makes our system reliable and robust.
Fig. 6 Shows the results of running the WMA-RRT tree and the coverage mechanism
in the presence of three agents.
Multi-Agent Area Coverage Using a Single Query Roadmap 109
Table 2 Represents all the information available in the nodes of the WMA-RRT
roadmap.
Table 3 Represents all the information available in the nodes of the WMA-RRT
roadmap.
References
1. Abraham, A., Guo, H., Liu, H.: Swarm intelligence: Foundations, perspectives
and applications. In: Swarm Intelligent Systems. Studies in Computational In-
telligence, pp. 3–25. Springer, Heidelberg (2006)
2. Bayazit, O.B.: Roadmap-based flocking for complex environments. In: Proc.
10th Pacific Conference on Computer Graphics and Applications, PG 2002,
pp. 104–113 (2004)
3. Beni, G., Wang, J.: Swarm intelligence. In: Proceedings Seventh Annual Meet-
ing of the Robotics Society of Japan, pp. 425–428. RSJ Press, Tokyo (1989)
4. Choset, H.: Coverage for robotics - a survey of recent results. Ann. Math. Artif.
Intell. 31(1-4), 113–126 (2001)
5. Davoodi, A., Fazli, P., Pasquier, P., Mackworth, A.K.: On multi-robot area
coverage. In: 7th Japan Conference on Computational Geometry and Graphs,
JCCGG 2009, pp. 75–76 (2009)
6. Gordon, N., Wagner, I.A., Bruckstein, A.M.: Discrete bee dance algorithm for
pattern formation on a grid. In: IEEE International Conference on Intelligents
Agents Technologies, Toronto, CA, pp. 545–549 (2003)
7. Halász, Á.M., Hsieh, M.A., Berman, S., Kumar, V.: Dynamic redistribution
of a swarm of robots among multiple sites. In: International Conference on
Intelligent Robots and Systems, IROS, pp. 2320–2325 (2007)
8. Hazon, N., Kaminka, G.: On redundancy, efficiency, and robustness in coverage
for multiple robots. Robotics and Autonomous Systems (2008)
9. Kavraki, L.E., Svestka, P., Latombe, J.C., Overmars, M.H.: Probabilistic
roadmaps for path planning in high-dimensional configuration spaces, vol. 12,
pp. 566–580 (1996)
10. Lavalle, S.M., Kuffner, J.J.: Rapidly-exploring random trees: Progress and
prospects. In: 4th Workshop on the Algorithmic Foundations of Robotics, Al-
gorithmic and Computational Robotics: New Directions, pp. 293–308 (2000)
11. Oks, E., Sharir, M.: Minkowski sums of monotone and general simple polygons.
Discrete & Computational Geometry 35(2), 223–240 (2006)
12. Packer, E.: Computing multiple watchman routes. In: McGeoch, C.C. (ed.)
WEA 2008. LNCS, vol. 5038, pp. 114–128. Springer, Heidelberg (2008)
13. Preparata, F.P., Hong, S.J.: Convex hulls of finite sets of poin ts in two and
three dimensions. Commun. ACM 20(2), 87–93 (1977)
14. Reynolds, C.: Steering behaviors for autonomous characters. In: Game Devel-
opers Conference (1999)
15. Sahin, E., Labella, T.H., Trianni, V., Louis Deneubourg, J., Rasse, P., Floreano,
D., Gambardella, L., Mondada, F., Nolfi, S., Dorigo, M.: Swarm-bot: Pattern
formation in a swarm of self-assembling mobile robots. In: Proceedings of the
IEEE International Conference on Systems, Man and Cybernetics, Hammamet,
pp. 6–9. IEEE Press, Los Alamitos (2002)
16. Svennebring, J., Koenig, S.: Trail-laying robots for robust terrain coverage. In:
ICRA, pp. 75–82 (2003)
17. Tang, L.J.: A visibility-based algorithm for multi-robot boundary coverage.
International Journal of Advanced Robotic Systems 5, 63–68 (2008)
112 A.N. Nazif, A. Davoodi, and P. Pasquier
18. Tarasewich, P., McMullen, P.R.: Swarm intelligence: power in numbers. Com-
mun. ACM 45(8), 62–67 (2002)
19. Wurm, K.M., Stachniss, C., Burgard, W.: Coordinated multi-robot exploration
using a segmentation of the environment (2008)
20. Zlot, R., Stentz, A., tony Stentz, A., Dias, M.B., Thayer, S.: Multi-robot ex-
ploration controlled by a market economy. pp. 3016–3023 (2002)
An Approach for Learnable
Context-Awareness System by
Reflecting User Feedback
1 Introduction
Recently, as the ubiquitous computing becomes popular, context awareness
computing is also becoming interesting research issue. The context in context
awareness system is any entities that affect interaction between human and
computing devices in the environment. An entity can be a human, place, or
any object relevant to the interaction between human and application [1] [14].
Among the many research results on this issue, the most prominent sys-
tems are the CoBrA (Context Broker Architecture) [2], SOCAM (Service
Oriented Context-Aware Middleware) [3], Context-Toolkit [4]. Some of the
InWoo Jang · Chong-Woo Woo
School of Computer Science, Kookmin University
e-mail: zx1018@nate.com,cwwoo@kookmin.ac.kr
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 113–127.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
114 I. Jang and C.-W. Woo
systems were not flexible for providing formalized expression and also for the
extensibility and interoperability with other systems. To handle this issue,
CoBrA and SOCAM used ontological model that supports easy sharing in
different domains, and reusing the related knowledge. Therefore, the ontolog-
ical modeling approach becomes popular solution to represent the data in the
recent context awareness systems. But, when we try to build the system based
on the ontological modeling, we still have some problems as follows. First,
we often run into unreliable results that come from converting numerically
sensed data into symbolic expression. Second, the system always responds
with the same context to different users. And this could be much effective if
we could support personalized solution based on the each individuals prefer-
ence. And finally, if the system could learn the users behavior, then it will
perform much adaptively by providing a personalized service to the user.
In this paper, we are describing a design and development of agent based
context awareness system that can handle such limitations of the current con-
text awareness system. Our system is designed as three main layers, namely,
hardware layer, middleware layer, and application layer. The hardware layer
receives numerical raw data and makes data conversion using Fuzzy infer-
ence mechanism. Middleware layer takes care of ontology modeling, which
has been developed but will not be discussed in this paper. The applica-
tion layer makes inferences using case-based reasoning (CBR) engine, and
generates final solution to the user. Our approach focuses on the next two
main issues; building ontology based context modeling using fuzzy data, and
second, reusing adapted rules by interacting with users feedback. The sim-
ulated experimentation has been done, and results show some significant
importance.
2 Related Studies
2.1 Context-Awareness System
Recently, the context awareness computing has been studied explosively, and
CoBrA [2], SOCAM [3], Context-Toolkit [4][6], GAIA [5][9], CoolTown [7], are
the major research results in this area. These studies showed advent of some
context models for representing a context, such as, Attribute-Value, Web
Based model. And later, this has been developed into Entity-Relationship
model, UML model, and ontology model. This development is due to the
need for sharing not only the context information, but also the need for in-
ference of the users intention. Therefore, the context awareness computing
can be implemented in many different ways, including more formalized repre-
sentation of the context information, and the use of ontology becomes main
stream in this paradigm. The general process of the system that based on
the ontology model includes next few steps [2][3][5][9].
An Approach for Learnable Context-Awareness System 115
- First, the system collects numerical raw data from the physical sensors in
the distributed environments. Since the ubiquitous sensor network has been
built successfully, we could receive information from the sensors through the
network.
- Second, the collected data from the sensors needs to be further processed
to become ontology data that requires semantic value. For instance, if mea-
sured illumination value is 1000 lux, then this value must be converted into
semantic data, bright, and saved on the ontology model. This numeric value is
called low-level context and the semantic value is called as high-level context.
- Third, the data inference mechanism runs into some complicated onto-
logical process. For instance, if the processed data from the above stage is
room, then we need to get related context from the ontology model, and home
is the upper class from the ontology model.
- Fourth, we then need to generate appropriate solutions for the context
by applying domain specific rules. For example, ’if Tony comes into a room,
then turn on the light’ is the final solution for the current situation.
This is the most general procedure for the recent context awareness system,
which has strong advantages, such as, reusing data, sharing data, interoper-
ability, and so on. But the system using the ontology, still needs to be studied
further for its evaluation on the provided services. And also the sensed data
needs to be investigated further for its reliability, which will be discussed in
the next section.
3 System Architecture
The main structure of our system consists of three layers as in the Fig. 1,
namely, hardware layer, middleware layer, and application layer. And each
layer of the system works as follows.
116 I. Jang and C.-W. Woo
- First, the hardware layer is organized as physical sensor and software Con-
text Management Agent (CMA). The physical sensor receives raw data from
the environment, processes through the CMA, and sends the processed data
to the middleware of the system. To make this work, the CMA controls the
physical sensor and converts the raw data into the semantic data through the
fuzzy reasoning engine.
- Second, the middleware layer is in charge of sharing data through ontol-
ogy model, resource managing, and managing data input and output. Basi-
cally, it connects hardware layer and application layer through data I/O, and
generates intermediate data.
- Third, the application layer is composed of Situation Reasoning Agent
(SRA), and some computing devices for the user. This layer receives context
from the middleware, analyzes it, and provides solution to the user. The SRA
recognizes the context through the knowledge base and CBR engine, and
generates solution for the situation.
In this paper, we are not going to describe middleware layer, since it is out
of focus in this paper. The detailed descriptions for the agents of the system
are in the below.
fuzzy rules are saved in the internal knowledge base, and there are some
examples in the Table 1 below.
Index Rules
Rule1 IF x is A1 and y is B1 Then z is C1
Rule2 IF x is A2 and y is B2 Then z is C2
... ...
From the above table, the A1 stands for very dry, A2 is dry, B1 is warm,
and B2 is very warm. With this Fuzzy rules, the CMA generates dependent
context. Lets assume that fuzzy inference engine generates the discomfort
index by using the sensed humidity and temperature. Then the engine goes
through the next four steps as follows.
First, we seek for the suitability for the rules. This can be done by decid-
ing the membership using the membership function, and take the smallest
between value the A1 and B1, and find the result W1 as in the following
equations.
Suitability for Rule 1: W1 = μA1 (x) ∧ μB1 (y)
Suitability for Rule 2: W2 = μA2 (x) ∧ μB2 (y)
Second, we get the inference result for each rule by applying the generated
suitability value onto the Fuzzy sets. The fuzzy set indicates set for the dis-
comfort index; C1 is very discomfort, C2 is discomfort. This kind of informa-
tion also must be saved in the internal knowledge base. The inference results
for the each rule gets smaller value between the suitability value W1 and
value for Ci.
Inference Result for Rule1: μc (z) = W1 ∧ μc1 (z).∀z ∈ Z
1
Inference Result for Rule2: μc (z) = W2 ∧ μc2 (z).∀z ∈ Z
2
Third, find final inference result from the biggest value from acquired suitabil-
ity values. In other words, we will take the max value if C1 (very discomfort)
exists with more than a single value. Generally, for N rules, we can set the
equation as follows.
Inference Result for Rule1:
μc (z) = μc (z) ∨ μc (z) ∨ μc (z) · · · ∨μcn (z).∀z ∈ Z
1 2 3
Finally, the result from the previous steps is expressed as fuzzy set, and it
needs to go through the Defuzzification to get the final result. Among many
methods, we are going to use the Center for Gravity method as follows.
μc (z) · zdz
Z0 =
(1)
μc (z) dz
An Approach for Learnable Context-Awareness System 119
Therefore, the data type for the semantic data that the CMA sends to the
middleware layer is the highest membership status, and it is expressed as
normalized real value (0 1).
context by using this previous record, and use K-nearest algorithm to do this.
We are going to use the following equation to search for the most similar case.
Wi · f (d)(Cki , Cci )
Similarity = (2)
Wi
The parameter Wi is the weight for the context, and the function f represents
the distance value between the previous context and the current context.
When the function f gets implemented, the environmental information will
An Approach for Learnable Context-Awareness System 121
be calculated as distance value. And the information for when, and where
will be calculated through the similarity matrix. There are many ways to
compute the weight, and one of the methods could be the following equation.
|f (d)(Cki , Cci )|
W eighti = (3)
|f (d)(Cki , Cci )|
From the above procedure, we can compute the similarities for the every case,
and will choose one from them. But, if the selected one is the most similar
case, and there still exists differences from the current case. Then we need to
solve the differences by going through the adaptation stage.
Case Adaptation. The SRA performs the adaptation stage to resolve the
differences between the current context and the selected case, by applying
adaptation rules [12]. There are two types rule bases; one is the match rule
base, and the other is range rule base (see Table 2). The match rule represents
exact matches between current situation and similar case. The range rule
allows the temporary boundary values. These Two types of the rules are
pre-defined in the general knowledge base, and the learned result from the
feedback will be stored in the personal knowledge base. When the adaptation
is being done, then both the rule base will be searched. If more than one rule
is triggered, then the most recent rule gets fired. The rule adaptation process
is as follows.
If the Rule1 fires, the solution changes as follows. We can generate more
flexible rules like in table 3, and this approach could be worthwhile to produce
more appropriate solution.
4 Simulated Experimentation
We first designed the experimentation window as in the Fig. 9. In this figure,
the upper left corner of the window shows extracted case from the ontology,
which has semantic data type. The upper right corner of the window repre-
sents the result of comparison between the current case and past context in
Case DB. The last column of the table shows the similarity, and we can see
that second case is most similar one. The lower right corner of the window
shows the systems adaptation rule, and we can find out the last column fired
shows that which rule has been applied with the mark x and o. Finally, lower
right corner of the window shows the system log as the system progresses,
and the suggested solution.
The procedure for our simulated experimentation can be explained as fol-
lows. It first provides solution through the SRA, proceeds evaluation and
learning based on the solution, and finally shows the result after the learn-
ing. The scenario for this experimentation is as follows.
- Tony comes home from the school after finishing exercise.
- When Tony goes to his room, the CMA recognizes him and sends this in-
formation and surrounding context information to the system.
- The system recognizes the status of Tony, and turns on the air-conditioner
and dehumidifier.
- Tony move to the living room to study, and the CMA recognizes this and
provides proper illumination for studying
- But Tony does not satisfy the illumination and temperature of air-conditioner.
- Therefore, Tony controls the illumination and temperature of air-conditioner.
- The system learns this feedback, and saves it as personal rule.
From the example, using the five environmental information, four actual con-
texts and one dependent context is generated through the CMA. If we apply
it into real world application, additional attributes should be added based on
the characteristics of a domain.
From the Fig 8, the left window represents the Humidity among the actual
context and the right window shows Uncomfort, which is dependent context
by using the Temp and Humidity.
Now, the simulation procedure works as follows. For the first step, when
Tony comes into the room, the system recognizes and provides a solution
as the following three commands, Play ballad music, set the air-conditioner
level as 3, and set the dehumidifier level as 2, which can be seen in lower right
corner window in the Fig.9. In this simulation, since there is no adaptation
rule exists, the temperature context was not considered. Therefore, Tony set
the air-conditioner level at 5 by himself, and the system recognizes the change
and generates a new rule to solve this problem automatically.
From the Fig. 10 (extracted only the necessary sub-windows from the
original window), we can see that a new rule is added in the bottom left
An Approach for Learnable Context-Awareness System 125
of the window, which will be saved in Tonys personal knowledge base. The
stored solution is the air-conditioner level 5, which has been set by Tony as a
feedback. When the same situation happens for the next occasion, then the
system will apply the personal rule first, as in the Fig. 11. We can see the
system provides revised solution by adapting the differences from the current
situation.
126 I. Jang and C.-W. Woo
5 Conclusion
In this paper, we have designed and developed agent based context awareness
system that can handle some limitations of the previous context awareness
systems. The significance of our study is as follows. First, out system archi-
tecture does not need to be modified for any environmental change, and it
could be adapted into any application domains easily. Second, the system
employed agent based architecture to enhance the intelligence of the system.
For instance, the CMA treats the conversion of numerical data into seman-
tic data using Fuzzy inference mechanism. This approach can provide more
reliable solution to the user. Third, the SRA utilizes CBR engine to make
inference, and also reflects user feedback as adaptation rule to enhance the in-
ference capability. This approach brings advantage of providing personalized
service to the user. Fourth, we have performed our approach as a simulated
experimentation with a virtual scenario. Even if our approach shows some
successful results, it still needs to be enhanced in few issues. First, the Fuzzy
data conversion of the CMA needs to be tested with more sophisticated Fuzzy
theorem, and the SRA needs to handle conflicts of personalized rules as the
learning progresses.
An Approach for Learnable Context-Awareness System 127
References
1. Dey, A., Abowd, G.: Towards a Better Understanding of Context and Context-
Awareness. In: Gellersen, H.-W. (ed.) HUC 1999. LNCS, vol. 1707, pp. 304–307.
Springer, Heidelberg (1999)
2. Chen, H., Finin, T., Joshi, A.: An Ontology for Context-Aware Pervasive Com-
puting Environments, Special Issue on Ontologies for Distributed Systems.
Knowledge Engineering Review, 18, 197–207 (2003)
3. Gu, T., Pung, H., Zhang, D.: A Middleware for Building Context-aware Mo-
bile Services. In: Proceedings of Vehicular Technology Conference (VTC 2004),
vol. 5, pp. 2656–2660 (2004)
4. Dey, A., Abowd, G.: The Context Toolkit: Aiding the Development of Context-
Aware Applications, In: Proceedings of the SIGCHI Conference on Human
Factors in Computing Systems, pp. 434–441 (1999)
5. Ranganathan, A., Campell, R.: A Middleware for Context Aware
Agents in Ubiquitous Computiong Environments. In: Proceedings of
ACM/IFIP/USENIX International Middleware Conference, Rio de Janeiro,
Brazil (2003)
6. Dey, A., Salber, D., Abowd, G.: A Conceptual Framework and a Toolkit for Sup-
porting the Rapid Prototyping of Context-Aware Applications. Human Com-
puter Interaction 16(2-4), 97–166 (2001)
7. Kindberg, T., et al.: People, Places, Things: Web Presence for the Real World.
In: Proceedings of the 3rd IEEE Workshop on Mobile Computing Systems and
Applications, Monterey, California, USA, pp. 19–28 (2000)
8. Gu, T., Wang, X., Pung, H., Zhang, D.: An Ontology-based Context Model
in Intelligent Environments, In: Proceedings of Communication Networks and
Distributed Systems Modeling and Simulation Conference, San Diego, Califor-
nia, USA (January 2004)
9. Kim, Y., Uhm, Y., et al.: Context-Aware Multi-agent Service System for As-
sistive Home Applications. In: Ma, J., Jin, H., Yang, L.T., Tsai, J.J.-P. (eds.)
UIC 2006. LNCS, vol. 4159, pp. 736–745. Springer, Heidelberg (2006)
10. Roman, M., Compbell, R.: GAIA: Enabling Active Spaces. In: Proceedings of
the 9th ACM SIGOPS European Workshop, Kolding, Denmark (2000)
11. Mamdani, E., Assilian, S.: An Experiment in Linguistic Synthesis with a Fuzzy
Logic Controller. International Journal of Human-Computer Studies 51(2),
135–147 (1999)
12. Aamodt, A., Plaza, E.: Case-Based Reasoning: Foundational Issues, Method-
ological Variations, and System Approaches. AI Communications 7(1), 39–59
(1994)
13. Hanney, K., Keane, M.: The Adaptation Knowledge Bottleneck How to Ease it
by Learning from Cases. In: Leake, D.B., Plaza, E. (eds.) ICCBR 1997. LNCS,
vol. 1266, pp. 359–370. Springer, Heidelberg (1997)
14. Baldauf, M., Dustdar, S.: A Survey on Context-aware systems. International
Journal of Adhoc and Ubiquitous Computing 2, 263–277 (2004)
A Hybrid Multi-Agent Framework for
Load Management in Power Grid
Systems
Minjie Zhang, Dayong Ye, Quan Bai, Danny Sutanto, and Kashem Muttaqi
1 Introduction
In power grid systems, faults at certain nodes can lead to cascading blackouts
in the whole system. When a power system is facing the threat of a widespread
power outage, the system will be in a vulnerable state [5]. Catastrophic fail-
ures in power grid systems bring great inconvenience to people’s living and
are very harmful for both power industries and the social economics. To avoid
Minjie Zhang · Dayong Ye
School of Computer Science and Software Engineering, University of Wollongong,
Wollongong, NSW 2522 Australia
e-mail: {minjie,dy721}@uow.edu.au
Quan Bai
Tasmanian ICT Centre, CSIRO, Hobart, TAS 7001 Australia
e-mail: quan.bai@csiro.au
Danny Sutanto · Kashem Muttaqi
School of Electrical, Computer and Telecommunications Engineering, University of
Wollongong, Wollongong, NSW 2522 Australia
e-mail: {danny,kashem}@elec.uow.edu.au
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 129–143.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
130 M. Zhang et al.
2 Related Work
In the last decade, intelligent agent technology has been adopted for various
aspects of power systems management, such as restoration [15], relaying [2],
A Hybrid Multi-Agent Framework for Load Management 131
maintenance [6], substation automation [1] and state estimation [12]. In this
paper, we focus on power system restoration when one or several generators
in a power grid system are out of order.
Jung and Liu [5] presented a multi-agent framework that provides real-time
information acquisition and interpretation, quick vulnerability evaluation of
both power and communication systems, and preventive and corrective self-
healing strategies to avoid catastrophic failures of a power system. However,
their framework is only a preliminary work, and much details have to be
done.
Nagata and Sasaki [10] developed a multi-agent framework for power sys-
tem restoration. Their framework is in centralized design, which consists of
several bus agents and a single facilitator agent. Bus agents are used to decide
suboptimal target configuration after a fault occurrence, while the facilitator
agent acts as a manager in the decision process. Srivastava et al. [16] pro-
posed a similar method to use a coordinating agent with global information
was used for reconfiguration of the shipboard power system. Nagata et al. [9]
improved the restoration methodology proposed in [10], but facilitator agents
were still required for coordination of the agents. Momoh and Diouf [8] refined
the work done in [9]. They utilized power generation agents, bus agents, and
circuit breaker agents to distribute the reconfiguration functionalities. How-
ever, the system proposed in [8] still needs facilitator agents. The methods
proposed in [10] [16] [9] and [8] are centralized multi-agent systems. The cen-
tralized manner might bring the framework the potential of the single point
of failure.
To overcome the drawback of centralized methods, some decentralized ones
have also been presented. In [11], Nordman and Lehtonen proposed a new
agent concept for managing an electrical distribution network. Their concept
consisted of three aspects, which include secondary substation object, de-
centralized functionality and an information access model. However, because
all secondary substations are copies of the secondary substation objects, all
secondary substations are homogeneous with the same type of agent intelli-
gence. This homogeneous feature might limit the adaptation of this concept.
Moreover, according to their decentralized functionality, when a primary sub-
station wants to deliver a permission message to a specific secondary substa-
tion, the primary substation has to pass the permission message along the
communication path instead of directly to the target secondary substation.
This communication feature might rise communication costs, and delay the
decision making in emergency situations.
Solanki et al. [15] provided a multi-agent framework with detailed design
of each agent, which is used to restore a power system after a fault. The
framework is decentralized and topology independent which can overcome
the scalability of limitations of existing restoration techniques. Although the
decentralized manner can avoid single point of failure, the decision accuracy
usually cannot be guaranteed. This is because each node in the decentralized
132 M. Zhang et al.
architecture only has a limited view of the whole working environment and
makes decisions based only on its incomplete information.
A multi-agent system based reconfiguration approach for mesh structured
power systems was introduced in [4]. Although the authors claimed that the
architecture of the multi-agent system is decentralized, some global informa-
tion is still employed, e.g. the net power of the power system. Moreover, their
work overlooked negotiation process of the multi-agent system, which is very
important for a multi-agent system to perform properly.
In this research, we propose a Hybrid Multi-Agent Framework (HMAF)
for load management in power grid systems, which deploys intelligent agents
in different power grids, e.g. buses, generators and transformers, to execute
monitoring, analysing and maintaining activities. Agents in different nodes
can also interact with each other to exchange their local information and
decisions. Through binding intelligent agents to power grids, we can decom-
pose and allocate complex problems to local agents. In this case, a power
grid system can be considered as a Multi-Agent System (MAS), and many
agent coordination techniques could be borrowed to solve load balancing and
system protection problems in the power system domain. Compared with
centralized methods, e.g. [10] and [16], our framework can avoid the single
point of failure. In contrast to current decentralized approaches, such as [11]
and [15], our framework can provide sufficient information for nodes to make
accurate decisions.
since a central controller has a global view of the whole system. However, such
architectures may have low response speed especially when the system scale
is large. For large scale power grid systems, there are a large number of nodes
which need to be monitored and managed. This will bring huge computation
and communication overheads to the central controller of the system. In this
case, a central controller will not be able to give quick response to different
nodes in the system within a limited time.
To overcome the above limitation, decentralized system architecture is a
potential solution. In a decentralized architecture, a complex problem is de-
composed into a number of sub-problems and assigned to different computing
nodes. As each node only needs to process a small amount of data and han-
dle a “simple” sub-problem, it can solve the assigned task in shorter time.
Decentralized architecture enables a system short response time, but the ac-
curacy cannot be guaranteed. This is because each node in the system only
have a limited view of the whole working environment, and needs to provide
solutions based on its incomplete vision.
4. TakeMeasures: This plan reasons current state and makes decisions about
taking measures against faults. When this plan makes a decision, it will
send a message with conversation ID “DecMaking” to the Local Coor-
dinator Agent. This message is handled by the Local Coordinator Agent
DecEstimation plan. The countermeasures include transformer tap change,
switch open/close and Strategic Load Shedding, which were listed in [7].
LSA Cap, LSA Load), where AgentID is the agent identification number,
N umLSA is the number of Local Scheduler Agents in the same area with
the Local Coordinator Agent, LSA IDs stores the identification numbers of
these Local Scheduler Agents, and LSA Cap and LSA Load demonstrate
the capacity and current load of each Local Scheduler Agent respectively.
A Local Coordinator Agent has two plans.
1. InfoCollection: This plan collects and distributes information from/to the
Local Scheduler Agents. When this plan receives a “MoreNode” message, it
sends a message with conversation ID “NodeInfo” and content as states of
the nodes back to the requesting Local Scheduler Agent. The “NodeInfo”
message is handled by the Local Scheduler Agent DataAnalysis plan.
2. DecEstimation: This plan estimates whether decisions, made by LSAs, are
reasonable. The estimation is based on local balance in the associated area.
When this plan receives a “DecMaking” message, it responds a message
with conversation ID “EstResult” and content as the result of estimation
and its suggestion, if applicable, to the requesting Local Scheduler Agent.
The “EstResult” message is handled by the Local Scheduler Agent Take-
Measures plan.
People might argue that, in our framework, if the Local Coordinator Agent
is out of order, the associated area could not operate properly. However, the
Local Coordinator Agent in our framework is just an information provider
rather than a manager, which is different from the Facilitator Agent in [10].
Hence, if Local Coordinator Agents are out of order, this framework only
degrades to a general decentralized framework, which can still work as other
decentralized architectures.
A Hybrid Multi-Agent Framework for Load Management 139
4 Operation of HMAF
To demonstrate the operation of HMAF, an example is introduced in Fig. 4,
where Gen 1, 2, 3 and 4 denote Generator 1, 2, 3 and 4, respectively, and
the arrows indicate loads. As displayed in Fig. 4, Local View 1 consists of
Generator 1 and Load 1, 2 and 4; Local View 2 is composed of Generator 2,
Load 2 and Load 5 ; Local View 3 is formed with Generator 3, Load 7 and
Load 8 ; Local View 4 contains Generator 4, Load 6 and Load 9.1 Based on
HMAF, the power grid system shown in Fig. 4 can be mapped into a multi-
agent framework which is displayed in Fig. 5. It is assumed that there is a
fault on Generator 1 and the Local Scheduler Agent 1 (LSA 1) has detected
this fault.2 Then, the LSA 1 will attempt to request other LSAs for help to
restore its local power.
this example, the reasoning result is that Generator 4 can supply the power
20kw. Thereby, LSA 1 sends a request message to the selected agent, namely
LSA 4 in this example, according to its AgentID with the conversation ID
“Request” and content “Power Supply Request: 20kw”. As long as LSA 4
receives the request messages, it supplies the designated power to the area
which was supplied by Generator 1, and LSA 1 assigns this power to Load
2. The interaction process is displayed in Fig. 6. In this example, Generators
2, 3 and 4 can supply sufficient power for Generator 1. However, in some
cases, if other generators could not provide enough power for Generator 1,
Generator 1 would have to make a decision about discarding current load,
and asks LCA to estimate whether this decision is reasonable. The detailed
estimation process is one of our future works.
5 Conclusion
In this paper, a hybrid multi-agent framework for power grid systems load
management was proposed. The contribution of this framework is that it
combines centralized and decentralized architecture together. Compared with
centralized architectures, this framework can avoid single point of failure in
some extent. In contrast to complete decentralized architectures, this frame-
work can provide nodes sufficient information for making accurate decisions
and quick response when a fault occurs in the system.
142 M. Zhang et al.
In the future, a formal detection model for quick fault discovery in power
grid systems will be developed. In addition, an efficient resource allocation
algorithm will also be devised to efficiently distribute power to the local
view in which the generator is out of order. Finally, this framework will be
implemented and tested in some real cases.
References
1. Buse, D.P., Sun, P., Wu, Q.H., Fitch, J.: Agent-based substation automation.
IEEE Power Energy 1(2), 50–55 (2003)
2. Coury, D.V., Thorp, J.S., Hopkinson, K.M., Birman, K.P.: Agent technology
applied to adaptive relay setting for multi-terminal lines. In: Proc. IEEE Power
Eng. Soc. Summer Meeting, pp. 1196–1201 (2000)
3. Green, S., Hurst, L., Nangle, B., Cunningham, P., Somers, F., Evans, R.: Soft-
ware Agents: A Review. Tech. Rep. TCD-CS-1997-06, Trinity Collega, Dublin,
Ireland (1997)
4. Huang, K., Srivastava, S.K., Cartes, D.A., Sloderbeck, M.: Intelligent agents
applied to reconfiguration of mesh structured power systems. In: Proc. Interna-
tional Conference on Intelligent Systems Applications to Power Systems, Toki
Messe, Niigata, pp. 1–7 (2007)
5. Jung, J., Liu, C.: Multi-agent technology for vulnerability assessment and
control. In: Proc. IEEE Power Eng. Soc. Summer Meeting, Vancouver, BC,
Canada, vol. 2, pp. 1287–1292 (2001)
6. Kezunovic, M., Xu, X., Wong, D.: Improving circuit breaker maintenance man-
agement tasks by applying mobile agent software technology. In: Proc. IEEE
Power Eng. Soc. Asia Pacific Transm. Distrib. Conf., pp. 782–787 (2002)
7. Lachs, W., Sutanto, D.: Voltage instability in interconnected power systems:
A simulation approach. IEEE Transactions on Power Systems 7(2), 753–761
(1992)
8. Momoh, J., Diouf, O.: Optimal reconfiguration of the navy ship power sys-
tem using agents. In: Proceedings of IEEE PES Transmission and Distribution
Conference and Exhibition, pp. 562–567 (2006)
9. Nagata, T., Fujita, H., Sasaki, H.: Decentralized approach to normal operations
for power system network. In: Proceedings of the 13th International Conference
on Intelligent Systems Application on Power Systems, pp. 407–412 (2005)
10. Nagata, T., Sasaki, H.: A multi-agent approach to power system restoration.
IEEE Transactions on Power Systems 17(2), 457–462 (2002)
11. Nordman, M.M., Lehtonen, M.: An agent concept for managing electrical distri-
bution networks. IEEE Transactions on Power Delivery 20(2), 696–703 (2005)
12. Nordman, M.M., Lehtonen, M.: Distributed agent-based state estimation for
electrical distribution networks. IEEE Transactions on Power Systems 20(2),
652–658 (2005)
A Hybrid Multi-Agent Framework for Load Management 143
13. Poutakidis, D., Padgham, L., Winikoff, M.: Debugging Multi-Agent Systems
Using Design Artefacts: The Case of Interaction Protocols. In: Proceedings of
the First International Workshop on Challenges in Open Agent Systems at
AAMAS 2002, Bologna, Italy, pp. 960–967 (2002)
14. Rao, A., Georgeff, M.: An Abstract Architecture for Rational Agents. In: Pro-
ceedings of the Third International Conference on Principles of Knowledge
Representation and Reasoning, San Mateo, USA, pp. 439–449 (1992)
15. Solanki, J.M., Khushalani, S., Schulz, N.N.: A multi-agent solution to distri-
bution systems restoration. IEEE Transactions on Power Systems 22(3), 1026–
1034 (2007)
16. Srivastava, S.K., Xiao, H., Butler-Purry, K.L.: Multi-agent system for auto-
mated service restoration of shipboard power systems. In: Proceedings of the
15th International Conference on Computer Applications in Industry and En-
gineering, San Diego, CA, USA, pp. 119–123 (2002)
17. Wooldridge, M.: An Introduction to Multiagent Systems. John Wiley & Sons
Ltd., Chichester (2002)
18. Wooldridge, M., Jennings, N.: Intelligent Agents: Theory and Practice. The
Knowledge Engineering Review 10(2), 115–152 (1995)
Part II
Agent-Based Simulation for Complex
Systems: Application to Economics,
Finance and Social Sciences
146 Agent-Based Simulation for Complex Systems
Social Science domain is certainly one of the research domains that con-
tributes the most to the advance of Multi-Agent Systems. A lot of agent-based
models and simulators have been developed to study real social systems (e.g.
a town) with various points of view such as demography, transportation,
social organisation and so one. Following papers illustrate the use of agent
paradigm in various application fields of social sciences. El Saadi, Bah and
Agent-Based Simulation for Complex Systems 147
Takao Terano
Tohgoroh Matsui
Kiyoshi Izumi
MASFE’09 Workshop Chairs
Alexis Drogoul
Benoit Gaudou
Nicolas Marilleau
AAECS’09 Workshop Chairs
Financial Frictions and Money-Driven
Variability in Velocity
José J. Cao-Alvira
1 Introduction
Money-driven general equilibrium models with great difficulty are capable to
explain the variability on velocity observed in the data, due to the insufficient
propagation mechanisms associated with monetary shocks. Alas, a proper as-
sessment on the variability of consumption velocity of money is fundamental
to understanding the role of money in the business cycle and bestows insight
on the microfoundations of monetary policy.
Attempting to answer for a significant allotment of its variability, financial
frictions are introduced in the information structure of a cash-in-advance
(CIA) dynamic stochastic general equilibrium model with the interest of
studying the impact on velocity of serially correlated monetary shocks. On
José J. Cao-Alvira
Universidad de Puerto Rico, Rio Piedras PR 00931, USA
e-mail: JoseJulian.Cao@UPR.edu
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 149–166.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
150 J.J. Cao-Alvira
Mt Mt−1
ct + kt+1 + = Yt + (1 − δ)kt + + Tt (2)
pt−1 pt−1
The budget constraint in eq. (2) implies that current period’s consumption,
real money holdings and next period capital stock are financed by actual
production, undepreciated capital stock, and the real money balances held
from the previous period plus a lump-sum transfer. The CIA constraint
requires the agent to hold sufficient real money balances in order to purchase
the consumption good.
Defining λt as the Lagrangian multiplier associated with the budget con-
straint and μt as the multiplier for the CIA constraint, the planner’s problem
is identified by the first order conditions in eqs. (5) − (8):
λt = c−τ
t − μt (5)
Yt
γn nγt = λt (1 − α) (6)
nt
Yt+1
λt = βEt λt+1 α + (1 − δ) (7)
kt+1
Et {λt } c−τ
t+1
= βEt (8)
pt−1 pt
Variable Expression
Consumption Velocity Vt = MC t
t /pt
pt
Inflation πt = pt−1
Nominal Interest I
t = rt· Et {πt+1 }
Y
Real Interest t+1
rt = Et λt+1 α kt+1 + (1 − δ)
and market clearing conditions for the goods and the money markets, in eqs.
(10) − (11), respectively:
k ss αβ
= (12)
y ss 1 − β (1 − δ)
iss k ss
ss
= δ ss (13)
y y
css iss
ss
= 1 − ss (14)
y y
−α γ+τ
−τ ss τ1−α 1
β (1 − α) css y
nss = ss (15)
θ γn k ss k ss
ss
θ
μss = λss −1 . (16)
β
The steady state of the CIA constraint is derived from the equilibrium con-
dition of real money balances, eq. (8). Notice the constraint is binding in
the steady state for a steady state growth of money supply greater than the
time discount factor.
Q
ỹwz
K
Ruw k, M ; Υ , P̄ = Υuw (k, M )−β Πwz {Υwz (k̃uw , M̃ )[α +1−δ]} (17)
z=1 k̃uw
Q
Υuw (k, M )
Q
c̃−τ
RuM k, M ; Υ , P̄ = Πuw − Πwz wz
, (18)
w=1
Pu (k, M ) z=1 Pw uw , M̃ )
(k̃
for all u = {1, ..., Q} & w = {1, ..., Q}. The real variables are defined by
yuw = Ak α n1−α
uw (19)
uw + (1 − δ) k − cuw
k̃uw = Ak α n1−α (20)
1−α
(1 − α) γ+1
nuw = Υuw (k, M )Ak α (21)
γn
Υuw (k, M )−τ if μuw = 0
cuw = . (22)
M+1 /Pu (k, M̃ ) if μuw > 0
The money supplies ex-ante and ex-post the realization of the growth rate
shocks are defined by M̃ and M+1 , respectively
M̃ = θ(u)M (23)
156 J.J. Cao-Alvira
3 Solution Methodology
The solution procedure involves the usage of finite elements in its approxima-
tion of the policy functions and a parameterized expectations algorithm to
minimize the weighted absolute value of residual functions Ruw K
(k, M ; Υ , P̄ )
and Ru (k, M ; Υ , P̄ ), for all u & w; where the true decision rules Υuw (k, M )
M
h
& Pu (k, M ) are replaced by the parametric approximations υuw (k, M ) &
h h h
pu (k, M ). υuw & pu are approximated using an implementation of the finite
element method, that follows that advocated in McGrattan (1996), see [13].
h
To create
the approximate
time invariant functions υuw & phu , the space
Ω = k, k̄ × M, M̄ is divided in ne nonoverlapping rectangular subdomains
¯ ¯
called “elements”. Parameterization of the policy functions for each element,
at each realization of u & w, are constructed using linear combinations of
low order polynomials or “basis functions”. This procedure creates local
approximations for each function. Given the discrete nature of Λ, this state
space need not to be divided.
h
The parameterized functions υuw (k, M ) & phu (k, M ) are built as follows:
IJ
h uw
υuw (k, M ) = υij Wij (k, M ) (25)
ij
IJ
phu (k, M ) = puij Wij (k, M ) . (26)
ij
Where i = {1, ..., I} indicate capital nodes, j = {1, ..., J} indicate money
supply nodes, Wij (k, M ) is a set of linear basis functions dependent on the
element [ki , ki+1 ] × [Mj , Mj+1 ] , for all i, j, over which the local approxima-
uw
tions are performed. υij & puij are vectors of coefficients to be determined.
The parameterized value of the conditional expectation function and the price
function over the full state space are obtained by piecing together all the local
h
approximations. The approximate solutions for υuw (k, M ) & phu (k, M ) on Θ
are then “piecewise linear functions”.
Wij (k, M ) are the basis functions that the finite element method em-
ploys. These are constructed such that they take a value of zero for most
Financial Frictions and Money-Driven Variability in Velocity 157
of the space Ω, except for a small interval where they take a simple linear
form. The basis functions adopted for these approximations are set such that
Wij (k, M ) = Ψi (k) Φj (M ) , where
⎧ k−ki−1
⎪
⎨ ki −ki−1 if k ∈ [ki−1 , ki ]
ki+1 −k
Ψi (k) = if k ∈ [ki , ki+1 ] (27)
⎪
⎩ ki+1 −ki
0 elsewhere
⎧ M−1 −Mj−1
⎨ Mj −Mj−1 if M ∈ [Mj−1 , Mj ]
⎪
Mj+1 −M
Φj (M ) = M if M ∈ [Mj , Mj+1 ] , (28)
⎪
⎩ j+1 −Mj
0 elsewhere
for all i, j. Ψi (k) & Φj (M ) have the shape of a continuous pyramid which
peaks at nodal points k = ki & M = Mj , respectively, and are only non-zero
on the surrounding elements of these nodes.
h
The approximations υuw (k, M ) & phu (k, M ) are chosen to simultaneously
satisfy the equations:
M̄ k̄
K
ω (k, M ) Ruw k, M ; υ h , p̄h dkdM = 0 (29)
M k
¯ ¯
M̄ k̄
ω (k, M ) RuM k, M ; υ h , p̄h dkdM = 0, (30)
M k
¯ ¯
for all u = {1, ..., Q} & w = {1, ..., Q}. ω (k, M ) is a weighting function,
and RuwK
(k, M ; υ h , p̄h ) & RuM (k, M ; υ h , p̄h ) are the residuals from the Eu-
ler equation and the equilibrium condition for real money holdings, where
the true policy functions Υ & P̄ are replaced by the vectors of paramet-
ric approximations υ h & p̄h . A Galerkin scheme is employed to find the
uw
vectors of coefficients υij & puij , for all i, j and all u & w, which solves
for the weighted residual equations (29) & (30) over the complete space Θ.
The Galerkin scheme employs the basis functions Wij (k, M ) as weights for
K
Ruw (k, M ; υ h , p̄h ) & RuM (k, M ; υ h , p̄h ):
M̄ k̄
K
Wij (k, M ) Ruw k, M ; υ h , p̄h dkdM = 0 (31)
M k
¯ ¯
M̄ k̄
Wij (k, M ) RuM k, M ; υ h , p̄h dkdM = 0 (32)
M k
¯ ¯
for all i, j and all states u & w. Since the basis functions are only nonzero
surrounding their nodes, eqs. (31) & (32) can be rewritten in terms of the
individual elements:
158 J.J. Cao-Alvira
ne
K
Wij (k, M ) Ruw k, M ; υ h , p̄h dkdM = 0 (33)
e=1 Ωe
ne
Wij (k, M ) RuM k, M ; υ h , p̄h dkdM = 0, (34)
e=1 Ωe
for all i, j and all states u & w. ne is the total number of elements and Ωe
is the capital and money stock domain covered by the element e.
A Newton algorithm is used to find the coefficients for [υs , p̄s ] which solve
for the nonlinear system of equations H:
Where H ([υs , p̄s ]) is denoted by eqs. (33) & (34). The first step is to choose
initial vectors of coefficients [υs0 , p̄s0 ], and iterate as follows:
−1
υsl+1 , p̄sl+1 = [υsl , p̄sl ] − J ([υsl , p̄sl ]) H ([υsl , p̄ul ]) . (36)
J is the Jacobian matrix of H, and [υsl , p̄sl ] is the lth iteration of [υs , p̄s ].
The algorithm solves for the parameterized version of the conditional expec-
h
tation function of the Euler equation υuw (k, M ) and for the price function
pu (k, M ) , for {θ (u) , θ!(w)}
h
⊂ Λ and
{k, M } ⊂ !Ω. Convergence is assumed
to have occurred once ! υsl+1 , p̄sl+1 − [υsl , p̄sl ]! < 10−7 .
5. Compute the values of employment nuw , actual output yuw , next period
capital k̃uw , using eqs. (21), (19) and (20) respectively, and of real money
holdings M+1 /phu (Ωt ). Where M+1 = θ (w) · M̃ , and M̃ = θ (u) · M .
6. Check whether the initial conjecture was correct. If “yes”, go to next
step. If not, the constraint binds: μuw > 0 & cuw = M+1 /phu (Ωt ). The
value of the CIA multiplier becomes μuw = [c−τ uw − υuw (Ωt )].
h
7. For each realization θt = θ (w) approximate Pw with phw , and for each
h
θt+1 = θ (z) approximate Υwz with υwz , for all Gauss-Legendre capital
and money stock abscissas of each element along the state space Ωt+1 .
8. Create a conjecture that the CIA constraint does not bind at time t + 1,
h −τ
i.e. μwz = 0. Consumption becomes cwz = υwz (Ωt+1 ) .
h
9. Compute the value of each possible nwz , ywz , and M̃+1 /pw (Ωt+1 ). Where
M̃+1 = θ (z) · M+1 .
10. Check whether the conjecture in Step 8 was correct. If “yes”, go to next
step. If not, the constraint binds: μwz > 0, & cwz = M̃+1 /phw (Ωt+1 ).
The value of the CIA multiplier becomes μwz = [c−τ
wz − υwz
h
(Ωt+1 )].
11. Construct the residual functions Ruw k, M ; υ , p̄ & Ru k, M ; υ h , p̄h
K h h M
A α β δ γ τ θss
1.00 0.33 0.99 0.02 0.00 3.00 1.01727
160 J.J. Cao-Alvira
Fig. 1 The left and right panels illustrate the values assumed by μt on the bench-
mark economy when frictions are absent and present on the information structure,
respectively. On the environment with frictions, the CIA-constraint binds at the
steady state money growth rate but it does not at lesser growth rates.
162 J.J. Cao-Alvira
Fig. 2 The two panels depict those cases where the previously realized growth
parameter is greater than the steady state value. The lesser the realization of the
previous period’s money growth parameter, the more possible cases will occur that
the agent will decide to hold on cash for next period and the CIA-constraint will
not bind.
possible to notice that below a critical level of capital stock, the real interest
is sufficiently high to dissuade the agent from holding any money for pre-
cautionary reasons, and instead acquire interest bearing physical capital in
its portfolio. Cao-Alvira (2010b) presents a thorough description of a projec-
tion methodology that solves for a similar cash-in-advance model economy
with no frictions in its information structure, as well as a robustness analysis
and cutoff points for the existence of variability on velocity. The considered
model economy in Cao-Alvira (2010b) contains a more comprehensive state
space partition and, as a consequence, includes more possible realizations of
the money growth rate where the slackness multiplier is zero.
Observing the case depicted in panel 2 of Fig. 1, the environment where
frictions are present, it is possible to notice two key features. First, μt is
positive, or the CIA constraint binds, at the steady state of the benchmark
economy. This result is consistent with those findings reported in Table 3 and,
as expected, information or financial frictions do not affect the steady state
specification of the model. And second, μt equals zero, on all considered values
of Ωt , for those parameters θt < θss . At the considered state space partition,
serially correlated shocks on monetary growth, combined with frictions on the
information structure, pushes the agent to accumulate excessive real money
holdings on states of nature that exhibit a lower than average money growth
rate. On these states, cash is being transferred from one period to the next
in order to smooth consumption and decrease its expected marginal utility
variability.
Financial Frictions and Money-Driven Variability in Velocity 163
Fig. 3 The two panels depict those cases where the previously realized growth
parameter is greater than the steady state value. Only on the scenario where previ-
ously the highest possible growth parameter is realized, the right panel, the agent
will decide to never carry excessive cash holdings and the CIA constraint will always
bind.
Table 4 Simulated moments of three key indexes for the friction and frictionless
version of the benchmark economy and those for the US economy for the time
period 1959:I to 1998:III. Numbers in parenthesis are standard deviations over 500
simulations. US economy values as reported by Wang W., Shi S. (2006).
5 Conclusion
This paper introduces a modeling environment where frictions are incorpo-
rated on the selection of assets on an agent’s portfolio with the interest of
studying the impact of serially correlated monetary shocks on the variability
of velocity. The solution technique proposed and developed in finding the
equilibrium of the system employs the usage of finite elements in the ap-
proximation of the policy functions, and a parameterization of the agent’s
expectations over the choice of future real money balances. The method con-
verges with rapid speed and is efficient approximating the studied highly
nonlinear environment. Because the algorithm allows for the cash-in-advance
constraint on consumption expenditures to selectively bind, instead of exoge-
nously assuming it always does as other more conventional methods require,
it is possible for the equilibrium conditions to achieve variability in the ve-
locity of money holdings. By considering frictions on the financial structure
of the economy, a solely money-driven monetary benchmark economy is able
to significantly increase the explanation of observed variability on the data,
when judged against a comparable frictionless environments.
References
1. Aruoba, S.B., Rubio, J., Fernandez-Villaverde, J.: Comparing Solution Methods
for Dynamic Equilibrium Economies. J. Econ. Dyn. Control 30, 2477–2508
(1998)
2. Cao-Alvira, J.: Finite elements in the presence of occasionally binding con-
straints. Comput. Econ. 35(4), 355–370 (2010)
3. Cao-Alvira, J.: Velocity variability on cash-in-advance economies due to
precautionary demand. EGAE UPR-RP Working paper (2010)
4. Cao-Alvira, J.: Non-Linear Solution Methods, Occasionally Binding Con-
straints, Information Frictions, Inventories, Cash-In-Advance Economies & Ve-
locity of Money. Economics Department, Cornell University. Ph.D. Dissertation
(2006)
5. Christiano, L., Eichenbaum, M.: Liquidity Effects and the Monetary Transmis-
sion Mechanism. Am. Econ. Rev. 82(2), 346–353 (1992)
6. Cooley, T.F., Hansen, G.: The inflation tax in a real business cycle model. Am.
Econ. Rev. 79, 733–748 (1989)
7. den Haan, W., Marcet, A.: Solving the stochastic growth model by parameter-
izing expectations J. Bus. Econ. Stat. 8, 31–34 (1990)
166 J.J. Cao-Alvira
Abstract. This paper designs a novel fuzzy competition and attitude based
bidding strategy (FCA-Bid), in which the final best bid is calculated on the
basis of the attitude of the bidders and the competition for the goods in the
market. The estimation of attitude is based on the bidding item’s attribute
assessment, which adapts the fuzzy sets technique to handle uncertainty of
the bidding process as well it uses heuristic rules to determine attitude of
bidding agents. The bidding strategy also uses and determines competition
in the market (based on the two factors i.e. no. of the bidders participating
and the total time elapsed for an auction) using Mamdani’s Direct Method.
Then the final price of the best bid will be determined based on the assessed
attitude and the competition in the market using fuzzy reasoning technique.
1 Introduction
Online auctions have become increasingly important area of research with
its popularity, because it provides the traders the flexibility of time and geo-
graphical location for trading. Software agent technology is one of the most
popular mechanisms used in online auctions for buying and selling the goods.
Software agent is a software component that can execute autonomously,
communicates with other agents or the user and monitors the state of its
Madhu Goyal
Faculty of Engineering and Information Technology, UTS, Australia
e-mail: madhu@it.uts.edu.au
Saroj Kaushik
Indian Institute of Technology, Delhi, India
e-mail: saroj@cse.iitd.ernet.in
Preetinder Kaur
Faculty of Engineering and Information Technology, UTS, Australia
e-mail: preetinder.kaur@student.uts.edu.au
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 167–180.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
168 M. Goyal, S. Kaushik, and P. Kaur
execution environment effectively [2] [4] [7]. The agents can use different auc-
tion mechanisms (e.g. English, Dutch, Vickery etc.) for procurement of goods
or reaching agreement between agents. The agent makes decisions on behalf
of consumer and endeavors to guarantee the delivery of item according to
the buyer’s preferences. In these auctions buyers are faced with difficult task
of deciding amount to bid in order to get the desired item matching their
preferences. The bidding strategies for the software agents can be static or
it may be dynamic [13]. The static agents may not be appropriate for the
negotiating market situations like extent of competition may vary as traders
leave or enter into the market, deadlines and new opportunities may increase
the pressure. The dynamic or we can say flexible negotiation capabilities for
software agents in the online auctions have become a central concern [11].
Agents need to be able to prepare bids and evaluate offers on behalf of the
users they represent with the aim of obtaining the maximum benefit [8] for
their users according to the changing market situation.
Much research has already been done by the researchers to formulate dif-
ferent bidding strategies according to the changing market situations [9] [10]
[1] [15] [14]. Strategies based on flexible negotiation agents perform better as
compared to the strategies based on fixed negotiation agents [11] [5]. Faratin
et al in [5] developed strategies based on time, attitude, resources, but many
more factors such as competition, trading alternatives are not considered. In
this paper we focus on the design of a novel bidding strategy based on the
above mentioned factors to be used by the software agent in online auction.
A fuzzy competition and attitude based bidding strategy(FCA-Bid) is de-
signed, in which the final best bid is calculated on the basis of the attitude
of the bidders as well as the competition for the goods in the market.
the attitude of the agents. The higher bid price dampens the attitude of the
agents towards the goods. Also the increasing competition for the goods in
the market increases the attitude for that good. The competition in turn
depends on the number of bidders and the time elapsed for the auction. As
the number of bidders increases, the competition among them also increases,
resulting in a higher price. In the beginning of the auction the competition is
less and it increases as time elapses and it is at the peak when time approaches
approximately in the middle of the auction period. At the end of the auction
period the competition among the bidders decreases. The steps of the design
of fuzzy competition and attitude based bidding strategy (FCA-Bid) are as
follows:
• first, each attribute is evaluated and then the assessment of all these at-
tributes will be aggregated
• then attitude of the agent will be found based on these assessments,
• next the level of competition as the function of no. of bidders and time
elapsed for the auction will be found
• Finally the best bid is calculated on the basis of the above attitude of the
agents and the competition for the goods in the market.
In this paper we have used fuzzy set methods to deal with the uncertainty,
which exists during the determination of overall assessment of the goods for
their attributes, the attitude of the agent based on the assessment of goods
and the level of competition in the market. First of all, this paper uses a
satisfactory degree measure as the common universe of assessment, i.e., an
assessment is treated as a fuzzy set on the satisfactory degree. Secondly,
an attitude is expressed as a fuzzy set on the set of assessments, i.e., the
170 M. Goyal, S. Kaushik, and P. Kaur
Attribute Assessment
The assessment of the attributes is expressed in terms of a fuzzy set. Let
A = {a1 , a2 , ..., an } be a set of assessment terms on the universe i.e. the
satisfactory degree [0,1]. This is the satisfactory degree of the agent to a par-
ticular attribute. All the fuzzy sets have same universe which is convenient for
the aggregation of various assessments. Let gk (k = 0, 1, ..., K) is the satisfac-
tory degree measure for attribute ck . Then an agent’s opinion on the goods in
terms of attribute ck is denoted by gk (u) where u(∈ U k) is the real attribute
value of attribute ck and Uk is the real universe for attribute ck . For instance,
departing time is an attribute for a flight ticket. The possible departing time
in a day is from 0:00 to 23:59. For any time slot u, a client may present a
satisfactory degree such as departing at 7:30 is with satisfactory degree 0.9
and departing at 3:00 is with 0.3. In the following, A = {a1 , ..., an } be the
set of used assessment terms which are fuzzy sets on satisfactory degree [0,1].
Then a numeric satisfactory degree is transformed to a linguistic term. In the
above example [11] a7 is with the biggest the membership degree for 0.9, the
assessment for departing at 7:30 is a6 by the maximum membership degree
principle. Similarly, the assessment for 0.3 is a2 .
Aggregation of Assessments
All the goods have a number of different attribute. So to find the overall
estimation on the good, the assessment of these all attributes will be aggre-
gated together. Take booking a flight ticket for example, an assessment is
made on a ticket usually based on the airlines, flight departure and arrival
time, flight type, aircraft types, seat positions, as well as price. The change
of an attribute’s value may leads to the alternation of an assessment. In-
stinct natures of different attributes increase the difficulty and uncertainty
for obtaining an overall assessment. Notice that an agent’s preference on
an individual attribute can be expressed through the agent’s satisfactory
degree on that attribute. This paper uses a satisfactory degree measure as
the common universe of assessment. Based on assessment on each individ-
ual attribute, an overall assessment can be obtained as follows. Suppose the
individual assessments of all attributes are v0 , v1 , ..., vK and the weights of
Automated Fuzzy Bidding Strategy 171
To find the definite value for the conclusion, here center of gravity of the
fuzzy set has been applied as follows
μz(c)cdc
c=
(5)
μz(c)dc
174 M. Goyal, S. Kaushik, and P. Kaur
R = R1 ∪ R2 ... ∪ Rn
For the input of fuzzy set E’ on E and fuzzy set C’ on C, the output fuzzy
set P on P can be obtained as follows
P = (E and C ) ◦ R = E ◦ (C ◦ R) = C ◦ (E ◦ R) (7)
and then the final price for the bid will be
3 Experimental Evaluations
In this section, an experiment implements the fuzzy bidding strategy in a
scenario in which an agent intends to book flight tickets. Six factors (as
shown in Table 2) are concerned in this situation, i.e. ticket price (c0 ), depart
time (c1 ), arrival time (c2 ), number of stops (c3 ), seat positions (c4 ), and
travel season (c5 ). The flight ticket bid for is a return ticket to destination D
with the following properties:
• price: $800 - $2000;
• depart time: 18:00 PM, Wednesday;
• return arrival time: 10:00 AM, Friday;
Automated Fuzzy Bidding Strategy 175
• number of stops: 1;
• seat position: window;
• travel season: April (off-peak season).
Suppose the identified perspective of an agent is summarized as below:
• The agent prefers to a cheaper ticket and agrees to that the cheaper the
better.
• The agent prefers to travel at the weekend rather than at working day.
• The agent prefers to no stop travel.
• The agent prefers to aisle seat then window seat.
• The agent prefers to travel during off-peak season rather than peak season.
• The agent thinks the flight price is the most important factor, secondly
the travel season, and other factors are of same importance.
Based on the agent’s perspective, the agent evaluates the attitudes using
seven terms, very bad (a1 ), bad (a2 ), slightly bad (a3 ), acceptable (a4 ), fairly
good (a5 ), good (a6 ), and very good (a7 ).The seven terms are expressed by
fuzzy sets on the satisfactory degree [0,1] as below
Attribute Assessment
c0 (no assessment)
c1 good (a6 )
c2 fairly good (a5 )
c3 slightly bad (a3 )
c4 acceptable (a4 )
c5 good (a6 )
176 M. Goyal, S. Kaushik, and P. Kaur
Attitude
Ass. e1 e2 e3 e4 e5
a1 0.17 0.23 0.2 0.27 0.13
a2 0.1 0.28 0.22 0.26 0.13
a3 0.1 0.26 0.18 0.32 0.13
a4 0.17 0.26 0.23 0.23 0.12
a5 0.12 0.25 0.27 0.21 0.16
a6 0.12 0.26 0.26 0.23 0.13
a7 0.12 0.24 0.31 0.24 0.1
the agent uses five terms to distinguish the attitude, i.e., none (e1 ), slightly
(e2 ), medium(e3 ), strong (e4 ), and very strong (e5 ). In order to estimate the
agent’s attitude, a set of rules are extracted from a historical auction records,
which are illustrated in Table 3 and Fig. 4.
By Fig. 4, the agent’s attitudes at this moment are e2 and e3 because they
have the highest reliability. Because e3 is stronger than e2 , the agent first
searches possible bids under the attitude e3 .
For finding the price of the good, we will apply fuzzy logic by considering
two factors attitude and competition as described in the section 2.4. Let us
consider the following set of rules for the logic using various fuzzy sets.
These fuzzy sets represents the linguistic variables as follows: attitudes low
as E1 and high as E2 , Competition less as C1 and more as C2 and Neg-
ative displacement as P1 , no displacement as P2 and positive displace-
ment as P3 . We assume that the set of attitudes for buying any item as
E = {e1 , e2 , e3 } = {0, 0.5, 1} and set of competition for the good in the
market as C = {c1 , c2 , c3 } = {0, 0.5, 1}. Also, the bid displacement as
P = {p1 , p2 , p3 } = {−100, 0, +100}. The fuzzy sets used in the preced-
ing four rules can be quantized as shown in the Fig. 5:
E1 = [1.0, 0.5, 0] C1 = [1.0, 0.5, 0] P1 = [1.0, 0, 0]
E2 = [0, 0.5, 1.0] C2 = [0, 0.5, 1.0] P2 = [0, 1.0, 0]
P 3 = [0, 0, 1.0]
Note that the number of elements in E, C and P are three and the fuzzy
sets are also quantized into three elements. Now let us construct fuzzy rela-
tions by Mamdani’s Method for fuzzy relations
By the preceding conversion formula we get the fuzzy relation R1 from the
first rule as in Fig. 6.
Similarly we can convert Rules 2, 3 and 4 into fuzzy relations R2 , R3 and
R4 accordingly. The total fuzzy relation R is given using Mamdani’s Method
for compilation of fuzzy relations as shown in Fig.7 by
R = R1 ∪ R2 ∪ R3 ∪ R4
Automated Fuzzy Bidding Strategy 179
Let attitude of agent for buying the goods is high i.e. 1 and the competition
in the market for that product is more i.e. 1. Such a situation can be described
by fuzzy sets E’ and C’ as E = [0, 0, 1] C = [0, 0, 1]. Now the conclusion of
the reasoning can be calculated by applying Mamdani’s compositional rule
of Inference [3] as follows P = C ◦ (E ◦ R) where ◦ is the composition
process. After implementing this we will get P = [0, 0, 1]. Defuzzification of
P’ by taking center of gravity with the weighted mean we get definite value
for the bid displacement factor P as +100. So the final bid price will be
P = P + P .
4 Conclusions
In this paper we have designed a fuzzy competition and attitude based bid-
ding strategy (FCA-Bid), which uses a soft computing method i.e. fuzzy logic
technique to compute the final bid price based on the attitude of the agent
and the competition in the market. Another unique idea presented in this pa-
per is that to deal quantitatively the imprecision or uncertainty of multiple
attributes of items to acquire in auctions, fuzzy set technique is used. The
bidding strategy also allows for flexible heuristics both for the overall gain
and for individual attribute evaluations. Specifically, the bidding strategy is
adaptive to the environment as the agent can change the bid amount based
its assessment of the various attributes of item, eagerness of agent as well as
competition in the auction . The attitude of the agents is found with respect
to the goods’ attributes and the competition is calculated based on the num-
ber of bidders and the time elapsed for the auction. It was noticed that the
strategies in which agent’s behavior depends on attitudes and competition,
are easily adaptable to the dynamic situations of the market [11] [16]. An
experimental evaluation is also conducted to find the final bid price on the
bases of the historical auction records and by considering some set of rules
for the attitude and the competition factors. In future we will investigate
about the development of the bidding strategies for multiple auctions. We
will also compare our bidding techniques with the other strategies to find out
the relative strengths and the weaknesses.
References
1. Anthony, P., Jennings, N.R.: Developing a Bidding Agent for Multiple Het-
erogeneous Auctions. ACM transactions on Internet Technology 3(3), 185–217
(2003)
2. Anthony, P., Jennings, N.R.: Evolving Bidding Strategies for Multiple Auctions.
In: Proceedings of the 15th European Conference on Artificial Intelligence, pp.
187–192 (2002)
3. Bapna, R., Jank, W., Shmueli, G.: Price Formation and its Dynamics in Online
Auctions. Decision Support Systems 44(3), 641–656 (2008)
180 M. Goyal, S. Kaushik, and P. Kaur
4. Byde, A., Priest, C., Jennings, N.R.: Decision Procedures for Multiple Auctions.
In: Proceedings of the First International Joint Conference on Autonomous
agents and Multiagent Systems, pp. 613–620 (2002)
5. Faratin, P., Sierra, C., Jennings, N.R.: Negotiations Decision Functions for
Autonomous Agents. International Journal of Robotics and Autonomous Sys-
tems 24(3), 159–182 (1998)
6. Goyal, M.L., Ma, J.: Using Agents Attitudes and Assessments in Automated
Fuzzy Bidding Strategy. In: Proceedings of the 1st International Conference on
Agents and Artificial Intelligence (2009)
7. Greenwald, A., Stone, P.: Autonomous Bidding Agents in the Trading Agent
Competition. IEEE Internet Computing 5(2), 52–60 (2001)
8. He, M., Leung, H.f., Jennings, N.R.: A Fuzzy-Logic Based Bidding Strategy
for Autonomous Agents in Continuous Double Auctions. IEEE Transaction on
Knowledge and Data Engineering 15(6), 1345–1363 (2003)
9. Kowalczyk, R., Bui, V.: On Fuzzy e-Negotiation Agents: Autonomous Negoti-
ation with Incomplete and Imprecise Information. In: Proceedings of the 11th
International Workshop on Database and Expert Systems Applications, p. 1034
(2000)
10. Luo, X., Jennings, N.R., Shadbolt, N., Leung, H.F., Lee, J.H.M.: A Fuzzy Con-
straint based Model for Bilateral, Multi-issue Negotiations in Semi-competitive
Environments. Artifical Intelligence 148(1-2), 53–102 (2003)
11. Ma, H., Leung, H.F.: An Adaptive Attitude Bidding Strategy for Agents in
Continuous Double Auctions. Electronic Commerce Research and Applica-
tions 6(4), 383–398 (2007)
12. Ma, H., Leung, H.f.: Bidding Strategies in Agent-Based Continuous Double
Auctions. Birkhäuser, Basel (2008)
13. Murugesan, S.: Negotiation by Software Agents in Electronic Marketplace. In:
Proceedings of the 2000 IEEE Region 10 Conference on Intelligent Systems and
Technologies for the new Millenium, pp. 286–290 (2000)
14. Sim, K.M., Wang, S.Y.: Flexible Negotiation Agent With Relaxed Decision
Rules. IEEE Transactions on Systems, Man and Cybernetics- Part B: Cyber-
netics 34(3), 1602–1608 (2004)
15. Stone, P., Littman, M.L., Singh, S., Kearns, M.: ATTac-2000: an Adaptive
Autonomous Bidding Agent. Journal of Artificial Intelligence Research 15, 189–
206 (2001)
16. Tagra, H.: Generic Architecture for Agents in E-Commerce (Auction), Ph. D
dissertation. Indian Institute of Technology New Delhi (2006)
17. Tanaka, K.: An Introduction to Fuzzy Logic for Practical Applications.
Springer, Heidelberg (1991)
Resource Allocation Analysis in
Perfectly Competitive Virtual Market
with Demand Constraints of
Consumers
1 Introduction
Recently, it is difficult to manage and control artificial systems with
conventional approaches because these systems become large-scale and com-
plex. We focus on the concept of market in social science approach as a
technique to achieve the adaptability of artificial systems as well as the opti-
mality [1]. In perfectly competitive market defined by the neo-classical eco-
nomics, it is guaranteed that the equilibrium resource allocation is Pareto
optimal [2]. Market-Oriented programming (MOP) is the technique that
obtains the Pareto optimal solution by the construction of a virtual per-
fectly competitive market (virtual market) in the computer [3] [4]. MOP
has been also extended into oligopolistic virtual market in our previous
study [5].
Tetsuya Matsuda · Toshiya Kaihara · Nobutada Fujii
Kobe University, 1-1, Rokkodai, Nada, Kobe 657-8501, Japan
e-mail: matsuda@kaede.cs.kobe-u.ac.jp,kaihara@kobe-u.ac.jp,
nfujii@phoenix.kobe-u.ac.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 181–200.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
182 T. Matsuda, T. Kaihara, and N. Fujii
pi F price of good i
Gnum F number of kind of goods in the market
Snum F number of kind of producers in the market
Cnum F number of kind of consumers in the market
I Ck F class of goods in the market
C
ei k F amount of consumer Ck ’s initial good i
C
xi k F demand to good i of consumer Ck
Ck
yj F supply to good j of consumer Ck
B Ck F budget of consumer Ck
uCk F utility function of consumer Ck
x̄C
l
k
F demand limit to good l of consumer Ck
Sk
γ F profit of producer Sk
cSk F cost of producer Sk
r Sk F income of producer Sk
S
xi k F demand to good i of producer Sk
Sk
yj F supply to good i of producer Sk
f Sk F production function of producer Sk
3 Agent Formulations
3.1 Consumer Agents
3.1.1 Utility Function
The utility function is defined as a scalar function of a bundle of consuming
goods. In this model, the Cobb-Douglass utility function uCk , that is basic
function in Microeconomics, for consumer Ck is represented as follows:
+ C Ck C
uCk = ACk (xi k )αi (0 < ACk , αi k = 1) (1)
i i∈I Ck
3.1.2 Constraints
Budget constraint
B Ck = p · e Ck + γ Ck (2)
Therefore the constraint of budget for consumer Ck is represented as the
following equation:
pi xi ≤ B Ck (3)
i∈I Ck
184 T. Matsuda, T. Kaihara, and N. Fujii
Demand constraint
xC
i
k
≤ x̄i Ck (i ∈ I Ck ) (4)
+ Ck
max uCk = ACk (xC k αi
i ) (5)
i∈I Ck
s.t. pi xi = B Ck (6)
i∈I Ck
xCi
k
≤ x̄C
i
k
(i ∈ I Ck ) (7)
xC
i
k
= x̄C
i
k
(i ∈ θCk ) (8)
αC k
i (B
Ck
− pj xCk
j )
j∈θ Ck
orxC k
i (pi ) = (i ∈
/ θ Ck ) (9)
pn (1 − αC k
j )
j∈θ Ck
Resource Allocation Analysis in Perfectly Competitive Virtual Market 185
The set θCk includes the exceeded goods over x̄i Ck . At initial conditions we
assume θCk = {φ}. The demand of goods i in xC i
k
≥ x̄C
i
k
(Fig.1) is defined
Ck
as equation (8) and add goods i to θ . On the other hand, the demand of
goods i in xC
i
k
< x̄C
i
k
(Fig.2) is defined as equation (9).
In this model, we assume each consumer supplies all its endowments into
the market to try to attain maximum utility. So the supply function yjCk is
represented as follows:
yjCk = eC
j
k
(10)
In this equation xSi k and yjSk represent input goods i by Sk and output
goods j, respectively. The power variable αSk must be below 1 to satisfy the
fundamental theorem of welfare economics defined by micro economics, and
that means production function must follow the law of diminishing return [2].
γ S k = r S k − cS k (14)
pi 1
xSi k (pi ) = ( ) αSk −1 (17)
ASk αSk pj
S
(pi )α k −1 1
yjSk (pj ) = ( S S αSk
αSk )
αSk −1 (18)
A (α ) (pj )
k k
STEP5
Producers bid supply functions.
STEP6
The auctioneer of goods set up the price corresponded with the supply
and demand.
STEP7
We define the previous price of goods i is pbefi
ore
, and updated current price
af ter
is pi . If all goods satisfied the equation (19), then paf
i
ter
is equilibrium
price and go STEP8( σ is very small number. In this model, σ = 10−13 ).
If we cannot set up equilibrium price, return to STEP2.
|paf
i
ter
− pbef
i
ore
|<σ (19)
STEP8
Prices of all the goods are cleared, and all agents start trading.
The demand function of consumers (in STEP 2) followed by equations (9),
(8) is finally represented as Fig.3. The maximum demand value faces the
ceiling at x̄C
i
k
shown in this figure.
(Input)
C C C C C C
Agent Utility function (uCk ) (e1 k , e2 k , e3 k ) (x̄1 k , x̄2 k , x̄3 k ) Utility value
C1 0.7 C1 0.2 C1 0.1
C1 (x1 ) (x2 ) (x3 ) (5.0, 8.0, 10.0) (∞, ∞, ∞) 5.8870
2 0.1 C2 0.1 C2 0.8
C2 (xC 1 ) (x 2 ) (x 3 ) (8.0,10.0,5.0) (∞, ∞, ∞) 5.6168
(Results)
Price Allocation Utility value
(pˆ1 , pˆ2 , pˆ3 ) (xC 1 C1 C1
1 , x2 , x3 ) (xC2 C2 C2
1 , x2 , x3 ) C1 C2
(1.4519,0.38773,1.3430) (11.470,12.272,1.7715) (1.5296,5.7277,13.229) 9.6452 9.7249
(Input)
Agent Utility function (uCk ) (eC k Ck Ck Ck Ck Ck
1 , e2 , e3 ) (x̄1 , x̄2 , x̄3 ) Utility value
C1 0.7 C1 0.2 C1 0.1
C1 (x1 ) (x2 ) (x3 ) (5.0, 8.0, 10.0) (9, ∞, ∞) 5.8870
2 0.1 C2 0.1 C2 0.8
C2 (xC 1 ) (x 2 ) (x 3 ) (8.0,10.0,5.0) (∞, ∞, ∞) 5.6168
(Results)
Price Allocation Utility value
(pˆ1 , pˆ2 , pˆ3 ) (xC
1
1
, x C1
2 , x C1
3 ) (x C2
1 , x C2
2 , x C2
3 ) C1 C2
(0.51865,0.87771,1.5639) (9.0000,15.636,4.3878) (4.0000,2.3636,10.612) 9.3545 8.4702
in Case B.
Consumer agent C1 bids into goods 1 by the constraint limit x̄C
1 at maxi-
1
C1
mum, so the allocation of x1 in Table 2 (Results) is 9.0000 and less than in
Table 1 (Results). Consequently, the surplus budget in agent C1 was trans-
ferred into the bids for both goods B and C. As a result, xC C1
2 , x3 in Table 2
1
Resource Allocation Analysis in Perfectly Competitive Virtual Market 189
(Input)
Agent Utility function (uCk ) (eC k Ck Ck Ck Ck Ck
1 , e2 , e3 ) (x̄1 , x̄2 , x̄3 ) Utility value
C1 0.7 C1 0.2 C1 0.1
C1 (x1 ) (x2 ) (x3 ) (5.0, 8.0, 10.0) (9, 13, ∞) 5.8870
2 0.1 C2 0.1 C2 0.8
C2 (xC 1 ) (x 2 ) (x 3 ) (8.0,10.0,5.0) (∞, ∞, ∞) 5.6168
(Results)
Price Allocation Utility value
(pˆ1 , pˆ2 , pˆ3 ) (xC
1
1
, x C1
2 , x C1
3 ) (x C2
1 , x C2
2 , x C2
3 ) C1 C2
(0.46278,0.37022,2.2213) (9.0000,13.000,8.3333) (4.0000,5.0000,6.6667) 9.6126 6.5093
(Input)
C C C C C C
Agent Utility function (uCk ) (e1 k , e2 k , e3 k ) (x̄1 k , x̄2 k , x̄3 k ) Utility value
C1 0.7 C1 0.2 C1 0.1
C1 (x1 ) (x2 ) (x3 ) (5.0, 8.0, 10.0) (9, 10, ∞) 5.8870
2 0.1 2 0.1 2 0.8
C2 (xC1 ) (xC
2 ) (xC
3 ) (8.0,10.0,5.0) (∞, ∞, ∞) 5.6168
(Results)
Price Allocation Utility value
(pˆ1 , pˆ2 , pˆ3 ) (xC1 C1 C1
1 , x2 , x3 ) (xC2 C2 C2
1 , x2 , x3 ) C1 C2
(0.44660,0.22330,2.4117) (9.0000,10.000,9.0741) (4.0000,8.0000,5.9259) 9.1992 6.3670
The difference between demand x and supply y is called excess demand, and
excess demand function E(p) is shown in equation (20).
E(p) = x − y (20)
Scarf algorithm
Scarf algorithm is one of the most basic fixed point algorithm. First, normalize
the number i which is kinds of goods and make price set T ((I-1)dimensional
basic elementary substanceSI ). Next, we divide each side of SI equally and
make small elementary substance which has I vertex. The number D is di-
viding number and called grid size. If D is large, solution is high-precision.
In Scarf algorithm, we search solution from the small elementary substance
which has the vertex of SI to next small elementary substance.
(p1 , p2 , p3 ) Time
(a) (0.45619492,0.12182478,0.42198030) 0.007
(b) (0.17520216,0.29649596,0.52830189) 0.029
(c) (0.15151515,0.12121212,0.72727273) 0.037
(d) (0.14492754,0.07246377,0.78260870) 0.031
192 T. Matsuda, T. Kaihara, and N. Fujii
[Experimental results]
Fig. 5 show allocations, utility of consumer, profit of producer and equi-
librium prices of goods 1 and goods 2 (pˆ1 , pˆ2 ) . Transverse is x¯2 C1 .
Goods 2 is transformed from goods 1 to maximize utility and profit. Be-
cause the consumer agent prefers goods 2 to goods 1, the producer agent
transforms from goods 1 to goods 2. Fig.5 (a) describes that the amount of
allocation of goods 2 increases as the demand constraint is larger, and goods 1
that is input goods decreases because consumer agent can be allocated goods
2 less or equal of demand constraint. The amount of the allocation doesn’t
change even if the demand constraint is increased more than 5 in Fig.5 (a).
That is because the consumer agent can get more utility from goods 2 than
from goods 1 in this case due to the law of diminishing marginal utility.
Fig.5 (b) tells us that the price of goods 1(/goods 2) decreases(/increases)
as the demand constraint increases. That is because the demand of goods 2
of consumer agent is high in case the demand constraint is large. Demand
of goods 1 of consumer agent is decreased in case the demand constraint is
Resource Allocation Analysis in Perfectly Competitive Virtual Market 193
(c)Utility (d)Profit
large, instead. Fig.5 (c) tells us that the utility of consumer agent increases
because the number of allocated goods 2 is getting higher due to the increased
production. Fig.5 (d) describes that the profit of producer agent is increased
because the producer agent can produce more volume of goods 2.
Degree of preference(αC1 C1
1 , α2 ) Experimental result
(0.8,0.2) (a)
(0.6,0.4) (b)
(0.5,0.5) (c)
(0.3,0.7) (d)
(0.1,0.9) (e)
194 T. Matsuda, T. Kaihara, and N. Fujii
[Experimental results]
Fig. 6 shows the allocations of each goods, and Fig. 7 shows the equilibrium
prices (pˆ1 , pˆ2 ). Fig. 8 and Fig. 9 show the utility of consumer agents and the
profit of producer agents, respectively. Horizontal axis in each figure shows
the upper constraint of consumer agent 1 (x¯1 C1 ).
(a) (b)
(c) (d)
(e)
Fig. 6 Allocation
Resource Allocation Analysis in Perfectly Competitive Virtual Market 195
(a) (b)
(c) (d)
(e)
Fig. 7 Price
allocations of each goods (Fig. 6 ), the price of each goods (Fig. 7), and the
profit of each producder (Fig. 9) are getting smaller, because the utility of
196 T. Matsuda, T. Kaihara, and N. Fujii
(a) (b)
(c) (d)
(e)
Fig. 8 Utility
(c)), the differences between the allocations of each goods, the price of each
goods, and the profit of each producder become zero, because the consumer
agents utility of each goods is completely the same.
(a) (b)
(c) (d)
(e)
Fig. 9 Profit
[Experimental results]
The average of calculation time in 100 trials are shown in Table 12.
198 T. Matsuda, T. Kaihara, and N. Fujii
Cnum Time
10 0.04044
50 0.09550
100 12.14
Finally we show one sample in case of Cnum = 10. Table 13 and Table 14
are both in its initial states, and Table 15 and Table 16 describe equilibrium
prices acquired after the convergence.
Table 15 Allocation
Agent xC
1
k
xC
2
k
xC
3
k
It has been confrimed that all of the consumer’s allocation are under their
demand constraints in Table 15. Additionally it has been also clarified that
the calculation time for Pareto optimal solution increases as the number of
consumer agents increases in Table 12. However, the calculation time is within
13 seconds in the biggest model, and that means the calculation time is fully
enough for the practical use to attain a Pareto optimal solution.
5 Conclusions
We formulated a perfectly competitive virtual market with demand con-
straints of consumers, and demonstrated its efficiency on resource allocation
problem in this paper. Firstly we mentioned our motivation about the re-
source allocation problem with the demand constraints, and explained the
agent formulation in the virtual market. Then market-oriented programming
mechanism with the demand constraints were explained. After a brief expla-
nation of a conventional analytical approach, named Scarf algorithm, several
simulation experiments were carried out to analyse the proposed approach.
Finally we have obtained the following results:
• It has been confirmed that the proposed perfectly competitive virtual
market with demand constraints successfully conducts Pareto optimal
solutions.
• After the investigation about the effects of demand constraint, rational
resource allocation have been realised by the proposed virtual market.
• The effect of the number of agents have clarified the proposed approach is
useful for practical problems both in optimality and calculation time.
The obvious next targets of this research are as follows:
• Application of the proposed approach into a practical problem with de-
mand constraint
• Modelling extension to handle resource allocation with demand constraints
both of consumers and producers
References
1. Kaihara, T., Fujii, S.: Towards intelligent artificial systems with adaptive be-
haviour - Sociological approach, Manufacturing Systems Division Conference
2003, pp. 61–62 (2003)
2. Arrow, K.J., Debreu, G.: Existence of an equilibrium for a competitive economy.
Econometrica 22(3), 265–290 (1954)
200 T. Matsuda, T. Kaihara, and N. Fujii
1 Introduction
Recently, it has become clear that it is difficult to explain all of the phenomena
observed in markets by established market models based only on conventional
human psychological reasoning. On the other hand, the study of artificial
markets is one of the promising approaches to explaining such phenomena.
An artificial market is an agent-based model of a financial market. Studies
on artificial markets have attained some successes in market analysis in re-
cent years. For example, the Santa Fe Artificial Stock Market could simulate
financial bubbles with agents using genetic algorithms [1, 10]. Using an ar-
tificial market, Lux explained the fat-tailed distribution of financial returns,
one of the characteristic styles of financial markets [9]. Artificial markets
have also been used to test existing economic theories such as efficient mar-
ket hypotheses [4]. In a market design using an artificial market, Darley and
Fujio Toriumi
Graduate School of Information Science, Nagoya University
e-mail: tori@is.nagoya-u.ac.jp
Kiyoshi Izumi
Digital HumanResearchCenter, AIST
Hiroki Matsui
CMD Laboratory, Inc
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 201–215.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
202 F. Toriumi, K. Izumi, and H. Matsui
Outkin examined the effect of tick-size reduction in the Nasdaq market [6].
From theoretical to empirical studies, a variety of results have been obtained
by artificial markets over the past ten years [2, 3, 5].
One of the most important requirements in designing artificial markets is
to create a market that is similar to the real market. If an artificial market has
no resemblance to a real market, the results derived from such an artificial
market would be unreliable. In particular, the combination of agents deter-
mines the reliability of an artificial market. Accordingly, when the agents’
behaviors in the artificial market are highly similar to the behavior of traders
in the real market, the characteristics of the artificial market become similar
to those of the real market.
The combination of agents can thus be considered a key parameter of
artificial markets. Therefore, the agent combination problem can be consid-
ered a parameter estimation problem. “Inverse Simulation” [8] is one of the
parameter estimation methods used for social simulation, and here we pro-
pose a market-participant estimation method using the inverse simulation
technique.
2 Artificial Market
2.1 Artificial Market Framework
An artificial market (e.g. [5]) is a virtual financial market run on a computer.
Agents, which are computer programs that play the role of virtual dealers,
participate in artificial markets.
The artificial market consists of the following components (Fig.1).
1. Market mechanism module
2. Order protocol
3. Agent interface
4. Agents
2.4 Agents
A virtual dealer that participates in an artificial market is called an “Agent.”
An agent is the computer program that buys and sells virtual stocks in the
artificial market.
Each agent receives such market information as past stock prices, trading
volumes, the results of its own orders, and so on. From this information, the
agent determines an order based on its trading rules. The order is sent to the
artificial market through the agent interface.
4 Trader Agents
In this simulation, we used two types of agents: Fundamentalists and Chartists.
Each type of agent has different parameters and trade strategies. The details
of each agent are described in this section.
Market Participant Estimation by Using Artificial Market 207
4.3.2 Activeness
Activeness A is a parameter that defines the activeness of the agents. When
an active agent estimates the upper trend, the agent orders with a higher
price limit than the current price.
The parameter Activeness takes one of two values, “active” or “passive.”
An active agent orders at a higher price than the current price for a buying
order and at a lower price than the current price for a selling order. On the
other hand, a passive agent orders at the current price for both buying and
selling orders.
Table 1 shows each parameter and the values that each agent can take. Fun-
damentalist agents use the parameters Trade Frequency, Belief, Trade Type,
Trade Quantity, and Chartist agents use the parameters Trade Frequency,
Belief, TrendBehavior, Measure Span.
Market Participant Estimation by Using Artificial Market 209
Fundamentalist Chartists
Trade Frequency 1,5,10,30
Activeness Active,Passive
Trade Type Buy,Sell -
Target Quantity 5,30,60,200 -
Trend Behavior - Followers/Contrarians
Measure Span - 10,60
Population Size 50
Generations 100
Length of gene 400
Selection Leave the highest-fitness individual
Select 5% by roulette selection
Crossover One-point crossover
Mutation rate 10%
Figure 6 shows the sample of price movement of the artificial market and
the real market on April 1, 2009. From the figure, the simulation price rises
after once dropping down, which also happened in the real market. As indi-
cated above, the simulation prices follow the rough shape of the real price
movement.
Furthermore, the active agents make more orders at a price that has a
higher chance of trades being executed. Therefore, a low rate of active agents
causes a smaller amount of trading volume in the market. In fact, there was
actually a small amount of trading volume in the real market on the simulated
day. This fact also backs up the validity of the simulation results.
From the above results, it is estimated that the types of agents in the
artificial market that simulates a market under special circumstances are
different from those in other markets. This observation suggests that it is
possible to estimate the market participants by using the proposed method.
Market Participant Estimation by Using Artificial Market 213
6 Conclusion
In this paper, we proposed a method to estimate market participants by
inverse simulation. We applied the proposed method to markets after the
worldwide recession caused by the subprime financial meltdown. The simula-
tion results indicate that the types of agents in the market under such special
circumstances are different from those in other markets.
However, there are several points that merit further consideration in this
simulation. The first is the sufficiency of the agent types. We used three types
of agents in the simulation; however, it has not been verified whether these
agent types cover all types of traders. There might be traders who trades
using completely different algorithms. To ability to generate enough types of
agents to fully represent a real market is one of the future works of this study.
A second point is the adequacy of the evaluation function. In this paper,
we use the similarity of price movements as the evaluation function. However,
price movement does not reflect all characteristics of a market. For example,
the standard deviation, ARCH coefficient, Skewness and Kurtosis have been
used to evaluate the price movement of stocks, and a board is also expected to
evaluate the market. Developing an evaluation function that can more fully
express market characteristics is also a future work of this study.
References
1. Arthur, W., Holland, J., LeBaron, B., Palmer, R., Tayler, P.: Asset pricing
under endogenous expectations in an artificial stock market. In: Arthur, B.,
Durlauf, S., Lane, D. (eds.) The Economy as an Evolving Complex System II,
pp. 15–44. Addison-Wesley, Reading (1997)
2. Chen, S.-H., Wang, P.P. (eds.): Computational Intelligence in Economics and
Finance. Springer, Heidelberg (2003)
3. Chen, S.-H., Wang, P.P., Kuo, T.-W. (eds.): Computational Intelligence in Eco-
nomics and Finance: Volume II. Springer, Heidelberg (2007)
4. Chen, S.-H., Yeh, C.-H.: On the emergent properties of artificial stock mar-
kets: the efficient market hypothesis and the rational expectations hypothesis.
Journal of Economic Behavior & Organization 49(2), 217–239 (2002)
5. Consiglio, A. (ed.): Artificial Markets Modelling: Methods and Applications.
Springer, Heidelberg (2007)
6. Darley, V., Outkin, A.: A NASDAQ Market Simulation. World Scientific, Sin-
gapore (2007)
7. FIX Protocol Ltd. The FIX Guide: Implementing the FIX Protocol. Xlibris
(2005)
8. Kurahashi, S., Minami, U., Terano, T.: Why not multiple solutions: Agent-
based social interaction analysis via inverse simulation. In: IEEE International
Conference on System, Man, and Cybernetics, page 2048 (1999)
9. Lux, T.: The socio-economic dynamics of speculative markets: Interacting
agents, chaos, and the fat tails of return distributions. Journal of Economic
Behavior & Organization 33, 143–165 (1998)
10. Palmer, R.G., Arthur, W.B., Holland, J.H., LeBaron, B., Tayler, P.: Artificial
economic life: A simple model of a stock market. Physica D 75, 264–274 (1994)
214 F. Toriumi, K. Izumi, and H. Matsui
Trend := Upward
else if p(t) <LB then
Trend := Downward
end if
3. RSI Strategy
n := 6, 30, 60
gain := 0
loss := 0
i := 0
while i < n do
dp := p(t − i) − p(t − (i + 1))
if dp > 0 then
gain := gain + |dp|
else if dp < 0 then
loss := loss + |dp|
end if
i := i + 1
end while
if gain/(gain + loss) > 0.75 then
Trend := Upward
else if gain/(gain + loss) < 0.25 then
Trend := Downward
end if
4. BollingerBand Strategy
n := 6, 30, 60
M A(t) := Moving Average from t − n to t − 1
V (t) := Standard Deviation from t − n to t − 1
if p(t) > M A(t) + V (t) then
Trend := Upward
else if p(t) < M A(t) − V (t) then
Trend := Downward
end if
A Study on the Market Impact of
Short-Selling Regulation Using
Artificial Markets
Abstract. Since the subprime mortgage crisis in the United Sates, stock
markets around the world have crashed, revealing their instability. To stem
the decline in stock prices, short-selling regulations have been implemented
in many markets. However, their effectiveness remains unclear. In this paper,
we discuss the effectiveness of short-selling regulation using artificial markets.
An artificial market is an agent-based model of financial markets. We con-
structed an artificial market that allows short-selling and an artificial market
with short-selling regulation and have observed the stock prices in both of
these markets. We found that the market in which short-selling was allowed
was more stable than the market with short-selling regulation, and a bub-
ble emerged in the regulated market. We evaluated the values of assets of
agents who used three trading strategies, specifically, these agents were fun-
damentalists, chartists, and noise traders. The fundamentalists had the best
performance among the three types of agents. Finally, we observe the price
variations when the market price affects the theoretical price.
Isao Yagi
Interdisciplinary Graduate School of Science and Engineering,
Tokyo Institute of Technology, J2-52, 4259, Nagatsuda-Cho, Midori-ku,
Yokohama 226-8502, Japan
e-mail: iyagi2005@gmail.com
Takanobu Mizuta
SPARX Asset Management Co., Ltd., Gate City Ohsaki,
East Tower 16F 1-11-2 Ohsaki, Shinagawa-ku, Tokyo 141-0032, Japan
e-mail: takanobu.mizuta@sparxgroup.com
Kiyoshi Izumi
Digital Human Research Center, National Institute of Advanced Industrial Science
and Technology, 2-3-26, Aomi, Koto-ku, Tokyo 135-0064, Japan
e-mail: kiyoshi@ni.mints.ne.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 217–231.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
218 I. Yagi, T. Mizuta, and K. Izumi
1 Introduction
As a consequence of the subprime mortgage crisis and the Lehman Brothers
shock in the United States, stock markets around the world have crashed,
revealing their instability. To stem the decline in stock price, short-selling
regulations were implemented in a number of markets. Generally, these reg-
ulations are thought to inhibit the decline of stock prices. The effectiveness
of such regulations has been evaluated in previous studies [2, 9]. A report on
short-selling regulation in the United Kingdom, United States, and so on did
not consider the market mechanisms for the case in which short selling is
restricted [8].
An artificial market that is an agent-based model of financial markets
is useful to observe the market mechanism. That is, it is effective for ana-
lyzing causal relationship between the behaviors of market participants and
the transition of market price. The behavior of an artificial market can be
observed when useful models of actual market participants are introduced
to the market. Moreover, when institution models to stem market declines,
such as short-selling regulation, are introduced to the market, we can observe
the behavior of the market participants. In addition, we can observe market
phenomena on which the above participant behaviors have a great effect. A
number of studies have investigated phenomena in financial markets using
artificial markets [1, 3, 4, 6, 7].
In this paper, we discuss the effectiveness of short-selling regulation us-
ing an artificial market. For simplicity, in Sect. 2, we constructed an artifi-
cial market with short-selling regulation and an artificial market that allows
short-selling. We hereinafter refer to the market with short-selling regulation
as the regulated market and to the market that allows short-selling as the un-
regulated market. The trading strategy models of these market participants
(agents) are defined in Sect. 2.1 and Sect. 2.2. In Sect. 3, we analyze the
market mechanisms based on the results of the artificial markets simulations.
First, in Sect. 3.1, we observe the stock price variations of two markets. The
unregulated market was found to be more stable than the regulated market,
and a stock market bubble emerged in the regulated market. By comparing
our proposed market to past actual market data (TOPIX), we demonstrate
that our artificial market has some properties of actual markets. Next, we
observe agent transactions in Sect. 3.2. We analyze how agent transactions
influence our two markets and why a bubble emerges in the regulated mar-
ket. Then, we discuss the performances of agents in our two markets. The
proposed market considers agents with three trading strategies: fundamen-
talists, chartists and noise traders. We demonstrate that fundamentalist have
the best performance among the three agent types. The theoretical price in
above market model does not depend on the market price. In Sect. 3.3, we
observe price variations under the market model in which the market price
affects the theoretical price.
A Study on the Market Impact of Short-Selling Regulation 219
2.1.1 Fundamentalist
In our markets, fundamentalists predict current market price based on the-
oretical price and control the number of shares that they hold in order to
maximize their asset values. Theoretical price is a given value and is read-
justed periodically. The initial theoretical price is set to 300 and readjusted
according to a normal distribution N (1, 0.12 ) every 1,000 periods.
Let the expected stock price of agent i at period t be P̃i,t , and
let the theoretical price be Pt . Then, P̃i,t is decided according to
N (Pt + i,t , (α(Pt + i,t ))2 ) and readjusted every 100 periods, where i,t de-
notes the degree of bullishness2 of i at t, and α is a coefficient that describes
the degree of spread of expected prices, which is decided by each agent, and
is set to 0.1. Qi,t−1 denotes the amount of cash that i holds before a transac-
tion at t. qi,t−1 denotes the amount of shares that i holds before a transaction
at t. Pt−1 denotes the market price at t − 1. The assets of agent i before a
transaction at t are as follows:
Let the amount of shares held by i at t that maximizes the subjective expected
utility under condition (1) be q̃i,t . Then, q̃i,t is as follows:
Pt + i,t − Pt−1
q̃i,t = .
a(α(Pt + i,t ))2
Note that a constant a(> 0) is a coefficient of risk aversion and the larger
the value of a is, the smaller the number of shares that fundamentalists hold
to averse risk. Fundamentalists attempt to decide their trading rules based
on q̃i,t . If q̃i,t > qi,t−1 , i places an order to buy q̃i,t − qi,t−1 shares at P̃i,t ,
where the maximum number of shares that can be bought is Smax − qi,t−1 .
If q̃i,t < qi,t−1 , i places an order to sell qi,t−1 − q̃i,t shares at P̃i,t , where the
maximum number of shares that can be sold in the unregulated market is
qi,t−1 − Smin , and the maximum number of shares that can be sold in the
regulated market is qi,t−1 . If q̃i,t = qi,t−1 , i does not trade at period t.
2.1.2 Chartists
In our markets, chartists trade based on the moving averages of stock prices.
Chartists consist of market followers and contrarians. Let ni,t be a number
that satisfies 1 ≤ ni,t ≤ 25. Let the ni,t period moving average at t used by
agent i be
1
ni,t
M At,ni,t = Pt−j ,
ni,t j=1
2.2.1 Fundamentalists
Let the ratio of change of the assets of fundamentalist agent i from period
t − 1 to t be Ri,t = Wi,t /Wi,t−1 . When the mean ratio of change of the assets
of i in the past N (= 5) periods
1
N
Ri,t = Ri,t−(j−1)
N j=1
1
N
i ,N = i ,t−j .
N j=0
222 I. Yagi, T. Mizuta, and K. Izumi
2.2.2 Chartists
When the mean ratio of change of the assets of chartist agent i in the past
N (= 5) periods Ri,t ranks in the bottom NL (= 20)% of all chartists, the
moving average period ni,t is changed with the following probability.
3 Discussion
3.1 Price Variation in the Markets
We observed market price variation in the unregulated and regulated markets
(refer to Fig. 1 and Fig. 2). These figures reveal that prices in the unregulated
market did not diverge significantly from the theoretical price. On the other
hand, prices in the regulated market diverged greatly from the theoretical
price and showed repeated sharp increases and decreases. In the regulated
market, once the price rose sharply, it never returned to the theoretical price.
These indicate that the short-selling regulation has the property that it not
only stems the decline in the prices but also increases the prices excessively.
However, if our market models do not reflect the actual markets, the result
is meaningless. Thus, we discuss whether our market models reflect the actual
markets. The price variation statistics of our markets were first compared
with those of past market data to analyze the problem. We used the Tokyo
stock price index (TOPIX) from 16th May 1949 to 16th June 2009 (14,926
A Study on the Market Impact of Short-Selling Regulation 223
600
500
400
Price
300
200
100
0
0 2000 4000 6000 8000 10000 12000 14000
Period
600000
500000
400000
Price
300000
200000
100000
0
0 2000 4000 6000 8000 10000 12000 14000
Period
days) as past market data. Table 1 shows the parameters of the rate of return
in these markets. Fig. 3, Fig. 4, and Fig. 5 show histograms of the rate of
return in the unregulated market, the rate of return in the regulated market,
and the rate of return of the TOPIX, respectively. The distribution of the
rate of return of actual data is fat-tailed. (The kurtosis of the rate of return
distribution of the TOPIX is 11.673.) Since the kurtosis of the rate of return
distribution of the unregulated market is 128.65 and that of the regulated
market is 0.50732, the distributions of the rates of return of our markets
are also fat-tailed. The kurtosis of the unregulated market is larger than
that of the regulated market, as the standard deviation (SD) of the rate of
return in the unregulated market is smaller than that in the regulated market.
Therefore, we can consider that the price variation in the unregulated market
is more stable than that in the regulated market.
Since short-selling is seldom regulated in actual markets, a histogram of
the rate of return in the unregulated market is more similar to that in the
TOPIX than that in the regulated market. This indicates that our market
models reflect the actual markets.
224 I. Yagi, T. Mizuta, and K. Izumi
Unregulated Regulated
TOPIX
market market
Mean 0.00045 0.00317 0.00031
Standard
0.03098 0.07745 0.01064
deviation
Kurtosis 128.65 0.50732 11.673
Skewness 3.8347 0.10937 -0.13232
Fig. 3 Histogram of 9
the rate of return in the
unregulated market 8
5
Frequency
0
-4 .5%
-3 .3%
-2 .2%
-1 .6%
-1 %
.7%
14 %
22 %
29 %
35 %
41 %
47 %
52 %
57 %
62 %
67 %
.1%
1.2
6.9
.8
.2
.0
.4
.4
.1
.5
.6
.5
1
6
3
1
-6
Rate of return
Fig. 4 Histogram of 9
the rate of return in the
regulated market 8
Frequency
4
0
-3 .4%
-2 .0%
-1 .3%
-1 .2%
-4 %
.2%
%
12 %
17 %
22 %
26 %
31 %
35 %
39 %
43 %
46 %
.9%
0.5
1.6
7.2
.5
.5
.2
.8
.1
.3
.3
.2
0
2
4
7
-4
Rate of return
5
Frequency
0
Fig. 5 Histogram of
-1 .8%
-1 .8%
-9 %
-8 %
-6 %
-4 %
-2 %
-0 %
.9%
%
%
%
%
%
10 %
12 %
.1%
1.9
.9
.1
.2
.4
.6
0.8
2.5
4.2
5.8
7.4
9.0
.6
5
3
-1
0.8
0.6
Coefficient
0.4
0.2
0
0 5 10 15 20 25 30 35 40 45 50
-0.2
-0.4
Lag
0.8
Coefficient
0.6
0.4
0.2
0
0 5 10 15 20 25 30 35 40 45 50
-0.2
Lag
regulated market decreases slowly, which indicates that the rates of return
depend on past rates of return. Therefore, the volatility of returns in the
regulated market is considered to increase.
Position rate
0.05
400
Price
300
0
200
-0.05
100
-0.1 0
0 2000 4000 6000 8000 10000 12000 14000
Period
On the other hand, the positions of chartists correlate directly with market
price (0.839) in the regulated market. In addition, in the regulated market,
fundamentalists seldom trade is found. Fundamentalists sell all of their shares
upon the emergence of a bubble. However, the price is seldom lower than the
theoretical price, even if the bubble collapses. Therefore, they cannot buy
shares after the first bubble. The reason for this is not analyzed in this paper.
Volume
Price
1000
150
100
100
50 10
0 1
2301 2351 2401 2451
Period
0.5
Price
1000
0.4
0.3 100
0.2
10
0.1
0 1
2301 2351 2401 2451
Period
to place sell orders as the trend and price decrease. When the price has fallen
sufficiently, agents who were unable to place buy orders due to a cash shortage
become able to buy some shares and price ceases to decrease (refer to Fig. 9).
Fig. 10 indicates that noise traders place orders to buy and their positions
begin to increase around the time prices cease to decrease.
Regarding the mechanism of the bubble emergence, in the theory of fi-
nance, there is reported to be a high possibility of bubble emergence if short-
term investors or noise traders exist in markets [5, 10] and prices affect fun-
damental values [11].
Unregulated Regulated
market market
Fundamentalists 14.670% 0.03906%
Chartists -15.077% 187.54%
Noise traders 1.1808% -99.752%
Table 3 Return ranking of chartists and noise traders in the regulated market
fundamentalists make slight profits, chartists make large profits, and noise
traders lose most of their assets. Table 3 shows the five highest and five lowest
returns of chartists and noise traders, respectively, in the regulated market.
Not all chartists make large profits, and all noise traders lose most of their
assets. The strategies of chartists in the regulated market can be considered
to be high-risk-high-return strategies.
Therefore, in both markets, the strategies of fundamentalists have the most
stable performance among three agent types.
Pt+1 = Pt + αp (Pt − Pt )
230 I. Yagi, T. Mizuta, and K. Izumi
0.08
SD of rates of return 0.07
0.06
0.05
0.04
0.03
0.02
0.01
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
αp
Fig. 11 The SD of rates of return in the unregulated market under the new theo-
retical price model
where Pt is more than three times of Pt or less than one third of Pt . Fig. 11
shows the SD of rates of return in the unregulated market when αp ranges
from 0 to 1.0 at intervals of 0.1. These values are the arithmetic means of
various values of SD obtained from the execution of 10 simulations in the
unregulated market. The SD of the rates of return increases with increas-
ing value of αp . This means that the more market prices affect theoretical
prices, the more instable markets are. Therefore, this property bears out
Soros’ prospect [11].
In addition, we do not have the scenario in which the market price is more
than three times of the theoretical price in the unregulated market.
In the regulated market, we do not have any features under the new the-
oretical price model.
4 Conclusion
In this paper, we discussed the effectiveness of short-selling regulation using
an artificial market. We first constructed regulated and unregulated markets.
We found that the unregulated market was more stable than the regulated
market, and that a bubble emerged in the regulated market. This indicates
that the short-selling regulation has the property that it not only stems the
decline in the prices but also increases the prices excessively. To confirm the
adequacy of our market models, we compared the price variation statistics
of our markets with those of past market data (TOPIX) and found that
our market models reflected the actual markets. Next, we discussed agent
performance in the two markets and showed that the performance of funda-
mentalists was the best among the three agent types. Moreover, we discussed
the reason why a bubble emerges in our market models. Finally, we discussed
A Study on the Market Impact of Short-Selling Regulation 231
the price variations under the market model in which the market price affects
the theoretical price. In the future, we intend to clarify the reason why mar-
ket price diverges from theoretical price in the unregulated market, prospect
the price transition for the case in which the ratio of the three strategies
(fundamentalists, chartists, and noise traders) in our markets change flexi-
bly, identify whether properties of the artificial markets using other values for
each parameter (e.g., the maximum number of shares) satisfy that of the ac-
tual markets, and discuss the effectiveness of other regulations (e.g., leverage
regulation) using our market models.
References
1. Arthur, W., Holland, J., LeBaron, B., Palmer, R., Tayler, P.: Asset pricing
under endogenous expectations in an artificial stock market. In: The Economy
as an Evolving Complex System II, pp. 15–44. Addison-Wesley, Reading (1997)
2. Arturo, B., Goetzmann, W.N., Zhu, N.: Efficiency and the bear: short sales and
markets around the world, Yale ICF Working Paper No.02-45 (2004)
3. Chen, S.-H., Yeh, C.-H.: On the emergent properties of artificial stock mar-
kets: the efficient market hypothesis and the rational expectations hypothesis.
Journal of Economic Behavior & Organization 49(2), 217–239 (2002)
4. Darley, V., Outkin, A.V.: A NASDAQ Market Simulation: Insights on a Major
Market from the Science of Complex Adaptive Systems. World Scientific Pub.
Co. Inc., Singapore (2007)
5. De Long, J.B., Shleifer, A., Summers, L.H., Waldmann, R.J.: Noise Trader Risk
in Financial Markets. Journal of Political Economy 68(4), 703–738 (1990)
6. Hara, A., Nagao, T.: Construction and analysis of stock markets model using
ADG; automatically defined groups. International Journal of Computational
Intelligence and Application 2(4), 433–446 (2002)
7. Izumi, K., Toriumi, F., Matsui, H.: Evaluation of automated-trading programs
using an artificial market. NeuroComputing (forthcoming)
8. Marsh, I.W., Niemer, N.: The impact of short sales restrictions, The London
Investment Banking Association (2008)
9. Saffi, P.A.C., Sigurdsson, K.: Price efficiency and short-selling, AFA 2008 New
Orleans Meeting Paper (2007)
10. Shiller, R.J.: Bubbles, Human Judgement, and Expert Opinion. Financial An-
alysts Journal 58, 18–26 (2001)
11. Soros, G.: The Alchemy of Finance. John Wiley & Sons. Inc., Chichester (2003)
Learning a Pension Investment in
Consideration of Liability through
Business Game
1 Introduction
The company has introduced the pension system from the character of the
complement of social security and the mitigation of the burden of individ-
ual security which retirement benefits1 mean, or a personnel policy. In the
pension system, although the organization which became independent of cor-
porate management is usually performing investment and management, the
sponsoring company takes the final risk of management. In Japan, it became
clear that the shortfall to projected benefit obligation2 comes to be added up
Yasuo Yamashita · Takao Terano
Department of Computational Intelligence and Systems Science,
Tokyo Institute of Technology
e-mail: yamasita.y.aa@m.titech.ac.jp
Hiroshi Takahashi
Graduate School of Business Administration, Keio University
1
Retirement benefits say the thing of retirement lump sum grants or a business
annuity in Japan.
2
It is also called PBO. It is the current value of the Retirement benefits which a
company will pay to employees in the future.
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 233–249.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
234 Y. Yamashita, H. Takahashi, and T. Terano
as that research on a finance is done enough [1]. So far, some researches are
made regarding learning through business game as the technique of the finan-
cial education in the finance field [10] [11]. However, it cannot say that the
quantity of research is still enough, and research on learning of the pension
asset management in consideration of liability is not done, either. Therefore,
the approach of this research in consideration of liability on learning of pen-
sion asset management can be called meaningful thing. In this research, it is
shown clearly by experiencing a business game that practitioners learn the
difficult asset allocation method effectively.
In this research, in order to aim at showing the technique of learning the
pension-assets investment in consideration of liability using the framework of
business game, first, the model of a business game is built, next it experiments
by using an actual human being as a player. After explaining the model used
for analysis in the following chapter, results are shown in chapter 3. Chapter 4
is Conclusion.
2 Method
2.1 System of Business Game
The environment required for development of the system in this research,
and execution of an experiment is constituted by business model descrip-
tive language (BMDL) and business model development system (BMDS) [5].
BMDL is the programming descriptive language of a short form. HTML files,
CGI files, and so on can be created in BMDS by describing BMDL for game
managers (facilitator) and for game users (player). A business model writ-
ten in BMDL is first translated into corresponding CGI scripts and C pro-
grams by BMDL translator. The programs are then run on a host computer
with a WWW server and the model variable data in the form of spread-
sheets. Finally, players execute the business game through browsers on the
WWW environment. Fig.1 shows development and execution environment of
an experiment. Players input decision-making in each round through WWW
browsers, and a facilitator also advances a game through a WWW browser.
bond, domestic short-term bond, foreign bond, domestic stock, and foreign
stock at the beginning of every term6 , and inputs into a model as “Input
information”.
A player refers to “Reference information” and decides asset allocation.
“Reference information” is offered every term. The expected return of asset,
the risk, the covariance, the duration, and so on are displayed on “Refer-
ence information”. “Reference information” is provided also with the example
(suppose that it is called a reference portfolio in this paper) of the portfolio
used as reference which maximizes the expected return of a surplus and so
forth.
As a reference portfolio, the case where the “expected return / risk” of a
surplus7 is maximized, the case where an expected return is maximized, and
the case where a risk is minimized are raised according to the kind to optimize.
The reference portfolio of a short period (the past one year), the middle (the
past three years), and a long period of time (all the available data, in general
five years or more) is raised regarding the period of the past return used for
expected return estimation. The examples of the target duration regarding
6
One term is set to one round, and the period of one term is considered as the
period of three months.
7
A surplus here is the difference of pension assets and a pension liability. In plus,
a surplus means the savings surplus of pension assets, and in minus, it means
the unfunded liabilities of pension assets.
Learning a Pension Investment in Consideration of Liability 237
pension asset are given as for the cases where the percentage to the duration
of pension liability is 80%, 100%, and 120%8 .
As a person in charge of a pension fund, the reservation of pension assets
which provides pension liability is given to a player as a target. Although a
savings ratio should just always change into not less than 100% of state (state
of plus surplus), when a risk is taken too much in order to raise a return,
sometimes, a reserve ratio may fall. When a reserve ratio falls, a sponsoring
company compensates funds as retirement benefit costs, and a sponsoring
company receives a loss by this costs donation. Therefore, a player must de-
cide the asset allocation of pension assets, after also taking into consideration
the influence which it has on the profit and loss of a sponsoring company.
When a savings ratio becomes not less than 100%, it is a setup by which pen-
sion assets are balanced to pension liability by reducing the contribution of
the pension which is inflow to pension assets, and a reserve ratio is adjusted
to 100% at the time of every term and a start.
8
Furthermore, in a Excel file which offers “Reference information”, a program is
able to calculate a portfolio by specifying the percentage of duration and expected
returns (the short period, the middle, long period of time).
238 Y. Yamashita, H. Takahashi, and T. Terano
That is, when a reserve ratio is less than 100%, the insufficiency of pension
assets is added up as a loss of a sponsoring company. When a reserve ratio is
not less than 100%, it is adjusted so that an exceeded part of pension assets
may become the same amount of pension liability, and there is not reduc-
tion to a sponsoring company9 . It is explaining to a player being exposed to
the situation of competing with the pension fund staff (other players in an
experiment) of the other companies of a same sector at the time of an ex-
periment start. It is explaining that the result of a pension-assets investment
of a player may inflict a loss on a sponsoring company, and may make it the
disadvantageous situation for competition with the other company.
The performance evaluation of a player is measured from the impact (track-
ing error of retirement benefit costs) given to the profit and loss of a sponsor-
ing company. Ranking of the management capabilities of a player is carried
out by this evaluation. In addition to the impact given to the profit and loss
of a sponsoring company, players are provided also with information ratio
(which is the average of excess return of pension assets to pension liability
divided by the tracking error), the average of excess return of pension as-
sets, and the average of excess return for every round to pension liability as
“Output information” after the end of each round.
Messages are given to a player when every round finishes. The ranking
based on the impact given to the profit and loss of the sponsoring company
of a player is displayed on a message. Moreover, the links for acquiring the
information for deciding the strategy of the next round are displayed. If
equity ratio falls greatly, warning is emitted. As for a slight warning, a yellow
card will be displayed, and, as for a serious warning, a red card will be
displayed. Such a message is a device for giving the incentive which makes
a player a competitive spirit. A sponsoring company may go bankrupt when
defamation of pension assets continues and the loss done to a sponsoring
company exceeds capital. Even when it goes bankrupt, a player can be made
decisions succeedingly.
Experiments are two cases. One is a case where a player is a student,
and the other is a case where a player is an institutional investor affiliation
member. The experiment which uses a student as a player was conducted as a
representative of the beginner who there is no knowledge not much regarding
finance theory, and have not received professional training. Students are five
Tokyo Institute of Technology graduate students (three first-year graduate
students, two two-year graduate students). All five persons can be regarded
9
Although adjustment by change of a premium and so on is performed only once
in about five years, in this experiment, it shall be carried out every term. More-
over, it is assumed that a capital structure change by funding of a sponsoring
company is not made every term, either. Since the pension premium increase
for insufficiency stopgap of pension assets cannot be done by subscriber’s strong
opposition, either, it is assumed that there is only donation from a sponsoring
company.
Learning a Pension Investment in Consideration of Liability 239
as beginners mostly regarding finance theory. On the other hand, the exper-
iment which uses the three affiliation members in an institutional investor’s
investment section as a player, as a representative of those who it is knowl-
edgeable regarding finance theory and have received professional training was
conducted.
All three institutional investor affiliation members are financial analyst
qualification holders. They hold the knowledge regarding finance theory, and
since they have also experienced the training as a fund manager, it can be
considered that they received professional training. The experiment was con-
ducted 3 times10 regarding each case where a player is a student, or an
institutional investor affiliation member. One experiment took 10 rounds. It
took about 1 hour. In order to avoid the end effect to a player, a player has
not been told the end time beforehand.
The rational asset allocation method based on the finance theory in the
experiment of this research is asset allocation which minimizes a prior risk11 .
Asset allocation which minimizes impact (tracking error of retirement benefit
costs) given to the profit and loss of a sponsoring company ex post may
occur besides the asset allocation which minimizes a prior risk. However, the
rational asset allocation which can be made decisions in advance is asset
allocation which minimizes a prior risk. Therefore, in the experiment of this
research, it is judged that there is learning effect with the asset allocation of
a player approaching prior risk minimization asset allocation.
3 Result
This chapter explains the result of having conducted the experiment regard-
ing the pension-assets investment in consideration of liability , in order to find
a difference between the case where professional training has been done and
the case where professional training has not been done regarding finance the-
ory. First, it explains that it was insufficient for bringing regarding learning
effect as for the pension-assets investment which took liability into considera-
tion in the case of the student, where the player has not received professional
training. Subsequently, it explains that there was learning effect through this
experiment regarding the pension-assets investment in consideration of lia-
bility in the case of the institutional investor affiliation member, where the
player has received professional training.
10
The 1st, the 2nd, and the 3rd experiment used respectively the actual market
data from December 2001 to March 2004, from June 2004 to September 2006,
from December
√ 2006 to March 2009.
11
“TE = r 2 + σ 2 ” defines the prior risk. TE : Prior risk. r: The expected return
presumed from a past return. σ: The standard deviation presumed from a past
return.
240 Y. Yamashita, H. Takahashi, and T. Terano
Fig.9 to Fig.11 shows transition of the rate of deviation of the prior risk
based on the asset allocation for which each player in experiment A3 from
the experiment A1 in case a player is a student. The rate of deviation of a
prior risk13 is the difference of the prior risk of a player and the prior risk of
minimal risk asset allocation divided by the prior risk of minimal risk asset
allocation. It is shown that the rate of deviation of a prior risk is carrying
out good decision-making, so that it is close to zero. The label of “p1” in a
figure is a notation for distinguishing five players.
Even if it carries out comparison with Fig.9 and Fig.10, or comparison with
Fig.9 and Fig.11, it is hard to find out the tendency for the rate of deviation
of the prior risk of a player to decrease as it passes through an experiment.
On the contrary, the round in which the rate of deviation of a prior risk is
increasing can also be seen at a player p2 or a player p4.
12
In this paper, the 1st experiment in case a player is a student, the 2nd experi-
ment, and the 3rd experiment are called as experiment A1, experiment A2, and
experiment A3, respectively. It is supposed that a player is called as a player p1
and so forth in order to distinguish five student players.
13
A prior risk is measured as a presumed tracking error.
Learning a Pension Investment in Consideration of Liability 241
Table 1 The average of the rate of deviation of a prior risk (student) (Unit:%)
Experiment p1 p2 p3 p4 p5
A1 39 32 108 35 61
A2 40 115 37 99 90
A3 66 126 15 32 76
Table 2 The rate of the number of having thought as important on the decision-
making (student) (Unit:%)
A1 A2 A3
Reference portfolio(return/risk:max) 21 17 13
Reference portfolio(return:max) 2 1 1
Reference portfolio(risk:min) 2 1 7
Expected return and risk 17 20 19
Graph 15 14 10
Others 43 47 50
Total 100 100 100
Table 1 is the average of the rate of deviation of the prior risk in each
experiment of each player. In the 2nd or the 3rd experiment, as for any
players other than player p3, the rate-of-deviation average of a prior risk
is not decreasing. As this reason, there is no notice that it must carry out
decision-making that a player minimizes a risk in the asset allocation in
consideration of pension liability.
Table 2 is a rate for every experiment which totaled the number of times of
a reply of the item thought as important on the occasion of decision-making of
a player. When a student is used as a player, the rates to which it is supposed
that the reference portfolio which maximizes the return/risk of the 1st line
was thought as important are 21% and a high value in experiment A1. Also in
experiment A3, it has a high rate as compared with 1% of return maximum
244 Y. Yamashita, H. Takahashi, and T. Terano
Fig. 12 Transition of the foreign stock weight (p1 - p5) and of the foreign stock
return (right axis) in experiment A2
Fig.12 shows the transition of the foreign stock weight of each player, and
the foreign stock return of a pre-round in experiment A2. Especially after
round 6, when a return is high, it is shown that the player decide the raising
the weight of the asset. Also from this, a player is considered to become
difficult to notice better decision-making of minimizing a risk, by thinking an
asset return as important.
As mentioned above, from this experiment, when a student is used as a
player, it has checked that there is not much learning effect.
3rd experiment, the asset allocation of player q1 is quite near minimal risk
asset allocation.
Fig.16 to Fig.18 is transition of the rate of deviation of the prior risk
based on the asset allocation as which each player in the experiment B3 from
experiment B1 in case a player is an institutional investor affiliation member.
With comparison of Fig.16 and Fig.17 or comparison of Fig.16 and Fig.18,
Learning a Pension Investment in Consideration of Liability 247
Table 3 The average of the rate of deviation of a prior risk (institutional investor
affiliation member) (Unit:%)
Experiment q1 q2 q3 p-value
B1 102 116 73 —
B2 6 60 18 0.036
B3 2 19 5 0.013
P-value of exp.B2 is calculated by paired t-test with exp.B1 and exp.B2.
P-value of exp.B3 is calculated by paired t-test with exp.B1 and exp.B3.
Table 4 The rate of the number of having thought as important on the decision-
making (institutional investor affiliation member) (Unit:%)
B1 B2 B3
Reference portfolio(return/risk:max) 31 6 0
Reference portfolio(return:max) 0 0 0
Reference portfolio(risk:min) 2 18 22
Expected return and risk 23 17 0
Graph 7 4 11
Others 37 55 67
Total 100 100 100
the tendency for the rate of deviation of the prior risk of a player to decrease
will be seen as it passes through an experiment.
Table 3 is the average of the rate of deviation of the prior risk in each
experiment of each player. In the rightmost column of experiment B2, p-
value is tested for the difference of the average by paired t-test regarding
the average of the prior risk of experiment B1 and experiment B2. In the
rightmost column of experiment B3, p-value is tested for the difference of the
average by paired t-test regarding the average of the prior risk of experiment
B1 and experiment B3. All are significant at 5% of a level of significance.
Through a player’s experience of experiments, a player learns that a minimal
risk strategy is better decision-making.
Table 4 is a rate for each experiment which totaled the number of times of a
reply of the item thought as important on the occasion of decision-making of
a player. When a player is an institutional investor affiliation member, it can
check at the 3rd line that the rate to which it is supposed that the reference
portfolio which minimizes the risk was thought as important is increasing as
it passes through an experiment. By experiencing experiments, this shows
that a player recognized the importance of risk minimization gradually. Also
from this, when a player is an institutional investor affiliation member, it is
shown that there is learning effect by experiencing an experiment.
248 Y. Yamashita, H. Takahashi, and T. Terano
3.3 Consideration
As a result of the experiment, between those who received the professional
training regarding asset management, and those who have not received pro-
fessional training, even if it experienced business game, the difference was
found by the learning effect. In the case of the institutional investor affilia-
tion member which is the representation of those who received training, bet-
ter decision-making has been able to be done regarding the pension-assets
investment in consideration of liability by experiencing business game. On
the other hand, in the case of the student who is the representation of those
who have not received training, it was the result of the ability to seldom say
that there is learning effect even if it experiences business game.
As a reason of this difference, the existence of the professional training
regarding asset management is raised. It is thought that the institutional in-
vestor affiliation member can apply now the rational view based on finance
theory by having received professional training. From the experimental result
of this research, the meaningful result that for those who received professional
training it was effective for learning the decision-making method regarding
the pension-assets investment in consideration of liability experiencing busi-
ness game has been checked . On the other hand, since the student had not
received professional training, it has checked that it was not enough just to
have experienced business game to learn the decision-making method regard-
ing the pension-assets investment in consideration of liability. In order to
also make those who have not received professional training learn effectively
the decision-making method regarding the pension-assets investment which
took liability into consideration by experiencing business game, using a case
method together can be considered. But I would like to consider it as a future
subject regarding it.
4 Conclusion
In this research, through the business game technique, analysis aiming at
learning of the pension-assets investment technique in which pension liabil-
ity was taken into consideration was carried out. Regarding the participant
who received professional training, interesting phenomena – an understand-
ing progresses regarding the asset allocation technique at the time of taking
pension liability into consideration – were seen as a result of analysis. This
result shows the effectiveness of the business game technique to learning of
the pension-assets investment technique. On the other hand, in the partici-
pant who has not received professional training, it has checked that learning
effect was seldom acquired. Although the method which used the case method
together can be considered to learning more effectively by experiencing busi-
ness game to the beginner who have not received training, I would like to
consider it as a future subject regarding it.
Learning a Pension Investment in Consideration of Liability 249
References
1. Faria, A.J.: Business simulation games: Current usage levels—an update. Sim-
ulation & Gaming 29, 295–308 (1998)
2. Greenblat, C.S.: Designing games and simulations. Sage Publications, Inc.,
Thousand Oaks (1988)
3. Kingsland, L.: Projecting the financial condition of a pension plan using simu-
lation analysis. Journal of Finance 37(2), 577–584 (1982)
4. Sharpe, W.F., Tint, L.G.: Liabilities-a new approach. Journal of Portfolio Man-
agement 16(2), 5–10 (1990)
5. Terano, T., Suzuki, H., Kuno, Y., Fujimori, H., Shirai, H., Nishio, H., Ogura,
N., Takahashi, M.: Understanding your business through home-maid simulatior
development. Developments in Business Simulation and Experimential Learn-
ing (26), 65–71 (1999)
6. Tompson, G.H., Dass, P.: Improving students’ self-efficacy in strategic manage-
ment: the relative impact of cases and simulations. Simulation & Gaming 31(1),
22–41 (2000)
7. Walters, B.A., Coalter, T.M., Rasheed, A.M.A.: Simulation games in business
policy courses: is there value for students? Journal of Education for Busi-
ness 72(3), 170–174 (1997)
8. Winklevoss, H.E.: Plasm: Pension liability and asset simulation model. Journal
of Finance 37(2), 585–594 (1982)
9. Wolfe, J.: The effectiveness of business games in strategic management course
work. Simulation & Gaming 28(4), 360–376 (1997)
10. Yamashita, Y., Takahashi, H., Terano, T.: The development of the financial
learning tool through business game. In: Lovrek, I., Howlett, R.J., Jain, L.C.
(eds.) KES 2008, Part II. LNCS (LNAI), vol. 5178, pp. 986–993. Springer,
Heidelberg (2008)
11. Yamashita, Y., Takahashi, H., Terano, T.: Learning a selection problem of in-
vestment projects and capital structure through business game. In: Proceedings
of First KES International Symposium on Intelligent Decision Technologies,
KES IDT 2009, Himeji, Japan, April 23-24 (2009)
An Agent-Based Implementation of
the Todaro Model
Abstract. The problem of internal migration and its effect on urban un-
employment and underemployment has been the subject of an abundant
theoretical literature on economic development. However, most discussions
have been largely qualitative and have not provided enough rigorous frame-
works with which to analyze the mechanism of labor migration and urban
unemployment. In this paper, we build up an economic behavioral model of
rural-urban migration which is an agent-based version of the analytical To-
daro model described by deterministic ordinary differential equations. The
agent-based model allows to explore the rural-urban labor migration process
and give quantitative results on the equilibrium proportion of the labor force
that is not absorbed by the modern industrial economy.
1 Introduction
Economic development generates significant structural transformations such
as changes in the demographic condition and in the production structure.
The most important structural feature of developing economies is the dis-
tinction between rural and urban sectors and particularly, economic develop-
ment is often defined in terms of the transfer of a large proportion of workers
from agricultural to industrial activities [8]. The study of rural-urban labor
migration has for long been an important area of research in development
economics and a large body of literature has grown up in recent year around
this topic in particular for less developed countries (see [1] for a survey on
the theoretical literature).
Nadjia El Saadi · Yacine Belarbi
CREAD, Algiers, Algeria
e-mail: enadjia@gmail.com,yacinebelarbi@yahoo.fr
Alassane Bah
UMI 209, UMMISCO (IRD-Paris 6), ESP-UCAD, Dakar, Senegal
e-mail: alassane.bah@gmail.com
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 251–265.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
252 N. El Saadi, A. Bah, and Y. Belarbi
formulated too close to the analytic model so that comparisons between the
two approaches results can be made. The aim of such a formulation is to
reproduce explicitly the migratory dynamics at individuals scale and analyze
and visualize migration process. By simulating the agent based model, we
explore the behavior of the model and hence check if from an agent-based
approach the Todaro paradox holds or not.
The paper is organized as follows: in section 2, we recall the theoretical
Todaro model. Section 3 is devoted to the agent-based model constructed
from the Todaro model. We first present the simulator model and then de-
scribe the structure of the simulator. Section 4 presents the scenarios tests
and the simulations results followed by a discussion. A conclusion is presented
in section 5.
dS(t)
= {βu + π(t)F (Δ(t))} dt (1)
S(t)
with
Yu (t) − Yr (t)
Δ(t) =
Yr (t)
where S(t) is the size of the total urban labor force in period t, βu the natural
rate of increase in the urban labor force and π(t) the probability of obtaining
a job in the “formal sector” in period t. Yr (t) is the net rural real income
in period t while Yu (t) is the net urban real income in period t. F (Δ(t))
is a function such that dF/dΔ > 0. Δ(t) is the percentage urban-rural real
income differential and the product π(t)F (Δ(t)) represents the rate of urban
labor force increase as a result of migration.
As mentioned earlier, the probability π(t) of obtaining a job in the “formal
sector” is defined as being equal to the ratio of new employment openings in
254 N. El Saadi, A. Bah, and Y. Belarbi
the “formal sector” relative to the number of accumulated job seekers in the
urban informal sector at date t, that is:
γN (t)
π(t) = (2)
S(t) − N (t)
with N (t) the total employment in the urban sector in period t and γ the
rate of job creation in this sector. The difference S(t) − N (t) measures the
size of the “informal sector”.
The model considers that the number of new jobs created increases at a
constant exponential rate over time, specifically
dN (t)
= γ dt. (3)
N (t)
The proportion of the urban labor force employed in the formal sector at
time t (the employment rate) is denoted by E(t), where
N (t)
E(t) = . (4)
S(t)
and the proportionate size of the urban informal sector (unemployment rate)
is denoted by T (t):
T (t) = 1 − E(t). (5)
Todaro [10] solves for the equilibrium rate of employment E ∗ in the simple
case that the income differential D(t) remains constant over time ie D(t) = D
and F (D) = D. He solves for:
dE dN dS
(t) = (t) − (t) = 0. (6)
E N S
The solution for E ∗ is:
γ−β
E∗ = (7)
γΔ + γ − β
Todaro claims that this equilibrium is stable. His intuition behind the equa-
tions above can be explained as follows:
for a developing country in the very early stages of industrialization such
that almost the entire population resides in rural areas, when the urbaniza-
tion process is just beginning, the pool of the urban unemployed is relatively
An Agent-Based Implementation of the Todaro Model 255
small so that the probability of obtaining a job is high. Therefore, for a sig-
nificantly positive Δ and a positive rate of urban job creation exceeding the
natural rate of urban population growth (ie γ > β), the resulting urban ex-
pected real income induces rural-urban migration such that the urban labor
force grows at a faster rate than that of job creation, that is, β + π(t)Δ > γ.
This more rapid growth of labor supply results in an increase in the size of
the urban traditional sector with the result that the probability of a rural
migrant finding a job in the next period is lower (π(t + 1) < π(t)). Assum-
ing Δ and γ remain constant, this lower probability should slow down the
rate of urban labor force growth although dS(t)/S(t) may continue to exceed
dN (t)/N (t). Eventually, π(t) stabilizes the urban unemployment rate at some
level 1 − E ∗ depending upon the values of Δ, β and γ. If the unemployment
rate falls below 1 − E ∗ , equilibrating forces in the form of rising π will be set
in motion to restore the equilibrium.
migration from rural sector to informal sector, transition from informal sec-
tor to formal sector or inversely from formal sector to informal sector. From
this, π(t) can be regarded as the probability of transition from informal sector
to the formal one. Using our notations, formula for π(t) given in (2) becomes
γNf (t)
π(t) = (9)
Ni (t)
with γ the rate of job creation considered constant along a period of sim-
ulation. Employment and unemployment rates defined in (4 ) and (5) are
given by:
Nf (t)
E(t) = Nf (t)+Ni (t) and T (t) = 1 − E(t).
To carry out the simulation of this economic system, workers are placed in the
3 different sectors according initial chosen values for Nr , Nf and Ni . Given
the values of parameters Δ , βu (rate of natural growth) and γ , the first step
in the simulation is to add to urban informal sector the new seekers for jobs
as a result of natural urban labor force growth at the rate βu . Then, the prob-
ability of finding a job π(t) is calculated using (9). If π(t) is strictly positive,
a fraction of individuals in the urban informal sector are selected randomly
to join the formal urban sector, each with probability π(t). If π(t) < 0 (due
to a negative value of γ expressing jobs suppression), a fraction of individuals
in the urban formal sector would be ejected to the informal sector with prob-
ability −π(t) . When π(t) = 0, no change takes place. Then, if both Δ and
π(t) are strictly positive, every individual i in the rural sector is a potential
migrant with probability P (t) = Δπ(t). Otherwise, individual i in the rural
sector does not migrate. To conclude the transition process from one sector
to another one, a random number is generated from an uniform distribution
on ]0,1[. If this number is less than the transition probability assessed for an
individual, than the transition of the corresponding individual holds, other-
wise, no change takes place. Hence, a new sectorial configuration is obtained.
Knowledge of the new populations allows the system to be reset. Therefore,
the state variables of the sectors have to be calculated again and the whole
procedure will be repeated as many times as we set in the simulation.
to manage the sequences of individual actions and the space (Fig. 1). The
dynamic of individuals transitions between the three sectors is summarized
in Fig. 2.
Our migration individual-based model is implemented in the Object-
Oriented language Smalltalk using Visual Works Environment.Time evolves
in a discrete way by steps t.
3.3 Interface
The simulator is provided with a three windows-interface:
(i) the first window allows to initialize, run and visualize evolution of
the sectors and different indicators of migration (Fig. 3),
(ii) the second window allows the user the set up of the global variables and
parameters (Fig. 4),
(iii) the third window is the simulation space, it shows the situation of the
different sectors in time (Fig. 5).
258 N. El Saadi, A. Bah, and Y. Belarbi
Fig. 5 The simulator interface: window 3. The rural sector is in green (its inhab-
itants are colored in blue), the urban informal sector is in yellow (its inhabitants
are colored in yellow) and the urban formal sector is in red (its inhabitants are in
black).
4 The Simulations
Fig. 6 Initial situation in the three sectors when Δ = 1, γ = 0.05 and Ni = 1000
(on the left). The three sectors situation after T = 200Δt (on the right).
An Agent-Based Implementation of the Todaro Model 261
Fig. 7 Unemployment rate evolution in time for different values of γ: γ = 0.1 (solid
line); γ = 0.05 (dashed line) and γ = 0.03 (dotted line).
4.3 Discussion
The simulation of the migration agent-based model shows that a positive
percentage urban-rural real income differential (Δ > 0) and a rate of
urban job creation exceeding the natural rate of urban population growth
(γ > βu ) lead the economic system to an equilibrium characterized by a
null unemployment rate (full employment). Indeed, simulations show that in
the first stage of the urban development, the combination (Δ > 0 and γ > βu )
An Agent-Based Implementation of the Todaro Model 263
Fig. 9 Sectors evolution in time: urban formal sector (solid line); urban informal
sector (dashed line) and rural urban sector (dotted line).
induces rural-urban migration such that the urban labor force grows at a
γN (t)
faster rate than that of job creation, that is βu + Δ Nif(t) > γ. This growth
of labor supply results in an increase in the size of the informal sector with
the consequence that the probability of finding a job in the next period is
lower. Till now, our simulations results agree with the intuitive explanation of
Todaro presented in section 2. But differently to Todaro intuition, simulations
show clearly that the lower probability of obtaining an urban job slows down
the migration rate and hence the rate of urban labor force such that from a
certain threshold, the rate of urban job creation will irreversibly exceed the
γN (t)
urban labor force (γ > βu + D Nif(t) ) leading at long run to an equilibrium
in which the formal sector absorbs all the urban labor force of the informal
sector. But as soon as γ < βu , the resulting equilibrium is unemployment 1.
On another hand, simulations have permitted to test implications of accel-
erated industrial growth and alternative rural-urban real income differentials.
It has been shown that a high urban job creation rate has the effect of ac-
celerating the convergence of the system to the full employment equilibrium
while a high percentage urban-rural real income differential, through inciting
to migration, has the opposite effect.
264 N. El Saadi, A. Bah, and Y. Belarbi
Fig. 10 (a) Employment (solid line) and unemployment (dashed line) rates evo-
lution in time for Δ = 1, γ = 0.01, βu = 0.02, Nr = 500, Nf = 20, Ni = 20, (b)
the rural sector evolution, (c) the urban informal sector evolution, (d) the urban
formal sector evolution.
5 Conclusion
In this paper, we have conceived and implemented an agent-based model
to study rural-urban labor migration in developing countries. This agent-
based model has been derived from the analytic model of Todaro described
by ordinary differential equations representing the links between rural-urban
migration and urban unemployment. The agent-based model permits the
analysis and visualization of the migration process for a large parametric
range. It is also performed in order to give many quantitative results on
equilibrium sectors sizes, equilibrium employment and unemployment rates.
Our main result in this paper is that the Todaro Paradox does not hold
in the long run. Indeed, simulations results stemmed from our agent-based
model show that as long as the urban job creation rate exceeds the natural
urban labor force growth, the economic system converges to an equilibrium
characterized by an unemployment rate 0 although the urban-rural real in-
come differential is high and induces too much migration. Even if this finding
is in contrast with the classical result of Todaro where inter-labor market
(rural-urban) equilibrium mandates urban unemployment, it sustains many
An Agent-Based Implementation of the Todaro Model 265
works (see for instance [7], [12]) reporting that Todaro model predicts too
high an employment rate in less developed countries.
References
1. Bhattacharya, B.: Rural-urban migration in economic development. Journal of
economic surveys 7(3), 243–281 (1993)
2. Blomqvist, A.G.: Urban job creation and unemployment in LDCs: Todaro vs
Harris and Todaro. Journal of Development Economics 5(1), 3–18 (1978)
3. Bousquet, F., Bakam, I., Proton, H., Le Page, C.: Cormas: common-pool re-
sources and multi-agent systems. In: Mira, J., Moonis, A., de Pobil, A.P. (eds.)
IEA/AIE 1998. LNCS(LNAI), vol. 1416, pp. 826–837. Springer, Heidelberg
(1998)
4. Collier, P.: Migration and unemployment, a general equilibrium analysis applied
to Tanzania. Oxford Economic Papers 31, 205–236 (1979)
5. Fields, G.S.: Rural-urban migration, urban unemployment and underemploy-
ment, and job-search activity in LDCs. Journal of Development Economics 2(2),
165–187 (1975)
6. Harris, J.R., Todaro, M.P.: Migration, unemployment and development: a two-
sector analysis. The American Economic Review 60, 126–142 (1970)
7. Mazumdar, D.: The theory of urban unemployment in less developed countries.
World Bank Staff Working Paper 1980 (1975)
8. Ranis, G., Fei, J.C.H.: Development of the labor-surplus economy: theory and
policy. Homewood (1964)
9. Stiglitz, J.E.: Rural-urban migration, surplus labour and the relationship be-
tween urban and rural wages. Easter Africa Economic Review 1(2), 1–27 (1969)
10. Todaro, M.P.: A model of labour migration and urban unemployment in less
developed countries. The American Economic Review 59(1), 138–148 (1969)
11. Weber, A.: The growth of cities in the nineteenth century: A study in statistics,
New York, pp. 184–186 (1899)
12. Zarembka, P.: Labour migration and urban unemployment: comment. Ameri-
can Economic Review 60(1), 184–186 (1970)
New Types of Metrics for Urban Road
Networks Explored with S3: An
Agent-Based Simulation Platform
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 267–285.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
268 C. Genre-Grandpierre and A. Banos
Fig. 1 Variations of the automobile efficacy with the range of the travels
hierarchised by speed, the farther you go, the more you stay on roads which
allows you to drive faster when you are looking for a shortest path in time.
This fact is rarely noticed because the performances of the road networks
are almost always evaluated through global index (such mean speed) without
making any differentiation according to the range of the travel.
Comparing the performances provided by several road networks for differ-
ent range of travels in terms of efficacy, which is an average speed defined as
the Euclidean distance between origin and destination of a trip divided by
the duration of the trip [7], shows evidence of this phenomenon. If we plot the
index of efficacy against the distance travelled (Fig. 1), we can notice that,
on average, the level of performance increases non linearly with the distance
travelled.
So, this metric goes against the objectives of urban planning, as it allows
and even encourages car use, separation between the various places of life
(as home and work) and finally urban sprawl. As all the networks tested are
more or less characterized by the speed metric, we can wonder if another type
of metric is possible, in order to avoid the bad effects of the current one.
Fig. 2 An example of slow metric for the network 40km around the town of
Carpentras
Traffic lights appears to be the only mean to change radically the logic
of the current metric. Indeed, changing the topology of the network or the
distribution of speeds (even if we homogenize circulation speeds) allows just
to reduce the increasing of the performances of the trips according to their
range, but not to invert the ratio of efficacy between short and long-range
trips.
By modifying the number and duration of stops, we thus obtain various
efficacy curves, favouring short-distance trips. These first encouraging results
encouraged us exploring more dynamic and microscopic models, in order to
address some keys issues identified so far such as: (a) the number, duration
270 C. Genre-Grandpierre and A. Banos
and location of stops, (b) the possible structural effect induced by networks
topology and (c) the possible impact on traffic, including congestion and
traffic jams and conversely the impact of the nature of traffic (density, types
of interactions) on the slow metric. For this last point agent-based simulations
seams to be the only way of exploration.
(a) α = 1 (b) α = 2
(c) α = 3 (d) α = 4
• Step 1: Acceleration
If Vn < V max, V n → min(V n+1, V max), i.e. the speed of the nth vehicle
is increased by one.
• Step 2: Deceleration (due to other vehicles/traffic signal)
Suppose Dn is the gap in between the nth vehicle and the vehicle in front
of it, and Dtn is the gap between the car under consideration and the red
light in front of it on the road, then:
if dn ≤ V n or dtn ≤ V n, then V n → min(V n, Dn − 1, Dtn − 1)
• Step 3: Randomisation
If V n > 0, the speed of the car under consideration is decreased randomly
by unity (i.e., V n → V n − 1) with probability p(0 ≤ p ≤ 1). This ran-
dom deceleration probability p is identical for all the vehicles, and does not
change during updating. Three different behavioural patterns are then em-
bedded in that single computational rule: fluctuations at maximum speed,
retarded acceleration and over-reaction at braking.
• Step 4: Movement
Each vehicle moves forward with the given speed i.e. Xn → Xn + V n,
where Xn denotes the position of the nth vehicle at any time t.
Figure 6 illustrates the kind of traffic patterns generated by this simple thus
powerful model.
Fig. 6 Examples of traffic patterns obtained from the extended NaSch model
3 First Results
3.1 The Metric of the Road Network Depends on ...
3.1.1 ... the Number and the Duration of the Traffic Lights
Imagine a hierarchised star network, defined by high speed corridors converg-
ing towards a center. On that basis, we explore the influence of number and
duration of traffic lights (determined in section 2.1) on network efficacy. As
expected, the “speed metric” occurs in the absence of traffic lights. Intro-
ducing such equipments allows reducing the discrepancy between short and
long trips, by penalising the long ones. However, the number of lights seems
to play a secondary role compared to their duration, as expressed by the
clustering of curves. This graph even suggests that a small number of traffic
lights may do the job quite well, if we calibrate their duration, and location,
correctly.
Fig. 7 Impact of number and duration of traffic lights on efficacy for a hierarchised
star network
Beyond this particular case, it appears that the slow metric does not occur
systematically when traffic lights are introduced. Indeed, traffic lights must
have a sufficient density, which must be higher when the network is homo-
geneous in terms of speed. However, this density should not exceed a given
threshold (of one traffic light / 2,5 edges in mean, this threshold varies ac-
cording to the structure of the networks), since in this case we just decrease
the level of efficacy of the network but without reaching a slow metric (the
network becomes homogeneous again).
276 C. Genre-Grandpierre and A. Banos
We also note that the slow metric is particularly obvious when the distribu-
tion of the traffic lights duration has high maximum and standard deviation,
but with a high variability in the individual level of efficacy as a consequence.
Fig. 8 Difference of efficacy for W1 (a) and homo50 (b) without and with 100
traffic lights.
New Types of Metrics for Urban Road Networks Explored with S3 277
Fig. 9 Patterns of the flows (number of shortest paths per edge represents pro-
portionnaly to the thickness of the edges) for 2000 random trips for a hierarchised
network (W1) and a homogeneous network (homo50) without traffic lights.
many shortest paths when the range of the trips increases (cf. 1.1.). However,
they appear on the other hand very sensitive to the introduction of traffic
lights. For example we can see in the Fig. 8 that the difference of efficacy
in the situations with or without traffic lights is higher for hierarchised net-
works (blue curve) than for homogeneous networks (red curve). Thus, for
hierarchised networks, a limited number of traffic lights are enough to obtain
a slow metric, especially if the networks are less connectives.
The sensitivity of the hierarchised networks can be explain by the fact that
their “betweenness centrality” (2.1) is very concentrated on the edges with
the higher speed, which belongs to a lot of shortest paths between origins and
destinations. Thus, few edges participate actively to the routing of the traffic
flows, which are very concentrated on particular roads (here the pattern of
the flows depends strongly of the structure of the network as shown by Penn
et al. [12] and Genre-Grandpierre [3]. Thus, if a traffic light is located on
these edges, it will deeply impact the efficacy. Remember that agents do not
adapt their shortest paths to the location of the traffic lights. Conversely,
the homogeneous networks are intrinsically less effective, but they are less
sensitive too to the traffic lights because the betweenness is more distributed
on the whole network (Fig. 9).
Fig. 10 Star network with 40% of the edges randomly deleted and square network
with 30% of the edges randomly deleted.
shortest paths are more direct. In this case making a detour in order to drive
on speed ways becomes irrelevant.
In order to test the effect of connectivity, we have successively and ran-
domly deleted 10%, 20%, 30%, 40% and 50% of the edges of star and rect-
angular networks more or less hierarchised by speed (Fig. 10).
On this basis, we perform analyses of efficacy with and without traffic
lights. The results shows that it is possible to delete numerous arcs with little
impact on efficacity when connectivity is high. Indeed, in this case, there are
a lot of alternative shortest paths which are very similar. For example, for
a star network with homogeneous speed, it is possible to delete 40% of the
edges with a very limited impact on efficacy.
However, for a less connective network or for a speed-based hierachised
network, if strategic edges are deleted, good solutions of shortest paths which
does not have any alternatives no longer exist. In this case the efficacy is much
lower. For example for a connective and hierarchised network without traffic
lights, we found that the efficacy decreases by 64% if we delete randomly
40% of the edges, but, if the deleted edges with low value of beetweeness
centrality, the efficacy decreases by only 3%!
Here, for the same number of traffic lights, homogeneous networks can quickly
become more efficient than hierarchised networks!
Fig. 12 Influence of traffic interactions on efficacy with and without traffic lights
(network W1 Fig. 8)
interactions between agents are numerous (which is the case for hierarchised
network with a lot of vehicles). In Fig. 13 we can see that it is even possible
to reach a slow metric without traffic lights, for simulations with numerous
agents (1,500) characterised by unhomogeneous behaviour (p = 0.9), but it
remains an extreme simulation case.
Currently we are collecting data with GPS on efficacy of vehicles in traffic
jam situations, in order to refine the calibration of the model and to see if
the slow metric already exists in such situations.
This analyse focuses on the importance of driving habits on the global
efficacy of the roads system. Thus, promoting homogeneous behaviours may
be as important as investing in speed ways in order to improve accessibility.
Concerning the impact of the number of agents in interactions on efficacy, it
differs according to the type of networks. For homogeneous and connective
Fig. 14 Influence of the number of agents on network efficacy with and without
traffic interactions for the network W2 Fig. 13
networks with 100 traffic lights, more than a thousand agents are required
to really impact efficacy. For example for the network W1 in Fig. 8 efficacy
decreases in mean of 0.8 m/s when the number of agents goes from 1,000
to 2,000 agents and the decrease of efficacy is of 1 m/s when the number
of agents goes from 2,000 to 3,000. However, for hierarchised networks the
impact of the number of agents is much more important, even if the number
of agents is limited, since in this case the probability of interaction is very
high (Fig. 14).
Here we can conclude that hierarchised networks are very efficient under
free flow conditions, but that they are unable to assume an increase in traffic
flows. At last, we can notice that the greater the number of agents, the higher
the variance of the value of efficacy at the individual level. The convergence
of individual situations seems to be reached in the case of free flow conditions
or generalised congestion.
with lk the length of edge k and vk its speed. Given step 1 (acceleration) of
the NaSch model, it is obvious that ti > τi . For each trip (and therefore each
agent i), we can then define its loss in fluidity fi as:
ti − τi
fi = , (0 ≤ fi ≤ 1)
ti
which varies between 0 and 1. From that point, we can then define an average
indicator of fluidity loss F, which will also varies between 0 (no loss) and 1
(maximum loss):
1 ti − τi
n
F =
n i=1 ti
Under free flow conditions, fluidity loss is null, as ti = τi , and it tends to-
wards high values when traffic interactions (NaSch rules) and traffic lights
are introduced. In extreme cases, F can go over 0.7. In these cases it may
seem unacceptable to support such a decrease in the efficacy of the system.
However, let us remember that F is a mean index whose value is lower for
short range trips as the slow metric impacts long range trips above all. More-
over, for simulations taking into account the real frequency of the range of
trips, which is very unsymmetrical (in urban areas trips lower than 3 kilo-
meters represent more than 50% of the total), and if in the same time we
improve the connectivity of the network for short range trips, we can reach a
slow metric but with a global accessibility very close to the situation without
traffic lights.
Here the question is, who should be encouraged through a high level of
efficacy: people who made short range trips, or people making long range
trips that urban policies try to restrain!
The more or less periodic shape of the graph is due to the traffic model
and the traffic lights. The data shows that the percentage of vehicles stopped
varies in mean from 5 to 40%. The values are once again higher for simulations
involving hierarchised networks. For example, for the simulation in figure 13,
for the hierarchised network W2 less than 30% of vehicles have come to a
halt for only 13% of the duration of the simulation, whereas this value is of
45% in the case of a homogeneous network.
Theses values may seem very high, but one should not forget that the
simulated traffic patterns are very important (a simulation with 500 agents
represents one vehicle every 50 meters), and that vehicles stopped is not
synonymous with congestion. Indeed, the vehicles stopped are not always the
same. Thus, the simulations shows that congestion appears when the density
of traffic reaches one vehicle every 8-10 meters. Until this extreme value the
performance of the network decreases but without massive congestion.
Here we have a lack of reference data relative to real traffic conditions in
order to know for example if a value of 30% of vehicles stopped represents a
normal or unacceptable situation.
4 Conclusion
These first results are very positive as they suggest that slow speed metric
could be reached with limited and well-targeted efforts, and this is particu-
larly true for hierarchised networks as current networks are. However, several
questions remain unanswered so far. The first one concerns the calibration of
the slow metric which is very complex as it depends on the number, duration
and location of the traffic lights, the structure of the network, the intensity
of the traffic and the nature of the interactions. In order to optimize this
calibration, we have to know more about what people will accept in terms of
maximum duration of the traffic lights, fluidity and accessibility.
Concerning the simulation model, a second issue pertains to the assump-
tion of non-adapting travelling behaviours: by imposing traffic lights with
284 C. Genre-Grandpierre and A. Banos
References
1. Banos, A., Godara, A., Lassarre, S.: Simulating Pedestrians and Cars Be-
haviours in a Virtual City: An Agent-based Approach. In: Proceedings of the
European Conference on Complex Systems, p. 4 (2005)
2. Freeman, L.C.: A Set of Measures of Centrality Based on Betweenness. Sociom-
etry 40, 35–41 (1977)
3. Genre-Grandpierre, C.: Forme et Fonctionnement des Réseaux de Transport:
Approche Fractale et Réflexion sur L’aménagement des Villes: Thèse de Doc-
torat en Géographie. Université de Franche-Comté (2000)
4. Genre-Grandpierre, C.: Des Réseaux Lents Contre la Dépendance Automobile?
Concept et Implications en Milieu Urbain. L’Espace Géographique 1, 27–39
(2007)
5. Girvan, M., Newman, M.E.J.: Community Structure in Social and Biologi-
cal Networks. Proceedings of the National Academy of Sciences of the United
States of America 99, 7821–7826 (2002)
6. Godara, A., Lassarre, S., Banos, A.: Simulating Granular Traffic Using Cellular
Automata and Multi-Agent Models, pp. 411–418. Springer, Berlin (2007)
7. Gutiérrez, J., Monzón, A., Pinéo, J.M.: Accessibility, Network Efficiency and
Transport Infrastructure Planning. Environment and Planning A 30(8), 1337–
1350 (1998)
New Types of Metrics for Urban Road Networks Explored with S3 285
8. Lämmer, S., Helbing, D.: Self-Control of Traffic Lights and Vehicle Flows in Ur-
ban Road Networks. Journal of Statistical Mechanics: Theory and Experiment
(2008), P04019
9. Levinson, D., Kumar, A.: The Rational Locator: Why Travel Times have Re-
mained Stable. Journal of the American Planning Association 60(3), 319–332
(1994)
10. Nagel, K., Schreckenberg, M.: Cellular Automaton Models for Freeway Traffic.
Journal de Physique I 2, 2221–2229 (1992)
11. Newman, M.E.J.: Power Laws, Pareto Distributions and Zipf’s Law. Contem-
porary Physics 46, 323–351 (2005)
12. Penn, A., Hillier, B., Banister, D., Xu, J.: Configurational Modelling of Urban
Movement Networks. Environment and Planning B: Planning and Design 25,
59–84 (1998)
13. Schadschneider, A.: Traffic Flow: A Statistical Physics Point of View. Physica
A: Statistical Mechanics and its Applications 313, 153–187 (2002)
Impact of Tenacity upon the Behaviors
of Social Actors
1 Introduction
In collaboration with sociologists, we have formalized the Sociology of the
Organized Action (SOA) of M. Crozier and E. Friedberg [10]. This theory
[3] [4] [5] is well experienced and it is the foundation of the methodologies
applied by most consultants when they are called to analyze the very origin
of dysfunctions within social organizations.
This formalization has led to the development of the environment SocLab 1
that allows the user to describe the structure of an organization, to explore
its state space and properties, and to find, by simulating the rationality of
social actors, how they could adapt their behaviors to each other [7].
This simulation algorithm, which is based on reinforcement self-learning
rules [12], implements a model of rational actors as they are considered by the
Joseph El-Gemayel · Christophe Sibertin-Blanc · Paul Chapron
Université de Toulouse Capitole / IRIT, 2 Place du Doyen G. Marty, F-31042
Toulouse Cedex 09, 0033 (0)5 61 63 35 00
e-mail: {joseph.el-gemayel,sibertin,paul.chapron}@univ-tlse1.fr
1
SocLab, available at : http://soclab.univ-tlse1.fr/
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 287–306.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
288 J. El-Gemayel, C. Sibertin-Blanc, and P. Chapron
SOA: every one cooperates with others while trying to obtain the means to
achieve his goals. This algorithm includes a set of parameters (ability to dis-
criminate situations, memory, ...) that correspond to cognitive or behavioral
properties of actors and have an impact on the expected simulation results: a
fast convergence of simulations and a good cooperation between the actors.
Among these parameters, the tenacity of an actor determines his persis-
tence, how hard he tries to get from others the capacity to reach his goals;
tenacity is the inverse of resignation. Tenacity seems to be the most influential
parameter, and this paper studies the influence of its value upon the obtained
results, by means of a sensitivity analysis applied to different organizations.
It appears that the more the actors are tenacious, the more the simulations
last and the more the actors cooperate. Thus, the appropriate value of actors’
tenacities in the simulation of an organization is a compromise between the
cost of simulations and the quality of the cooperation. It also appears that
this value depends on structural properties of organizations that are related
to the difficulty, for every actor, to find how to cooperate. These results are
useful for the study of organizations in the framework of the SOA, are they
actual organizations or organizations envisioned in the context of organiza-
tional changes. They are also significant outside the SAO theory, since the
social game may be applied in other domains, for example as a coordination
model for societies of autonomous software agents [9].
The rest of this paper is organized into six parts. After a brief presentation
of the SOA theory in Section 2, we present the main elements of the meta-
model in Section 3 and the cases used in the experimentations Section 4.
After a brief explanation of the actors’ behaviors, Section 5 introduces the
algorithm modeling this behavior and describes its various parameters. Then,
we study its results on two social games by varying the value of the tenacity
parameter in Section 6. Finally, we interpret these results in Section 7 before
to conclude.
the “terms of exchanges” for this resources: he makes his behavior more or
less unpredictable for the actors that depend on this resource and so urge
them to account for his own needs.
Therefore, the power relations structure the configurations of social or-
ganizations. These configurations feature a relative stability in a balance of
social relationships, resulting from the fact that each actor both controls some
resources and depends on others.
Describing an organization as a system of related Actors and Resources
produces a context of interaction rather precisely delimited; this context
structures the cooperation of the actors which, however, keep some level of
autonomy, their room for manoeuvre.
Since the SOA requires that each actor controls resources, another significant
quantity to be taken in consideration is the ability of an actor to influence the
Impact of Tenacity upon the Behaviors of Social Actors 291
In the social game, each actor seeks to obtain from others a high level of
satisfaction and, to this end, he adjusts the state of the resources he controls.
More precisely, at each step, every actor chooses moves of the values of the
states of the resources he controls, and this change of the game state modifies
the satisfaction of all actors. Let (sr1 , ..., srm ) be a state of the organization
and (cr )r∈control(a) be the moves chosen by actor a such that (cr + sr ) ∈
[−1; 1]. Once each actor has chosen such an action, the game goes to a new
state defined by
4 Illustrations
As an illustration, we present two types of organizations: a very simple ex-
ample functioning as a Prisoners’ Dilemma, and the Bolet case as the model
of a complex social organization.
(a) (b)
Fig. 2 The stakes of the actors upon the resources (boxes with black borders show
the controller) and the effect of the resources on the actors (x axis is the space of
choice of the resource, y axis is the resulting capacity of action for the actor).
equilibrium) is when they do not. When one actor cooperates and the other
does not, the former get -100 as satisfaction and the latter +100. (Due to the
symmetry of the game, satis(A1) = - satis(A2) when R1.state = - R2.state).
Although this case is the model of no social organization, it fulfills an
essential property of real social games: they are not zero-sum games, and high
global scores first require the actors’ cooperation and second are compatible
with high individual scores [2].
Fig. 3 The stakes of the actors upon the resources (boxes with black borders show
the controller of the resource).
Fig. 4 The Effect functions of the Bolet case (x axis is the space of choice of the
state of the resource, y axis is the resulting capacity of action for the actor)
whole set of all the possible states of the organization. Among them, some
are of a high interest in that they maximize a given criterion, e.g. the Nash
equilibriums, the Pareto optima or the extreme value of the satisfaction of a
particular actor or the sum of their satisfactions. This analysis is global (the
state space is sorted according to the criterion) and it is a good starting point
when an organization has to be studied. The SocLab environment allows the
user to see these particular states and to interactively get details upon all
states.
As an example, Table 2 shows the states of the organization that maximize
or minimize the satisfactions of every actor (in row), and the satisfactions
that this state provides to every actor (in column). For example, the max-
imum satisfaction that the HS could get is 100, and in this case the global
satisfaction is 87 while the Father’s satisfaction is 42.
294 J. El-Gemayel, C. Sibertin-Blanc, and P. Chapron
Table 2 Satisfaction values in particular states of the Bolet case that maximize or
minimize the actors or global satisfactions
5 Simulation
Among the states of an organization, the question is to determine the ones
that are socially feasible. If we consider a hypothetical organization where
two persons regularly interact according to the pattern of the Prisoners’
Dilemma, it is highly improbable that the one accepts to cooperate while
the other does not! More generally, the analytic approach does not take into
account the capacity of the actors to actually reach a specific state, for cog-
nitive, psychological or social acceptability reasons. To get insights into the
behaviors that the actors of an organization could adopt one to another, we
can make them to learn how to behave, and thus make simulation of these
join self-learning processes.
The design of the algorithm that gives rise to these self-learning processes
can be grounded upon four principles of the SOA regarding the actors’ be-
haviors. According to the SOA [3] [4] [5], the behavior of actors is strategic:
each actor tries to protect and increase his action capacity and to this end,
he handles the resources he controls. As a consequence, that behavior is quite
stable, and a constitutive property of organizations is the regularization of
the actors’ behaviors. However, actors only have a bounded rationality, due
to social relationship opaqueness, and to cognitive skills limitations [11], and
they seek to obtain a satisfactory situation that is not necessary an optimum.
Finally, this behavior is cooperative; indeed, cooperation is required for the
organization to work in a proper way, and every actor recognizes his interest
in this proper working.
current satisfaction in the course of the simulation. The bigger the gap be-
tween his current satisfaction and ambition, the more the actor explores, that
is undertake vigorous actions (that is important changes of resource states),
quickly forgets learned rules and rewards rules which increase widely his sat-
isfaction. When each player gets a satisfaction greater than his ambition, the
algorithm stops and then we consider that a regularized state is reached: ev-
ery actor thinks that he has the best level of satisfaction to achieve his goals
that he can obtain.
This algorithm requires very little knowledge about the state and the struc-
ture of the organization, as actors only perceive resources states and the sat-
isfaction level they get from. That makes this learning process socially and
cognitively likely, and it does not require any communication or deliberation
skills.
5.2.1 Tenacity
This parameter determines the rate at which the ambition of the actor comes
close to his satisfaction; it may vary from 0 to 999. The higher the tenacity,
the slower the ambition decreases, and the longer the actor explores and
has a chance to reach a higher satisfaction. As a consequence of a deeper
exploration, the simulation is likely to last longer before a stationary state is
reached. The value of this parameter is initialized according to the complexity
of the organization to simulate, i.e. depending on the number of actors and
resources and the distribution of the stakes of the actors: the more complex
the organization is, the longer it will take to an actor to find how to cooperate
with others, as there are more rules and more various situations to learn and
to test.
/* pr is the respective reward of the last and penultimate rules, α is the learning rate,
SRt is the Selected Rule at time t, M is the set of rules that are applicable at the
current step, minSatis is the minimum satisfaction of the actor, e.g. as in Table 2.*/
updating ambition
ambitiont =
ambitiont−1 + (satist -ambitiont−1 )×(1000 - tenacity)/105
updating Action-Range
if (satist <ambitiont−1) then
actionRange=
(ambitiont−1 −satist )
initialActionRange+0.5× (ambitiont−1 −minSatis)
else
actionRange = initialActionRange
selecting an action
if (M.isEmpty()) then
SRt = (CurrentSituation(),(atRandom()*actionRange),0,0)
RuleBase.add( SRt )
else
SRt = maximumStrengthRuleFrom(M)
a.action = SRt .action
5.2.3 Scope
This parameter stands for the actor cognitive ability to discriminate the
situations. A situation is a vector of the capacity of acting perceived from
every resource the actor depends on. If the Euclidean distance (weighted
by the stakes) between two situations exceeds the scope threshold, the two
situations are considered as different. At each step, the applicable rules are the
ones whose situation component is close to the current situation of the actor.
A small value of scope (near 0), means that the actor makes a clear distinction
between situations, so that rules have a small domain of application and the
actor creates many specific rules. Whereas an actor with a very large value
of scope (e. g. the amplitude of the range of the actor’s satisfaction) is not
able to distinguish the situations and all rules may be applied at every step.
6 Sensitivity Analysis
As tenacity impacts the exploration / exploitation ratio of the social actors
behaviors, we are looking for the values that provide actors with a good
satisfaction level, without spending too many simulation steps to stabilize
the social game.
To study the influence of the tenacity on the number of steps needed for
convergence and the level of satisfaction gained by the actors, we analyze
the results of the simulation algorithm on two types of organization mod-
els (the classical prisoners’ dilemma, and the model of the Bolet case), and
for each model, we carry out several simulations by varying the value of
tenacity.
Each experiment was conducted with 50 simulation runs. The parameters
of the algorithm were set to values (tested in previous simulations) needed
for good convergence (scope = 16, InitialActionRange = 1.0, and percentage
of distribution of reward = 30% for the last rule and 70% for the penultimate
rule). For the tenacity values below than 900, the number of steps and the
actors’ satisfactions grow linearly, while significant variations appear beyond
900.
Impact of Tenacity upon the Behaviors of Social Actors 299
Fig. 5 The variation of the average number of steps needed for the convergence,
depending on the variation of the tenacity of both actors.
Fig. 6 The variation of the average satisfaction of each actor, depending on the
variation of the tenacity of both actors.
The number of simulation steps varies from about 800 to 28000, with an
intermediate value of 2000 for a tenacity of 980 (Fig. 5).The satisfactions of
the two actors are quite similar, due to the symmetry of the game (Fig. 6)
and they vary from 55 (that is 84%2 of the maximum value) to 76 (97% of
the maximum value), with an intermediate value of 70 (94% of the maximum
value) for a 980 tenacity.
2
This proportion of the maximum satisfaction is given by (value -
min value)/(max value - min value) = 55+ 80/(80 + 80).
300 J. El-Gemayel, C. Sibertin-Blanc, and P. Chapron
It appears that for this game, making the tenacity parameter having a
value about 975 is a good compromise between the quality of the cooper-
ation and the length of the simulation, since a huge additional cost of the
computation gives a very little improvement of the actors’ satisfactions.
We can consider a variant of the standard PD consisting of four actors
that depend the one on another in a circular way: A2 depends on A1 that
depends on A4 that depends on A3 that depends on A2. In this case, it is
much more difficult for the actors to discover how they could cooperate, since
all of them must do simultaneously. The number of simulation steps varies
from about 2000 to 30000, with a value of 10000 for a tenacity of 990, while
the satisfactions of the four actors vary from 30 to 62, with a value of 50
for a tenacity of 990. These results are comparable to the ones of the two
prisoners case, but the improvement of tenacity is not enough to produce a
good cooperation.
Fig. 7 The variation of the average number of steps needed for the convergence,
depending on the variation of tenacity for all actors.
Impact of Tenacity upon the Behaviors of Social Actors 301
Fig. 9 Variation of the average number of steps needed for convergence depending
on the variation of HS’s tenacity.
Fig. 10 Variation of the average actors satisfaction depending on the HS’s tenacity.
7 Interpretation
For any organization, the level of the actors’ satisfactions and the number of
steps of simulations increase with the value of the tenacity parameter in a
Impact of Tenacity upon the Behaviors of Social Actors 303
Fig. 11 The variation of the mean satisfaction divided by the number of steps,
depending on the variation of the actors’ tenacity in the Bolet case.
quite linear way until the 900 threshold, and they increase in an exponential
way beyond this value. This is the reason to explore the influence of this
parameter within the 900 -999 range.
The tenacity parameter has been introduced in the simulation algorithm
to determine the level of exploration with regard to the level of exploitation:
the more the current satisfaction of an actor is below his ambition, the more
he explores, and his level of ambition decreases as the inverse of the tenacity.
The increase in tenacity produces an increase in exploration which in turn
produces an increase in the duration of simulations, and this increase is more
and more important beyond a threshold which is about 960 - 980: this fact
has been observed in the two cases presented in this paper and also in other
organizations.
Regarding the influence of the tenacity upon the level of cooperation of
actors, and their resulting individual levels of satisfaction, the two examples
figure different situations: an effective linear improvement in both the two and
four actors PD cases due to a better exploration and, in the Bolet case, no
change when all actors have the same tenacity and an important inversion,
surprising at first glance, of the HS’s and Jean-BE satisfactions when HS
improves his tenacity. To contrast the cases, the PD game is sensible to a
characteristic of the actors, i.e. their tenacity, and the sparsity of cooperative
states (in the four actors PD) requires a higher tenacity, while the Bolet game
is not when all actors feature the same tenacity. The Bolet organization is
strongly regulated and firmly constraints the behaviors of actors that are
engaged in numerous relations. However, a singular actor, HS, can escape
these constraints to obtain a much better situation to the detriment of a
vulnerable actor.
In any case, a relevant value of the tenacity parameter has to be found,
that is a compromise between the two effects of a high tenacity: a costlier
304 J. El-Gemayel, C. Sibertin-Blanc, and P. Chapron
simulation and a better cooperation. To this end we can consider the following
function:
Global Satisf action(tenacity)
T enacity Inf luence(tenacity) =
number steps(tenacity)
that compares the respective increases in the global satisfaction and the dura-
tion of the simulation (see Fig. 11). This function has a negative slope for any
simulation: the number of steps increases more quickly than the global satis-
faction while the tenacity is improved. A good value for the tenacity is when
this function comes under some threshold that is determined by the relative
importance of the cost of simulations and the quality of actors’ cooperation.
8 Conclusion
This SAO-based model of coordination is relevant in cases where the cooper-
ation of the actors is of first importance, notably in the contexts of Command
and Control [6] or of crisis management [13]. On the one hand, the success of
the operation requires the full cooperation of each member of the Headquar-
ter; on the other hand, each member has to promote the own interests of his
department, to protect the people acting on the ground, handle carefully the
resources or in the perspective of further operations. The cooperation is not
always so easy to achieve [14], [7], and a rigorous model of such interaction
contexts can highlight structural drawback of the organization. Moreover, the
tenacity directly refers to a psychological feature of the actors, and simula-
tions could help to select the most appropriate representative of a department
that is in a difficult position.
The distinctive feature of the algorithm presented in this paper is to al-
low the emergence of states featuring a high level of cooperation between
the actors, a global property, while it is very parsimonious of the cognitive
capacities of the actors and of their the knowledge of both the structure and
the current state of the organization they belong to. It is a distributed op-
timization algorithm that provides good results while it demands very few
information.
In this paper, we investigate the influence of the tenacity parameter on the
quality of results of this simulation algorithm, that is the number of steps
required for convergence and the actors’ capacities to achieve their goals.
It appears that the value of this parameter has a great influence on the
simulation results. If some principle have been highlighted, a deeper study is
necessary to relate the effects of this parameter with structural properties of
games.
Since our model focuses on the behavior of autonomous actors and the
value of tenacity influences the kind of regularization that is achieved, we have
also to study how each actor could compute his own level of tenacity according
Impact of Tenacity upon the Behaviors of Social Actors 305
to his satisfaction target, the best global satisfaction of the organization and
of given maximum cost of the simulation.
The behavior of the algorithm is constrained by several other parameters,
hence it will be necessary to study and analyze their influence on the al-
gorithm’s results (especially the Action Range’s parameter that matters in
exploration) and the effect of the join variation of several parameters.
Acknowledgements. This work has been supported by the ANR (French Re-
search Agency) through the IsyCri project.
References
1. Axelrod, R.: The Complexity of Cooperation: Agent-Based Models of Compe-
tition and Collaboration. Princeton University Press, Princeton (1997)
2. Clegg, S.R.: Handbook of organization studies. Sage Publ., Thousand Oaks
(2003), http://www.worldcat.org/oclc/249400076
3. Crozier, M.: The bureaucratic Phenomenon. Chicago University Press, Chicago
(1964)
4. Crozier, M., Friedberg, E.: L’acteur et le système. Seuil, Paris (1978)
5. Crozier, M., Friedberg, E.: Organizations and Collective Action: our Contribu-
tion to Organizational Analysis. In: Research in the Sociology of Organizations,
vol. 13, Jay-Press, Greenwich (1995)
6. Ioerger, T., Linli, H.: Modeling command and control in multiagents systems.
In: 8th Int. Command and Control Research and Technology Symposium,
Washington DC, June 17-19 (2003)
7. Mailliard, M., Amblard, F., Sibertin-Blanc, C., Roggero, P.: Cooperation is
not always so simple to learn. In: Terano, T., Kita, H., Deguchi, H., Kijima, K.
(eds.) Agent-based Approaches in Economic and Social Complex Systems IV,
pp. 147–154. Springer, Heidelberg (2007)
8. Mailliard, M., Roggero, P., Sibertin-Blanc, C.: Un modèle de la rationalité
limitée des acteurs sociaux. In: Chevrier, V., Huget, M.P. (eds.): Journées Fran-
cophones sur les Systèmes Multi-Agents, Annecy, 18/10/2006-20/10/2006, pp.
95–99 (2006), Hermès, http://www.editions-hermes.fr/
9. Sibertin-Blanc, C., Adreit, F., Chapron, P., Roggero, P.: A Model for the Social
Regularization of Autonomous Agents (2009) (submitted to publication)
10. Sibertin-Blanc, C., Amblard, F., Mailliard, M.: A coordination framework based
on the Sociology of Organized Action. In: Boissier, O., Padget, J., Dignum, V.,
Lindemann, G., Matson, E., Ossowski, S., Sichman, J.S., Vázquez-Salceda, J.
(eds.) ANIREM 2005 and OOOP 2005. LNCS (LNAI), vol. 3913, pp. 3–17.
Springer, Heidelberg (2006), http://www.springerlink.com
11. Simon, H.A.: The sciences of the artificial, 3rd edn. MIT Press, Cambridge
(1996)
12. Sutton, R.S., Barto, A.G.: Introduction to Reinforcement Learning. MIT Press,
Cambridge (1998)
306 J. El-Gemayel, C. Sibertin-Blanc, and P. Chapron
13. Tahir, O., Andonoff, E., Hanachi, C., Sibertin-Blanc, C., Benaben, F.: A Col-
laborative Information System Architecture for Process-based Crisis Manage-
ment. In: Lovrek, I., Howlett, R.J., Jain, L.C. (eds.) KES 2008, Part III. LNCS
(LNAI), vol. 5179, pp. 630–641. Springer, Heidelberg (2008)
14. Van Doren, P.H.: Decision making and cognitive analysis: Track a historic fail-
ure in the social domain. In: 10th Int. Command and Control Research and
Technology Symposium, Virginia Beach, VA (2005)
Part III
Agent Technology for Environmental
Monitoring and Disaster Management
308 Agent Technology for Environmental Monitoring and Disaster Management
Sarvapali D. Ramchurn
Serge Stinckwich
ATDM Workshop Chairs
Siddeswara Guru
IASNW Workshop Chair
Dynamic Role Assignment for
Large-Scale Multi-Agent Robotic
Systems
Van Tuan Le, Serge Stinckwich, Bouraqadi Noury, and Arnaud Doniec
1 Introduction
Traditional disaster relief operations involve a huge human resources, includ-
ing different on-site rescue teams, and remote coordinators. On-site rescue
teams are sent to explore the disaster area to find victims. Having located
victims, they send victims situation to remote coordinators informing about
the needs of medical cares or of any assistance else. Based on global infor-
mation received, coordinators should give instructions to rescuers, who have
usually a limited partial view. Hence, the most efficient cooperations between
on-site teams is achieved. Moreover, as the rate of victims survivability drops
drastically after 72 hours (thus called 72-golden-hours), and locating victims
within three first hours is desirable the most, the rescuers have to find vic-
tims in very strict timely manner. Therefore, it is ideally required to have
numerous teams in parallel and cooperatively exploring the disaster site.
In this context, the AROUND [3] project (Autonomous Robots for Ob-
servation of Urban Networks after a Disaster) aims at building a complete
Van Tuan Le · Serge Stinckwich
UMMISCO, UMI 209 (IRD, UPMC, IFI-MSI) and GREYC, UMR 6072
(CNRS, Université de Caen) and École des Mines de Douai, France
e-mail: {lvtuan@gmail.com,Serge.Stinckwich}@ird.fr
Noury Bouraqadi · Arnaud Doniec
Université Lille Nord de France and École des Mines de Douai, France
e-mail: {bouraqadi,doniec}@ensm-douai.fr
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 311–326.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
312 V.T. Le et al.
Participatory Geographic
Multi-Agent Information
Simulation System
robot team
Local level
Communications
rescue
team
real-time deci- sion support system for natural disaster management in ur-
ban areas. The system is composed of 2 subsystems: (i) a local level with
the deployment of autonomous mobile robots able to collect information in
impacted urban sites and to dynamically maintain the communication links
between rescuers, (ii) a global level based on a Spatial Decision Support Sys-
tem (SDSS).
The rescue operations impose a very quick exploration over disaster zone,
yet the environment in this context is extremely complex and dynamic. There-
fore, the number of the fleet of robots being deployed to assist the rescue oper-
ations is potentially very large, and a priori non-determined. We expect that
the robotic system should be able to be re-organize for different rescue tasks.
The proposed solution therefore should meet the following requirements:
• Scalability: we expect the solution to deal with a very large number of
robots and this without performance degradation,
• Robustness: if there is one robot that can not function anymore, the
others take over its task and the system can continue without any impact,
• Openness: system operators might easily add new robots to the system
on-the-fly,
• Flexibility and Reusability: we expect to reuse as much as possible
some parts of the solution for similar missions. Also, the set of robots
might vary in technology, or individual capabilities.
In this paper, we consider a robotic system as a social organization [7, 11, 2].
The organizational approach of multi-agent systems has a long history since
the 90s, when the term “agent society” was coined [7]. Organization is defined
as a supra-agent pattern of emergent cooperation or predefined cooperation
Dynamic Role Assignment for Large-Scale Multi-Agent Robotic Systems 313
Role Assignment
Engine installed on
Robot
<<described by>>
InterGroupConstraint OrganizationDescription Organization
* *
<<described by>>
GroupConstraint GroupDescription Group
plays role in
* *
<<described by>>
RoleConstraint RoleDescription Role Robot/Agent
plays
3.1 Concepts
• Neighbors : is a distance function whose goal is to compute which robots
are neighbors of a given robot. This function can be built in an ad hoc
way, that varies from an application to the other. In its simplest form, two
316 V.T. Le et al.
Right after having started up, robots are in an initial state and try to join
an organization. However, there should be at least one starting point : the
robot which initialize the organization by assigning it the first role of the
organization, along with the organization description.
• Init: every robot broadcasts the OrgDescRequest message to all its
neighbors; the broadcast procedure is repeated for some interval of time,
until they receive a response from an organization member.
• Role selection: received the OrgDescResponse, robot extracts the
organization description to rebuild the organization’s image. The next step
is to verify role constraints to find an appropriate role. According to this,
robot will broadcast NewGroup or Join message.
• Disappearance of a neighbor: whenever made aware of the disap-
pearance of one of its neighbor, if the robot is a group manager and the
newly disappeared robot is in its group, it removes the latter’s ID from
its group member list. If a robot is an ordinary member, and its group
manager has disappeared, it needs to re-execute the Role selection
procedure to adapt to the organization’s changes.
In the implementation, the Role selection procedure can be re-executed
on new events. For example, when robot detects that it is out of its group or
when it finished its task attached to the role.
As long as robots have not joined the organization yet, they could com-
municate merely with their neighbors by sending these organizational setup
messages. There are four messages; and all necessary information are sup-
posed to be encoded in the message. Here are the messages and the procedure
that robots will execute depending on the message received and on its own
role:
• OrgDescResquest: sent out by the robots that are not yet members of
the organization to their neighbors, and would like to join the organization.
They send this message to ask for the description of the organization (so
that it can select a role to play). A member, whatever its own role is, will
answer with the OrgDescResponse, otherwise, nothing to do.
• OrgDescResponse: contains the organization description, with all group
and role descriptions.This message is the response of the previous one. Re-
ceiving this message, robots do not thing if they have been organization
members already. Otherwise, they execute the role selection proce-
dure. This is the only message that a robot with non-determined role reacts
upon.
• NewGroup: used to inform the creation of a new group. This message is
created by a robot that decides to take the role of the manager of a new
group. Because this message indicates changes in the organization, the
robots that receive the messages need to check again the constraints of its
current roles, if the constraints are violated; they need to re-execute the
role selection procedure.
318 V.T. Le et al.
F
A
C
B
D
G
4 Validation
In this section, we sum up and analyze the results of different simulations we
have made to validate our proposal.
Number of steps 10 11 12 13 14 15 16 17 18
Frequencies over 1000 simulations 33 170 111 180 163 153 122 65 3
Fig. 4 Number of members, new members, role changes and messages sent in each
step.
Dynamic Role Assignment for Large-Scale Multi-Agent Robotic Systems 321
organization. In average of 1,000 simulations, each robot has changed its role
from 2 to 3 times.
Number of messages sent: the average rate of number of messages sent
in the simulations is around 18 times higher than the number of the robots.
There are totally from 15,000 to 20,000 messages sent with 1,000 robots in
the simulation.
The graph in figure 4 shows the relationship between the number of the
robots who have just join the organization, the number of role changes, and
the number of the messages of all kinds issued at each step. We might notice
that the variations of the number of role changes and the number of messages
sent are relatively close. This observation can be explained by the fact that, a
new “comer” to the organization will send one OrgDescRequest message, it
then decides to take part in the organization, and this might cause the changes
for the organization member close to it. For this, they have to exchange the
messages, mainly of kind NewGroup or Join.
Fig. 5 Impacts of robots’ radio range on number of messages sent and number of
role changes
only triggered by the external changes, but also by the internal changes of
robots themselves.
6 Related Work
In the literature, the role assignment in the domain of multi-robot systems is
a well-known problem already investigated in several works. Role assignment
is also known as task allocation [10] where the roles in the system is identified
as a set of (related) tasks. The most popular approach for task allocation lies
on a market-like protocol, namely the ContractNet Protocol [12]. Whenever
there is the task to be solved, a corresponding task proposal is generated and
submitted to all available robots. Robots, based on the information provided
along with the proposal, evaluate their appropriateness for the task to refuse
or accept it. In the latter case, robots submit their bid, and the one whose
bid value is the highest will be assigned for the task. This market-based
task allocation is implemented in [12], and particularly in the domain of
RoboCup [10, 6]. C. Candea et al. [4] have implemented the role assignment
mechanism to coordinate the RoboCup soccer team. The process consists
of two steps: (1) formation selection; and (2) role assignment. In the first
step, all robots vote for the best team formation. Then passing to the second
step, all roles are sorted in their order of importance. Using a greedy market-
based task allocation algorithm, the role is then assigned for each robot in
the fixed, total order. This is to insure that in case of robots’ failure, the
remaining member(s) will be always assigned the most “important” roles.
This market-based approach is useful to deal with task allocation where
the ordering of tasks, the availability of robots, and the suitability of a robot
for a particular task are not know. However, it does not scale well as the
space to search for robots includes all the robots in the system. To address
this shortcoming, C. McMillen and M. Veloso [10] implemented in their soc-
cer robots team a mechanism to limit the search space by introducing the
region-based task allocation: each robot is assigned to a region of the field.
A robot is primarily responsible for going after the ball whenever the ball is
in that robot’s region. In our approach, we view a role from the higher level
of abstraction than that of a task [11]. A role is not only a set of (related)
tasks, but also represents the relationship between different system entities.
We expect that robot enacting a role in an organization knows how to ful-
fill the task associated to that role (robot’s autonomy) in cooperating with
other robots in the system (thanks to the organization constraints, and role
relationship).
Chaimowicz et al. [5], proposed an architecture for tightly coupled multi-
robot cooperation - a system in which always exist a leader, and the others
follow this one. The role of leader and that of the followers are not fixed,
but might be changed on-the-fly. This work can be seen as a particular
case in our approach, where the organization is composed of a single group.
Dynamic Role Assignment for Large-Scale Multi-Agent Robotic Systems 325
Similarly to the approach presented in this paper, the role assignment and the
coordination between robots are guaranteed by the communication protocol
and control algorithms.
Automatic allocation of role is also largely investigated in pervasive com-
puting systems such as sensor networks. C. Frank et al [8], proposed a generic
algorithm to assign role to a node in a sensor network. They have developed
a framework allowing the system designers to specify the role specification
- a list of role-rule pairs in form of predicates. For each possible role, the
associated rule specifies the condition for assigning this role. These specifi-
cations are then used by node in the network to determine its proper role.
However, their solution is optimized for sensor networks, thus the mobility
is not taken into account. Furthermore, whilst the evolution of the organi-
zation might take place during its execution if we change the organization
description with our approach, it is not applicable for their algorithm.
7 Conclusion
We presented in this paper an organization-based approach for structuring
large-scale robotic system. This can be used to design systems that need to
be self-organized through the dynamic process role assignment.
Our proposal is built as an extension of the AGR meta-model. It unifies
the design process and deployment of multi-agent robotics systems. We call
our proposed approach “optimistic” because each agent immediately endorses
roles he choose. Potential conflicts are handled when they are encountered
(e.g., choosing the best robot to play the role of manager of a group). Another
possible strategy, known as “pessimistic”, is to anticipate potential conflicts.
Assigning a role to an agent is subject to validation by an another agent.
Another alternative strategy is to combine the “optimistic” and “pessimistic”
approaches. A future perspective for our work is to study these different
strategies and classes of problems they are best appropriate for.
References
1. Basagni, S.: Distributed clustering for ad hoc networks. In: Proceeding of the
1999 International Symposium on Parallel Architectures, Algorithms and Net-
works (I-SPAN 1999), pp. 310–315 (1999)
2. Boissier, O., Hübner, J.F., Sichman, J.S.: Organization Oriented Programming:
From Closed to Open Organizations. In: O’Hare, G.M.P., Ricci, A., O’Grady,
M.J., Dikenelli, O. (eds.) ESAW 2006. LNCS (LNAI), vol. 4457, pp. 86–105.
Springer, Heidelberg (2007)
3. Boucher, A., Canal, R., Chu, T.Q., Drogoul, A., Gaudou, B., Le, V.T., Moraru,
V., Nguyen, N.V., Vu, Q.A.N., Taillandier, P., Sempé, F., Stinckwich, S.: The
AROUND project: Adapting robotic disaster response to developing countries.
In: Proceedings of 2009 IEEE International Workshop on Safety, Security, and
Rescue Robotics, IEEE Computer Society, Los Alamitos (2009)
326 V.T. Le et al.
4. Candea, C., Hu, H., Iocchi, L., Nardi, D., Piaggio, M.: Coordination in multi-
agent robotcup teams. Robotics and Autonomous Systems 36, 67–86 (2001)
5. Chaimowicz, L., Sugar, T., Kumar, V., Campos, M.F.M.: An architecture for
tightly coupled multi-robot cooperation. In: Proceedings of the IEEE Interna-
tional Conference on Robotics and Automation (ICRA), vol. 3, pp. 2992–2997
(2001)
6. Emery, R., Sikorski, K., Balch, T.: Protocols for collaboration, coordination
and dynamic role assignment in a robot team. In: Proceeding of the IEEE
International Conference on Robotics and Automation (ICRA), pp. 3008–3015
(2002)
7. Ferber, J., Gutknecht, O., Michel, F.: From agents to organizations: an orga-
nizational view of multi-agent systems. In: Giorgini, J.O.P., Müller, J. (eds.)
Agent-Oriented Software Engineering (AOSE) IV, pp. 214–230 (2004)
8. Frank, C., Römer, K.: Algorithms for generic role assignment in wireless sensor
networks. In: Proceedings of the 3rd international conference on Embedded
networked sensor systems. ACM Press, New York (2005)
9. Hübner, J., Sichman, J., Boissier, O.: S-moise+: A middleware for develop-
ing organized multi-agent systems. In: Boissier, O., Dignum, V., Matson, E.,
Sichman, J. (eds.) International Workshop on Organizations in Multi-Agent
Systems: From Organizations to Organization Oriented Programming (OOOP
2005), pp. 107–120 (2005)
10. McMillen, C., Veloso, M.: Distributed, play-based role assignment for robot
teams in dynamic environments. In: Alami, R., Chatila, R., Asama, H. (eds.)
Proceedings of Distributed Autonomous Robotic Systems 6, pp. 145–154.
Springer, Heidelberg (2007)
11. Odell, J.J., Parunak, H.V.D., Fleischer, M.: The Role of Roles in Designing
Effective Agent Organizations. In: Garcia, A.F., de Lucena, C.J.P., Zambonelli,
F., Omicini, A., Castro, J. (eds.) Software Engineering for Large-Scale Multi-
Agent Systems. LNCS, vol. 2603, pp. 27–38. Springer, Heidelberg (2003)
12. Ulam, P., Endo, Y., Wagner, A., Arkin, R.: Integrated mission specification and
task allocation for robot teams - design and implementation. In: Proceeding
of the 2007 IEEE International Conference on Robotics and Automation, pp.
4428–4435 (2007)
Multi-agent System for Blackout
Prevention by Means of Computer
Simulations∗
1 Introduction
Multi-agent systems bring new possibilities into many research areas. The
main idea of decomposing the given problem into separated independent
task is not new and is already used in many other software design areas.
But multi-agent paradigm moves such an approach even further. In the clas-
sic software design paradigms (e.g. object-oriented programming, OOP), the
decomposition into separated units mainly helps the designer to decrease the
(logical) complexity of the problem but usually does not automatically make
Miroslav Prýmek · Aleš Horák · Adam Rambousek
Faculty of Informatics, Masaryk University
Botanická 68a, 602 00 Brno, Czech Republic
e-mail: {xprymek,hales,xrambous}@fi.muni.cz
∗
This work has been partly supported by the Czech Science Foundation under
the project 102/09/1842.
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 327–337.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
328 M. Prýmek, A. Horák, and A. Rambousek
the (physical, real) solution decomposed, modular and scalable. Even if the
designer uses the OOP approach well and extensively, the final software sys-
tem remains monolithic unless the explicit element borders and interfaces are
designed and implemented.
On the other hand, the multi-agent systems force the designer to think
about the software system units as really independent, self-contained and
autonomous entities. The overall system is then viewed as a community of
autonomous “organisms” rather than one complex with many subsystems.
In OOP, the main principle constituting the software system is (usually
synchronous) message sending or method calling. We can consider it to be
an instruction: you [the object] should do this [the message] and give me the
result of the action. The object is not instructed on how the action should be
taken but is supposed to perform the action. If the action is not performed,
the whole system will probably get broken.
The multi-agent approach is a little bit different. The “organisms” (agents)
constituting the “community” (system) are regarded as self-controlled au-
tonomous entities with a knowledge about their surrounding world and a
decision-making algorithm which controls the actions of the agent. The prob-
lem of the system design is then a problem of the appropriate agent design
and configuration. The binding between the agents should remain as loose
as possible – it is based mainly on the interchange of the particular agents’
knowledge about the surrounding world. A typical communication between
agents will be not an instruction but rather a question-answer pair followed
by the decision-making process:
Agent A: What is your current condition?
Agent B: Pre-failure. Probability of the failure in the near future is high.
Agent A decision-making: prepare for taking over the B’s functions
with the Technical University of Ostrava, Czech republic. The main aim of the
research is to develop a simulation tool for the explorations in the field of a
condition-based maintenance of the power distribution network facilities. The
multi-agent approach is suitable for highly flexible, modular and distributed
simulations. The flexibility of the system makes it possible to implement
simulations of the highly dynamic processes involving intelligent decision-
making.
The Rice simulator is based on autonomous agents implemented as inde-
pendent programs or processes [6]. The inter-agent communication is based
on the CORBA standard (Common Object Request Broker Architecture [3])
which enables the agents to communicate through local or even wide-area
network. Thus it is possible to deploy the system as highly distributed.
The meaning-layer of the communication is based on the KQML (Knowl-
edge Query and Manipulation Language [2]) in cooperation with two possible
content languages: python as a highly-expressive but expensive language and
a simple one-purpose declarative language as a standard communication in-
strument. For details about the communication languages see [7].
I1 I2
Uin Uout
R1 Agent C
R2
P1 Agent G
Agent A Agent B Agent F P3
Agent D
R2 I3
P2
Agent E
P =U ·I (1)
332 M. Prýmek, A. Horák, and A. Rambousek
There is also the maximal amount of electric power the agent can deliver to
the network as a pre-set value. When this amount is exceeded, the special
event is triggered.
The agents knowledge base will contain these items:
1. U – voltage (volts)
2. I – current (amperes)
3. P – load (watts) – computed according to the relation (1)
4. P max – maximal allowed load
5. overloaded – true/false, meaning that the source is actually overloaded
To make all the knowledge pieces correct and make the agent cooperate
with other agents in the network, we must define these behavioral reactive
primitives:
1. an agent tells me that it is connecting to my output −→ subscribe for its
I demand knowledge
2. subscribed I demand changes −→ change the current I accordingly
3. the current I changes −→ recalculate the value of the load P
4. the voltage U changes −→ recalculate the value of the load P
5. if the load P > P max −→ put true into overloaded
6. if the load P ≤ P max −→ put false into overloaded
Note that the rule 3 if separated from the rule 5. If the classic functional
paradigm were used, this behavior would be implemented in a function
changeI(i) which takes I, computes P and triggers an alarm if P ex-
ceeds P max. But with the behavior-primitives approach, each of these steps
is a distinct behavioral component. This approach allows to implement nodes
which do not monitor their overloading with just removing this specific rule 5
from the set of their attached behavioral primitives.
Let us now present the details of implementing the model behavior using
the functionalities of the Rice simulator (section 2.1).
The electric line which is going to be connected to the source node sends
a message
{tell
:sender <line-agent id>
:receiver <source-agent id>
:content output=<line-agent id>
}
which changes the value of the knowledge output in the agent represent-
ing the source. The rule 1 is then implemented as a change handler of this
knowledge. The Rice’s XML definition of this behavior looks like this:
<handler valueName="output" name="subscribe-I_demand">
<action type="sendKqmlMessage">
<kqml performative="subscribe" language="Kqml">
<kqml performative="ask" language="Trivial"
Multi-agent System for Blackout Prevention 333
reply-with="I_demand of output">
I_demand=?
</kqml>
</kqml>
</action>
</handler>
Whenever an agent connects to the source output, the source will send him
the subscription for the I demand knowledge. Whenever the I demand value
changes in the output, it will inform the source with the message:
{tell
:sender <line-agent id>
:receiver <source-agent id>
:in-reply-to "I_demand of output"
:content I_demand =<value>
}
When this message arrives to the source, the appropriate message handler is
triggered and will set the source’s I (rule 2).
The meaning of this messages passing chain can be interpreted as: “When-
ever an agent becomes my output, I will be interested in its power current
demands. I will saturate this need.” Note that the source node will not send
the message about changed I automatically. If the output agents need this
piece of knowledge, they must subscribe for it.
All the other rules are implemented as the change handlers defined by
the Python code. They ensure that the values of all the knowledge about
the power flow in the node remain consistent to the given relation (1). For
instance, the code of the rule 3 will look like this:
<handler valueName="I" name="I-handler">
<action type="python">
def action(data):
agent = data[’agent’]
agent[’P’] = agent[’U’] * data[’newValue’]
return True
</action>
</handler>
R which is the resistance of the line, and two voltage indicators: the input
voltage U and the output voltage U out.
The line resistance causes the power loss on the line. The loss is:
ΔU = R · I (2)
U out = U − ΔU (3)
threshold which effectively defines the desired accuracy of the energy flow
computation.
Throughout the whole process, the consumer agent simulates the power
demand variation in time. The Rice system contains the timer routines which
trigger the event in the pre-set time periods. As an reaction to this event
the agent computes (or retrieves from a table) a power consumption value
corresponding to the actual time of the day and sets the requested actual
power consumption P demand. All the power flow indicators get recomputed
automatically.
4 Conclusions
We have presented the network power flow model implementation in the
Rice multi-agent simulation environment. The system is based completely
on local interactions between autonomous agents and on-demand knowledge
propagation. No information is communicated without being requested.
Since each part of the simulation process is based on changes of agents’
knowledge and the reactive nature of the agents, it is possible to manually
change any of the involved indicators interactively during the simulation,
including the network topology.
The Rice system also offers tools to visualise the network,2 and to store
the computed values in any moment for off-line post-processing or statisti-
cal analysis. The concept of freely-combinable little pieces of code defining
the reactive behavior of the agents is suitable for machine generation of the
simulation definitions.
In the time of exploitation of electric appliances in all kinds of human
activity, any larger power outage, or blackout, may have a significant so-
cial and economical effects. Although it is impossible to prevent the black-
out completely or to predict it for a certainty, a computer simulation may
help to detect weak parts of the power network which should be replaced or
back-uped.
The future work will be focused on the implementation of complex models
involving intelligent decision making agents in highly dynamical environment.
References
1. Abel, E., et al.: A multiagent approach to analyze disturbances in electrical net-
works. In: Proceedings of the Fourth ESAP Conference, Melbourne, Australia,
pp. 606–611 (1993)
2. Finin, T., Fritzson, R., McKay, D., McEntire, R.: KQML as an agent com-
munication language. In: Proceedings of the third international conference on
Information and knowledge management, pp. 456–463. ACM, New York (1994)
2
See the Rice viewer in Figure 2.
Multi-agent System for Blackout Prevention 337
Abstract. Robotic search and rescue teams of the future will consist of both
robots and human operators. Operators are utilized for identifying victims,
by means of camera feeds from the robot, and for helping with navigation
when autonomy is insufficient. As the size of these robot teams increases, the
mental workload on operators increases, and robots find themselves in pre-
carious situations with no assistance for resolution. This paper presents an
approach that utilizes multiple levels of autonomy to allow a robot to consider
a range of options, including asking for operator assistance, for dealing with
problematic situations to maximize efficient use of the operator’s time. Indi-
vidual robots use self monitoring to determine failures in task progression, a
form of local autonomy. Upon this trigger, the robot evaluates decisions to
properly route asking for help, consensus autonomy. A Call Center alerts the
operator(s) to incoming requests for assistance. This results in a better use of
operator time by focusing attention where it is needed. Experiments explore
the effectiveness of agents’ decisions, both local and team-level, with multiple
simulators. A high fidelity simulator and user interface further evaluate how
effective robot information is relayed to the operator, through human trials.
1 Introduction
Applications for multirobot systems (MrS) such as interplanetary con-
struction [10], search and rescue [6], or cooperating aerial vehicles [1] will
require close coordination and control between human operator(s) and teams
of robots in uncertain environments. Some functions of a MrS, such as iden-
tifying victims among rubble, depend on human input. In general, humans
are needed for situations where robotic sensing is inadequate or reasoning is
Breelyn Kane · Prasanna Velagapudi · Paul Scerri
Carnegie Mellon University, Pittsburgh, PA 15213
e-mail: breelynk@cmu.edu,pkv@cs.cmu.edu,pscerri@cs.cmu.edu
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 339–357.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
340 B. Kane, P. Velagapudi, and P. Scerri
length of the queue of requests at the call center. The robot also attempts
to estimate the length of time the operator will need to help it out of the
current situation. Options for the robot include simply giving up its task,
getting another robot to take over its task, or getting another robot to provide
context to the current situation leading to a decrease in the amount of time
the operator will take to resolve the robot’s problem. The outcomes of all
these actions are uncertain and the aim is to maximize overall performance,
hence Markov Decision Processes are a natural choice for representing the
reasoning.
This approach was extensively evaluated in two ways. First, the approach
was evaluated in a series of pilot tests with real operators in a high fidelity
robot simulation1 . This determined a baseline for the need of human inter-
vention in searching highly cluttered environments versus robots searching
completely autonomously. Second, to understand the properties of the ap-
proach under a more diverse set of circumstances, a simple simulation of the
approach was implemented and tested under a range of parameters. This
included varying the probability of robots getting stuck, the number of op-
erators, and the number of robots.
2 Problem
We intend to address the following questions: What is the best way to go
about providing an operator with more information, and focusing their efforts
to where they can be utilized most effectively? Additionally, how can robots
in teams simultaneously make smart decisions, based on current resources,
to autonomously search zones while continuing to progress in their task?
Defining the problem more abstractly, consider a team consisting of a set of
robots R = {r1 , r2 , r3 , . . .} and operators O = {o1 , o2 , o3 , . . .}. The team ob-
jective is to maximize the number of identified victims and total area searched
with respect to time. This objective can be represented as the completion of
a discrete set of exploration and discovery tasks G = {g1 , g2 , g3 , . . .} over the
map. These tasks are of varying importance to the team goal, as determined
by the joint reward function u(g).
Robots operate in one of several states, depending on model fidelity, but
they are primarily fixed, stuck, or dead. In order to complete a task g and
gain reward u(g), a robot must spend some number of timesteps in the fixed
state. However, environmental circumstances can cause robots to transition
to the stuck or dead states at each timestep. Operators have the ability to
allocate a number of their time steps to restoring robots in the stuck state
to the fixed state. However, they can only do this for one robot at a time.
For time t, this is represented by the assignment Ator of operator o to robot
1
While future evaluation will involve real robots, it is not feasible at this time
to conduct extensive user evaluation with sixteen autonomous robots in a large,
complex environment.
342 B. Kane, P. Velagapudi, and P. Scerri
r at time t. Over a given time horizon T , the team must allocate operator
attention in a way that maximizes completed tasks:
n
Uteam = arg max u(gi )
A0or ,...,AT
or
i=1
Fig. 1 Various stuck positions are displayed: The views on the left, (a) and (c),
show what the operator sees from the robot camera vs. the actual position of the
robot in the real world, corresponding views (b) and (d).
Asking for Help through Adaptable Autonomy 343
operator might try to assist a robot that is stuck, and spend an arbitrarily
long time concentrating on how to get it out of a small, completely mapped
room rather than marking victims. Or, the operator may not realize that get-
ting this particular robot unstuck, instead of a different robot that is stuck
in an area that is not yet mapped, hurts their ability to search a larger en-
vironment. Furthermore, the degree to which a robot is stuck may also vary.
All of these instances present challenges in how the operator must go about
balancing their current tasks with knowing which robot is critical and the
time it may take to free the robot. Illustrating this point, Figure 1 shows a
side by side display where a robot is stuck, and what an operator sees through
the robot’s forward camera versus an exocentric view of the current state of
the robot. In robotic search and rescue domains, robots are often stuck in
situations where even the most skilled operators have trouble understand-
ing the severity or simplicity of the situation. This is because, provided only
with egocentric camera displays, human operators have difficulty establishing
complete situation awareness [12]. This leaves an opening for potential con-
tributions from individual robots to provide more context to the user. The
robot can assist the operator by autonomously making decisions, both indi-
vidually and through the team, to accomplish the overall goal. Robots can
delegate focused requests to the operator with a sort of scalable adjustable
autonomy. Thus, the degree to which a robot should make decisions for the
operator on where to concentrate is also an important question.
Figure 1 shows side by side displays illustrating the differences between
what an operator sees through the robot’s forward camera, versus the exo-
centric view of the current state of the robot.
3 Concept
Having self-awareness at the robot level can be beneficial to operators in
multi-robot domains in dealing with these issues. The ability for robots to
internally initiate decisions and act as a low pass filter over noisy information
allows operators to focus on assisting higher priority robots, allowing the con-
trol of larger teams with less operators. In this work, we propose a multiagent
approach that divides the problem of managing operator assistance into sev-
eral subtasks that are simultaneously processed both on the robot and at the
operator interface level. An additional actor, the queue meta-level manager,
is introduced as a coordinator for interaction between operators and robots.
The motivation for this architecture comes from the structure of a call
center. A call center is a system that allows multiple operators to service
many requests and is robust to varying call loads. In this domain, robots
would represent customers, and human tele-operators their servers. Rather
than providing advice the operator is responsible for assisting robots in need.
Robots, operators, and the queue meta-level manager all act as indepen-
dent agents. Each agent is a separate process. This modularity allows for
344 B. Kane, P. Velagapudi, and P. Scerri
independent interaction. For example, robots assess their situation and make
informed decisions, while operators service requests, and a queue manager
acts as a switching board for incoming requests.
As robots travel through the environment the evaluation of their stuck
state triggers the creation of an internal reasoning proxy. This occurs when
the percentage that a robot is stuck exceeds some threshold ω. The reasoner
proxy acts as a self-contained piece of software that combines team level
information to reason about its next action. This reasoning produces a pol-
icy, Π, which returns the optimal matching action u of the current state x.
The series of possible actions are described in the MDP formulation given in
4.2. As described in Algorithm 1, if the robot decides to place a request for
the operators help, it sends the request to the queue manager. This manager
Algorithm 1
1: if P r(stuck) > ω then
2: Π = generateReasoner(RobotID)
3: u = Π(x)
4: if u == ask operator for help then
5: sendMessage(QMessage, RobotID)
6: waitForUpdates()
7: ut+1 = getNextAction()
8: else
9: execute alternate u
10: end if
11: else
12: executeTask()
13: end if
Asking for Help through Adaptable Autonomy 345
extracts the incoming request message, and places it on a literal queue. Statis-
tic information is flagged for this request, and calculated as necessary. The
queue manager can also send updated changes for values such as predicted
service times, and the varying length of the queue as they change dynami-
cally. The queue is upper bounded by the number of robot agents multiplied
by their number of request types. It is useful for the robot to take updated
queue statistics into account when deciding whether or not to get additional
help from a team member, or try other behaviors while waiting for help from
the operator. The communication between these entities is illustrated in the
system level diagram, Figure 2.
4 Algorithm
The approach has two major components that we will describe in detail in the
following sections. In one piece, the robot locally decides through monitoring
if it is unable to progress in its current task. This can be thought of as the
classification of a full or partial failure. Another section gives an analysis of
how the robot takes information from itself, the team, and properties of the
call center queue to decide how to go about resolving its current situation.
4.1 Self-monitoring
There are many instances where something unexpected occurs and prevents
a robot from progression or even completion within their prescribed task.
As has been stated above, if robots could provide feedback on their current
state through a sense of self monitoring, it is easier to anticipate when things
are not going as planned. If a seemingly naive robot is able to determine a
fail or partial fail scenario this can act as a trigger for making decisions to
help resolve the current situation. The current implementation is the begin-
ning steps towards robot introspection. As described in [2] full introspection
incorporates both a monitoring phase and a meta-reasoning phase. In this
model, sensor level information is processed and then sent to a metareasoner
which in turn ‘thinks about thinking’. Control algorithms are replayed back
to be translated into actions for the robot to execute. It is stated that the
current implementation emulates the beginning steps, because our approach
incorporates just the monitoring phase. Even though the metareasoning is
employed at a team level, the individual agent does not really evaluate if an
action is good or bad in determining its initial stuck state.
The robot determines if it’s ‘stuck’ with different sensory information ob-
tained while obstacle avoiding. The z-axis of the root reference frame is par-
allel to gravity but points up. The y-axis of the root reference frame points
laterally and is chosen according to the right hand rule. The x-axis of the
root reference frame points behind the robot. The rotation matrix is utilized
to know if a robot is flipped completely over, facing the ceiling, or facing the
346 B. Kane, P. Velagapudi, and P. Scerri
Fig. 3 Robots can become stuck in a variety of ways, with different amounts of
effort required to extricate them.
Asking for Help through Adaptable Autonomy 347
In this case, the MDP has four thousand states, choosing from five possible
actions. It converges in approximately fifty iterations. Each state of our MDP
has five dimensions represented with a variable for queue time, stuck state,
distance of helping robot, distance of a robot providing more information,
and a boolean indicator of whether or not this robot is currently waiting
on the queue. The stuck states were determined after experimentation with
P2AT robots in the USARSim domain, USARSim is described in 5.1. See
Table 1 for the explicit state listings.
There are five possible actions a robot can take in each state: {AskHelp,
HelpSelf, AskSelf, AskRForInfo, AskRForHelp}. Each of these corresponds
to a potential action the robot can take in response to its current stuck state.
The robot can ask for operator help, which places a request on the queue. The
robot can try and help itself, meaning that it will attempt to move on its own
348 B. Kane, P. Velagapudi, and P. Scerri
States
Queue Stuck State Distance Distance On
Length Helper Robot Robot provid- Queue
ing info
Secshort facedown there there on
SecMedium faceup short short off
SeLong wedged medium medium
MinOne goingoverobj long long
MineTwo cornerstuck unknown unknown
MinFour flippeddead
MinSix notstuck
MinEight unknownstuck
MinTen
MinLong
to unstick itself. The robot can reassess its current state. This means that
the robot takes movements to help clarify its stuck state if it is uncertain of
this state previously. Also, another robot can be called in to help the current
robot, this is with the intention of the helper robot getting the calling robot
unstuck. Lastly, the robot may ask another teammate to provide information.
This is where another robot can come and provide camera imagery for the
current robot, to help give the operator more situational awareness for better
assisting the robot.
Transition probabilities are generated by doing trials in the physics sim-
ulation USARSim, described below, to gauge an accurate approximation of
how the environment affects the dynamics of a robot’s stuck state. These
probabilities are static heuristics, but as the state and action space increases
these transition probabilities could be learned through an algorithm such as
linear regression which will help maintain weights on the feature space.
Getting in a terminal stuck state of flipped over, or a not stuck state, act
as the absorbing states for this model. The single step cost for a given state
is dependent on the current action. For example, robots that ask for operator
help incur a higher cost when helper robots are near. This happens because
close robots may act as additional obstacles for getting stuck. Robots that
decide to ask for other robots to assist them have a lower cost when the
queue length is less and a helper robot is near, this is because helper robots
are better at providing a different viewpoint for the operator and assisting
the robot in need. It is not assumed that once assistance is employed to
a stuck robot that it will become unstuck. The model tries to account for
the percentages of times when the robot actually may get more stuck. The
amount of assistance a helper, operator, or the current robot can provide at
a given time is a direct result of the current state and action.
Asking for Help through Adaptable Autonomy 349
Trends emerge based on how the robot chooses actions from an optimal
policy. When queue lengths are short, and the robot is not on the queue, the
optimal action for the majority of states is to ask the operator for help. Also,
when the queue length tends to be long, and the robot is on the queue, the
best action is for the robot to help itself. If the distance of a helper robot
is right near the current stuck robot, the robot tends to ask for robot help
more often when the queue is of a higher length. A condensed version of this
MDP is compared to other policies in the Simple Sim described below in 5.2.
5 Results
In order to explore the complex potential interactions between robots and
operators as well as the gross effects of various system parameters, experi-
ments were conducted in both a high-fidelity simulation environment with
human participants and a low-fidelity discrete event simulation with simu-
lated human operators.
5.1 USARSim
Pilot experiments were conducted in a high fidelity robotic environment de-
veloped as a simulation of urban search and rescue (USAR) robots and envi-
ronments, and intended as a research tool for the study of human-robot in-
teraction (HRI) and multi-robot coordination. USARSim is open source and
freely available2 . It uses Epic Games’ UnrealEngine2 to provide a high fidelity
simulator at low cost. USARSim supports HRI by accurately rendering user
interface elements (particularly camera video), accurately representing robot
automation and behavior, and accurately representing the remote environ-
ment that links the operator’s awareness with the robot’s behaviors. MrCS
(Multi-robot Control System), a multirobot communications and control in-
frastructure with accompanying user interface developed for experiments in
multirobot control and RoboCup competition [11] was used with appropriate
modifications in both experimental conditions. MrCS provides facilities for
starting and controlling robots in the simulation, displaying camera and laser
output, and supporting inter-robot communication through Machinetta [8],
a distributed multiagent system. This distributed control enables us to scale
robot teams from small to large. The operator interface for this experiment
can be seen in Figure 4.
Initial experiments comparing robots autonomously searching with the as-
sistance of a human operator did not provide conclusive evidence of needing
operator assistance. The cubicle map was then customized to include many
more obstacles including desks, ramps, and chairs. This increased the diffi-
culty of the task. In this environment we ran a series of ten experiments for
2
http://www.sourceforge.net/projects/usarsim
350 B. Kane, P. Velagapudi, and P. Scerri
Fig. 4 Setup and user interface for USARSim experiments. Users are presented
with streaming video feeds from each robot (left), a global map panel (lower right),
teleoperation controls (middle right), and notification queue (upper right).
Table 2 Results of USARSim simulation trials. Fully autonomous robots are com-
pared to robots being assisted by human operations in a direct supervisory and call
center role.
the requesting robot, but is unable to do so. The ignore button allows the user
to multi-task and not address current request when they may be overloaded
with other tasks.The results are displayed in Table 2.
The total search area achieved in conditions in which an operator assisted
the robots showed an increase in more than 25% over robots simply searching
autonomously with no assistance at all, proving that operator assistance is
helpful. The results between the call center and operator conditions did not
yield signficant differences in performance.
The results from the USARSim experiments do not conclusively show that
the call center approach improves performance over that of an operator which
decides what robots to assist independently. There may be several reasons for
this result. The current experiments had the robots functioning very unintel-
ligently. Once the robot determined it was stuck it waited for the operator’s
help possibly indefinitely, a naive policy. This significantly decreases the ef-
ficiency of the robots and may not be the optimal thing to do when the
operator is not available to address a particular request. The reason robots
are left idle in their down time is to allow an incremental approach to ex-
periments. Bringing in too many additional variables may skew results. The
results also did not take into account metrics that may be specific to the
operator. For example, operators that take a very long time to address each
robot request will tend to have other requests queue up in the meantime. Also
operators that may not be very good at un-sticking robots could possibly not
benefit from a call center scheme, as all of the operators used in these pilot
experiments were inexperienced. It does seem that with more operators and
robots, where an operator can not possibly address every situation in an op-
timal way, the call center setup is beneficial. Since a good baseline has been
established without negative results, the next logical step is to test varia-
tions among human subjects which include varying the operators, number of
robots, and the complexity of decision autonomy the robots can employ. To
gain insight into how these variants effect the system a simpler simulator was
designed that allowed for exploring all these variations to provide validity for
future wide-scale human experiments.
352 B. Kane, P. Velagapudi, and P. Scerri
5.2 SimpleSim
A simple discrete event simulation was used to explore a large number of
parameters when evaluating the policies generated by the MDP, as well as
different dynamics related to how a call center functions. The simple simulator
is made up of Unmanned Ground Vehicles (UGVs), and Unmanned Aerial
Vehicles(UAVs) that travel the span of a grid world at each timestep. All
robots create paths to goals through an A* planner. The UGVs travel around
planning a path to a goal marked with a ‘T’. The ability of the UGV to travel
to that ‘T’, without getting stuck, is a completion of their task. UAVs also
travel through the environment and generate paths, but they do not get stuck,
they are purely used as helper robots in the context of these experiments.
Squares in the grid environment are marked with different colors to represent
different stuck states such as the robot being terminally stuck versus stuck
and needing help. An underlying queue collects incoming help requests for
the operator. A simple operator model based on a Gaussian distribution is
used for varying queue resolution times.
In generating our results, the MDP model is constrained to the state
space of the simple simulator. An example of how initial states might tran-
sition is shown in Figure 5. There are only six hundred states with three
possible actions. All state values remain the same accept for the stuck
state and distance of a helper robot. The stuck state changes to include
{dead, stuck, unstuck}, and the distance of helper robots no longer is im-
portant, using only {long, unknown} states. The actions now only allow the
robot to ask an operator for help, ask itself for help, or call in an additional
robot to provide more information. The optimal policy gets fed into the sim-
ulator that maps states and actions.
This MDP policy is compared over varying parameters to two other simple
policies. These policies are, a policy that puts every stuck robot on the queue
(Is it good to always ask an operator for help?), and a policy where the robot
always tries to help itself before placing itself on the queue. The ’operator
policy’, where every stuck robot places itself on the queue, was chosen as it
is similar to the policy used in the human trials that were described in 5.1.
A series of experiments studied the effects of varying the number of oper-
ators, the number of robots, the proportion of dangerous map cells (affecting
how often robots were stuck), and the average time for an operator to help a
robot. In each experiment, one parameter was varied, while the others were
held at the nominal values given in Table 3.
Parameter Value
# of robots 256
# of operators 2
% dangerous cells 4
Average assist time 30
At each setting 100 simulation runs were executed, each lasting 1000 time
steps. Robots and tasks were initialized uniformly random for each run, with
tasks varying randomly in value. The performance of each run was measured
by a sum of complete task value, additionally recording average queue length
over the run. More specifically, as a task is completed the value for that task
was accumulated in the total utility. The results of the experiments can be
seen in Figure 6.
Queue length increases and value decreases as conditions become adverse
for the robots, as the number of dangerous cells increases in Figures 6(e)
and 6(f).
The behavior of the MDP is apparent in its effect on queue length across
all conditions. While other policies show an increase in the queue length as
conditions become increasingly adverse, the MDP policy becomes increasingly
frugal with operator requests, efficiently keeping the operator workload down
as the queue length plateaus to a constant length, as operators take longer
to assist 6(b), and when the number of operators are low 6(d), and when
amount of dangerous cells increase 6(f), and as the number of robots increase
while operators are held constant 6(h). This is a good result and supports
the need for autonomous robot decision making using the environment’s state
information.
Since the MDP is additionally aware of task utility, it is also capable of
prioritizing requests to improve performance, yielding higher overall value
than the other policies. Figure 6(c) shows an interesting trend where the
MDP policy’s value decreases over time but still finishes at a higher value
354 B. Kane, P. Velagapudi, and P. Scerri
250 50
200 40
Queue length
150 30
Value
100 20
50 10
0 0
20 40 60 80 100 20 40 60 80 100
Average timesteps for operator to assist Average timesteps for operator to assist
(a) (b)
250 50
200 40
Queue length
150 30
Value
100 20
50 10
0 0
1 2 3 4 5 1 2 3 4 5
# of operators # of operators
(c) (d)
250 50
200 40
Queue length
150 30
Value
100 20
50 10
0 0
0.05 0.1 0.15 0.05 0.1 0.15
Proportion of dangerous map cells Proportion of dangerous map cells
(e) (f)
800 150
600
Queue length
100
Value
400
50
200
operator
0 0
self
50 100 150 200 50 100 150 200
# of robots # of robots mdp
(g) (h)
than the other policies. This may occur because increasing the number of
operators does not necessarily lead to the assistance of more robots, especially
when the MDP does not always ask for operator help. Also interesting is the
Asking for Help through Adaptable Autonomy 355
trend as operators take longer to assist in Figures 6(a), this shows that the
accumulation of task value increases without the need of an operator, for the
optimal policy.
6 Related Work
For human robot teams in uncertain environments, it is necessary to balance
between robot autonomy and the use of teleoperation for robot assistance.
Others [4] explore the effectiveness of a robot that navigates through the en-
vironment by teleoperation and asks an operator for input. This is referred to
as collaborative control. While the framework described above also employs a
line of communication between the operator and robot agents, it additionally
utilizes team autonomy information to make more informed decisions. Also,
the robots proactively take actions when an operator may be busy. This al-
lows an efficient use of an operator’s time by balancing when it is necessary for
an operator to interact with a robot versus the robot acting autonomously.
This balance is illustrated in models between neglect time and interaction
time [3]. The model describes a point in robot autonomy that reaches degra-
dation caused by neglect tolerance. This suggests that overall tasks would
improve with an operator causing the operator to devote interaction time to
the robot. The architecture just presented addresses this balance without the
need for dedicated operators.
To be clear the focus of this work was not to evaluate the graphical user
interface between the operator and the system. Many previous work addresses
interface design. Some of this work [7] touches on how questions should be
relayed with the support of robot state information to users to help their
understanding. The work presented here does provide context to the user,
but focuses more on the underlying intelligent decision making of the system.
Additionally, previous work [13] conducted tests for interface designs specific
to the task of robotic search and rescue. Users often felt more comfortable
with interfaces that provide more information about the robot’s orientation.
A preference for fixed windows was noted based on user feedback explaining
how this allowed them to better judge the current situation. Although, an
evaluation of the interface design is not the purpose of this work, the system
architecture supports the need for presenting more situational awareness to
the user, thus increasing their comfort level, and improving their ability to
complete the task.
This represents a baseline for future experiments and directions in teams with
higher scale and more dynamic environments.
The major contributions of this work include the presentation of a sys-
tem architecture that addresses neglect and interaction time profiles. Also, a
robots adaptable autonomy in this architecture efficiently allocates operator
time, provides situational awareness to the operator, improves team perfor-
mance, and is robust to partial system failure.
Future work includes extending experimentation to real robot platforms,
and delving deeper into investigating each subset of the problem. Extensions
on self monitoring might incorporate the ability of a robot to evaluate how
good their decisions are, and if they can learn, or set up a prediction of areas
they may want to avoid based on knowing histories of getting stuck. The
robot’s current decision model can be expanded with additional states and
actions. This may include the use of heterogeneous agents, and the ability for
an agent to give help in addition to asking for help. A call center formulation
does not have to be constrained to just agents requesting help. The queue
can also be utilized in other applications such as UAVs for holding important
video playback data that the operator may need for better completing their
task.
References
1. Chaimowicz, L., Kumar, V.: Aerial shepherds: Coordination among uavs and
swarms of robots (2004)
2. Cox, M., Raja, A.: Metareasoning: A manifesto. BBN Technical (2007)
3. Crandall, J., Goodrich, M., Olsen, D., Nielsen, C.: Validating human-robot
interaction schemes in multitasking environments. IEEE Transactions on Sys-
tems, Man, and Cybernetics, Part A: Systems and Humans 35(4), 438–449
(2005)
4. Fong, T., Thorpe, C., Baur, C.: Multi-robot remote driving with collaborative
control. IEEE Transactions on Industrial Electronics 50(4), 699–704 (2003)
5. Macfadzean, R.H., Barber, K.S., Interrante, L., Martin, C.E.: Supporting dy-
namic adaptive autonomy for agent-based systems (1996)
6. Murphy, R., Stover, S.: Rescue robots for mudslides: A descriptive study of the
2005 La Conchita mudslide response. Journal of Field Robotics 25 (2008)
7. Rosenthal, S., Dey, A., Veloso, M.: How Robots Questions Affect the Accuracy
of the Human Responses. In: Proc. International Symposium on Robot-Human
Interactive Communication, pp. 1137–1142 (2009)
8. Scerri, P., Liao, E., Lai, J., Sycara, K., Xu, Y., Lewis, M.: Coordinating very
large groups of wide area search munitions. Theory and Algorithms for Coop-
erative Systems (2005)
9. Schurr, N., Marecki, J., Lewis, J., Tambe, M., Scerri, P.: The defacto system:
Training tool for incident commanders. In: Proc of the AAAI, vol. 20 (2005)
10. Sellner, B., Heger, F., Hiatt, L., Simmons, R., Singh, S.: Coordinated Mul-
tiagent Teams and Sliding Autonomy for Large-Scale Assembly. Proc of the
IEEE 94(7), 1425–1444 (2006)
Asking for Help through Adaptable Autonomy 357
11. Wang, J., Lewis, M.: Human control for cooperating robot teams. In: Proc of
the ACM/IEEE international conference on Human-robot interaction, p. 16.
ACM, New York (2007)
12. Wickens, C.D., Lee, J., Liu, Y.D., Gordon-Becker, S.: Introduction to Human
Factors Engineering, 2nd edn. Prentice-Hall, Inc., Upper Saddle River (2003)
13. Yanco, H.A., Keyes, B., Drury, J.L., Nielsen, C.W., Few, D.A., Bruemmer,
D.J.: Evolving interface design for robot search tasks: Research articles. J.
Field Robot. 24(8-9), 779–799 (2007)
Conceptual Framework for Design of
Service Negotiation in Disaster
Management Applications
1 Introduction
The increased complexity of socio-economic systems in combination with nat-
ural hazards might create conditions that can evolve into dangerous chem-
ical incidents in or nearby urban areas. These are complex processes that
face involved stakeholders with handling of chemical substances threatening
themselves, the population, as well as the natural environment. Moreover,
as we deal with increasingly complex chemical substances, the risks of using
them in a wrong way are higher. Places such as harbors, as well as indus-
trial areas (e.g. the port of Rotterdam) are constantly at risk because many
companies are using, transporting and/or processing dangerous substances in
those areas. Agencies specialized in environmental protection and emergency
management are continuously monitoring the situation with the goal of tak-
ing corrective actions when things tend to go wrong. For example, DCMR1 in
Costin Bădică · Mihnea Scafeş
University of Craiova, Software Engineering Department,
Bvd.Decebal 107, Craiova, 200440, Romania
e-mail: {scafes_mihnea,badica_costin}@software.ucv.ro
1
DCMR Milieudienst Rijnmond.
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 359–375.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
360 C. Bădică and M. Scafeş
the Netherlands and DEMA2 in Denmark are two organizations that protect
the people and the environment against chemical hazards. Their goal is to
monitor the situation and, when needed, to discover the source of the disas-
ter, handle the chemical substances, secure the dangerous area and eventually
decide to evacuate the population or take other protection measures.
However, in responding to a threat caused by an incident, these agencies
cannot and usually do not act in isolation. Efficient response to a chemical
incident requires collaborative effort by exchange of information and expertise
between several stakeholders that can belong to different organizations. For
example, information from the Police (the traffic volume, congestions that
might be present near the incident location) and the City Hall (what is the
population in the area of the incident and how is it distributed?) help agencies
to take decisions whether to evacuate or shelter affected people. Additionally,
the agencies might also use volunteers that are present in the incident area
and are able and willing to give precious information about the source of the
incident (do they smell something unusual? do they see weird smoke in the
area?) apart from their own specially trained staff.
Among the difficult challenges that environmental and emergency organi-
zations have to face during their responses to incidents, we mention:
• Difficulty in finding the source of an incident, or at least obtaining a list
of potential sources.
• Inability to get an accurate on-site situational overview.
• Difficulty in finding useful resources and experts and in establishing com-
munication lines between them for information exchange.
• Inability to log the incident handling.
• Inability to try “What if” scenarios that help the decision makers.
See [3] for other problems and typical scenarios that are faced by these
organizations.
In order to tackle these kinds of problems faced by organizations such as
DEMA and DCMR, the DIADEM project3 has been started. DIADEM aims
to develop a distributed information system that supports experts from var-
ious geographically distributed organizations (e.g. environmental agencies,
police, fire fighters) with heterogeneous techniques from various domains in-
cluding software engineering, decision making, human-computer interaction,
chemical engineering, environmental management, and crisis management.
The core platform of the DIADEM distributed information system – the
DIADEM Framework – is modeled using multi-agent systems techniques [17].
Software agents support human experts to achieve their tasks. They can
provide user interfaces to various tools that run on different devices that
experts might utilize or carry with them (e.g. PCs, laptops, PDAs, mobile
phones, etc.).
2
Danish Emergency Management Agency.
3
DIADEM: Distributed information acquisition and decision-making for environ-
mental management. See project website at http://www.ist-diadem.eu/
Conceptual Framework for Design of Service Negotiation 361
incident and of the service providers, like for example up-to-dateness or com-
pleteness of data, can be included in the negotiation subject. See [9] for a list
of quality of context parameters and methods to compute them.
This paper is structured as follows. In Section 2 we give an introduction
to negotiation in multi-agent systems, with a focus on service negotiation.
Section 3 follows with a discussion about the framework we propose, including
examples of utility functions, and worked out example. In section 4 we briefly
compare our framework with other existing negotiation frameworks. Finally,
in Section 5 we draw some conclusions and point to future work.
2 Service Negotiation
Negotiation is a process that describes the interaction between one or more
participants that want to agree on a subject by exchanging deals or proposals
about this subject. Negotiation about a service that one or more participants
agrees to provide to other participants is called service negotiation.
There are at least two interpretations for services in software systems:
Software as a service and Service as a software. In the context of the DI-
ADEM system, we assume both interpretations. The first interpretation is
more suited as a software delivery model by application service providers for
component-based distributed applications. In DIADEM, we develop special-
ized components that are wrapped by agents and have their functionalities
provided as software services. Some examples are decision making compo-
nents and prognosis components. They can provide an overview of the situ-
ation, predict how the situation will evolve, rank the predictions according
to certain criteria, etc. These capabilities of software data processing are
provided to requester agents (representing humans as well as other software
components) as software services.
We use the second interpretation by viewing a service performed by hu-
mans or organizational entities as being delivered through software and IT
infrastructure, i.e. software agents in our particular situation. For example,
human experts or entire organizations can advertise their physical capabilities
as services. Therefore, they can provide their services by agreeing to perform
physical tasks included in their set of capabilities and returning results via
software agents assigned to them. Additionally, on-site experts and trained
staff can use computing devices (mobile phones, laptops, etc.) to interact
with their personal software agents that act as an external interface of the
DIADEM system.
There are three important aspects that need to be addressed in the defini-
tion of a negotiation problem: negotiation protocols, negotiation subject and
negotiation strategies [8].
The protocol describes the public rules of the agents’ interaction process.
Agents must comply to the protocol of the negotiation in order to be able to
communicate. The protocol specifies the steps of negotiations, the types of
Conceptual Framework for Design of Service Negotiation 363
messages that can be exchanged as well as when messages can be send and/or
received. Among well known negotiation protocols, we mention Contract Net
(CNET) [15], Monotonic Concession Protocol (MCP) [2], and different types
of auctions [17].
A negotiation subject describes what is being negotiated between the part-
ners. In service negotiation, the negotiation subject is the provided service
with its associated parameters. One of the preliminary steps of designing a ne-
gotiation for a specific problem is to identify the type of the subject: whether
it is single-issue (e.g. only the service name) or multi-issue (e.g. service name,
time, quality measures, etc). The operations that can be performed over the
range of issues must be also identified: whether the set of issues can be modi-
fied or not (e.g. participants can add or remove issues) and whether the values
of issues can be changed or not and in what way.
An agent’s strategy represents the private decisions that the agent makes
during the negotiation in order to reach the desired agreement that maximizes
some form of welfare. The chosen strategy must be defined in accordance with
the negotiation protocol.
An initial evaluation of different negotiation mechanisms with respect to
their usefulness for the DIADEM system is presented in [14] and [2].
3 Conceptual Framework
Negotiations in DIADEM are managed by a specialized software component
– negotiation engine that is instantiated inside each agent. The negotiation
engine supports the addition of new negotiation protocols by offering an
interface with support for negotiation demands in DIADEM. The negotiation
engine is generic, i.e. it can run negotiations described by various negotiation
protocols that were adapted for the DIADEM domain.
Negotiations in DIADEM are one-to-many and multi-issue. In a one-to-
many negotiation, a single agent is negotiating simultaneously with a group
of one or more agents. Therefore, one-to-many negotiations assume a special
agent that has a different role than his opponents. For example, in CNET,
there is one manager and possibly multiple contractors. The manager part
of CNET protocol is different from the contractor part. The manager sends
and eventually re-sends calls for proposals, selects the winner(s) and waits for
task results, while the contractor(s) make proposals or refusals and provide
task results.
Another important characteristic of service negotiations in DIADEM is
that they are cooperative. Cooperativity stems from the fact that the overall
goal of negotiation participants is the optimization of the response to the
chemical incident. Basically this means that during a negotiation: (i) the
manager is the leading decision factor that will try to optimize the assignment
of the negotiation task to the best available contractor; (ii) the contractor will
364 C. Bădică and M. Scafeş
try to make the best proposal for serving the manager, taking into account
his current duties and availability.
In DIADEM, a typical situation occurs when one agent requests some data
that can be made available through services provided by other agents in the
system. We found useful to describe such situations by borrowing the role
names from CNET. So in DIADEM there are only two negotiation roles:
(i) manager associated to the requester agent and (ii) contractor associated
to the provider agents. Protocol descriptions will be divided into multiple
sections to describe each role, in particular manager and contractor roles in
DIADEM. Agents should be able to identify their roles from the protocol
description and the context where the negotiation is applied.
determine their values. A subject issue that can be negotiated has the fol-
lowing properties:
• Name: a sequence of characters that uniquely identifies an issue of a given
subject.
• Data type: the type of the value the issue is allowed to take. For example,
possible values of this property are: NUMBER that describes an integer
or real number and STRING that describes a sequence of characters.
• Value type: specifies if the issue can be modified in proposals or it should
keep its original value set by the manager. Possible values are: FIXED (this
describes a non-negotiable issue, i.e. and agents cannot propose values for
this issue) and DYNAMIC (agents can propose values for this issue).
• Value: in case of a fixed issue, the manager sets this property when he
defines the subject.
Depending on their role – manager or contractor, participant agents can
privately configure the negotiation issues using additional properties:
• Monotony (for both manager and contractor): specifies whether the man-
ager or contractor prefers high values or low values of this issue. Possible
values are: INCREASING (the agent prefers high utility values of the is-
sue) and DECREASING (the agent prefers low utility values of this issue).
366 C. Bădică and M. Scafeş
ua : X → [0, 1]
|xi − x∗i |
fi (xi ) = max
xi − xmin
i
where x∗i is the reference value assigned by the manager to the issue i.
For text valued issues, the partial utility function might measure the se-
mantic difference between two texts or simply their lexical difference.
As issue utilities have values in the interval [0, 1] and the issues weights
are normalized, the utility function of the manager evaluates a proposal to
a real number in the interval [0, 1]. Value 0 means that the proposal is the
worst for the manager, while value 1 means the ideal proposal, i.e. the best
for the manager.
select the deals that he will propose to the manager. The utility function of
the contractor agent is a mapping from his deal space to a real number in
the interval [0, 1] as follows:
uc : X → [0, 1]
The utility function of the contractor agent can be defined as shown below:
uc (x) = 1 − ef f ort(x)
Here ef f ort(x) represents the total effort that needs to be deployed by con-
tractor agent c to be able to provide the terms and conditions specified by
proposal x of the manager. The contractor will obviously choose to propose a
deal that maximizes his utility, i.e. minimizes the required effort for achieving
the task required by the manager.
In practice function ef f ort(x) can take into account several factors, in-
cluding for example the cost of providing service x (i.e. cost(x)), the existing
commitments of agent c that were previously contracted and are not yet
finalized, and/or the amount of resources that are required to provide the
service.
Obviously, a higher number of commitments not finalized yet of agent c
or a higher cost of providing x will result in a higher value of ef f ort(x)
and consequently to a lower value of the utility function uc (x) of contractor
agent. The ef f ort function can be defined in a similar way to the optimizer
in [13]. For example, the agent might require a smaller effort if the task he
is negotiating depends on another task that the agent is currently working
on. For example, an agent located in the port of Rotterdam working on a
previous task would have to invest a smaller effort for a new task requiring
measurements of gas concentrations in the port of Rotterdam than for a
new task requiring the same type of measurements but in another location,
because in order to accomplish the former task the agent would not need to
move to another location.
Moreover, in the domain of disaster management, participation of contrac-
tor agents in negotiations might be constrained by the availability of required
resources. For example, an agent should not be able to carry out in parallel
two tasks that both require the associated expert to be in two different places
at the same time. Additionally, agents should not be able to perform in par-
allel too many tasks as they would be overwhelmed. To model such situations
we can endow agents with resources and constrain their participation in ne-
gotiations for service provisioning by the availability of these resources. The
load factor of an agent represents the degree of unavailability of the agent,
i.e. the higher is the agent load factor, the less available is the agent to pro-
vide services. This factor can be estimated by the percent of resources from
the total amount of resources available to the agent that are currently in use
by the agent to provide already contracted services. The agent load factor is
represented by a real number between 0 and 1. Using the load factor, the cost
Conceptual Framework for Design of Service Negotiation 369
for providing the service can be determined as the percent of resources that
will be added to the agent load factor if the agent will agree to provide the
services under the terms and conditions of the deal. If the agent has a load
factor of 0 then all his resources are available (this means he is not performing
any task at the moment), while if the load factor is 1 the agent cannot engage
in new negotiations until he finishes some of the currently performed tasks.
The load factor is very important to define the agent strategy for making
proposals, as the agent cannot propose a deal requiring more resources than
the available resources.
3.5 Example
In this section we consider a simple practical example of how the framework
supports negotiation in a disaster management scenario. Let us suppose that
a manager agent needs to find a provider for the service “Measure gas con-
centration”. For simplicity, we will not take into account the precision or
location of the measurement, so time is the only negotiable parameter of the
service. In this context, time means the time duration in which the mea-
surement of the gas concentration will be available. Therefore, the manager
agent starts a negotiation concerning the negotiation subject (service: “Mea-
sure gas concentration”, time). service is a fixed issue with the meaning that
agents cannot propose deals in which this issue has a value different than the
initial value specified by the manager. time is a dynamic issue and agents can
propose any value from their deal space for this issue. Let’s suppose there are
three other contractor agents in the system, as described in table 2.
Neg. Manager C1 C2 C3
step
1 Define subject (service: - - -
Measure gas concen-
tration, time) and role
(manager)
2 send CFP to C1, C2, C3 receive CFP receive CFP receive CFP
3 - search the deal search the deal search the deal
space and decide space and decide space and decide
4 receive proposals from propose deal 1 propose deal 2 refuse
C1, C2, C3
5 Compute utilities, send Agreement Conflict Conflict
agreement to C1, conflict
to C2 and C3
6 End negotiation, wait for Start processing End End
results
7 Receive task information Send task infor- - -
from C1, End mation, End
Table 4 Utilities of the deals received by the manager. As agent C3 did not send
a proposal, it is not mentioned in the table
Sender Utility
C1 0.5 + 0.5 * 0.99 = 0.995
C2 0.5 + 0.5 * 0.95 = 0.975
When computing the utilities in table 4 the manager takes into account
the following:
• the service issue takes text values;
• the time issue takes number values;
• the domain of the time issue is [0, 100];
• the time issue is descending;
• the reference value of the time issue is 0;
• all issues have equal weights of value 0.5.
Note that the issues (service and time) as well as the default values of their
properties (reference values, weights, monotony, data types, value types) are
defined at the first level of negotiation configuration.
The manager selects the sender of the deal with the highest utility as the
service provider. According to table 4, it is not difficult to observe that the
manager agent would pick the option proposed by agent C1 .
Note that in this example we have included among the possible contractors
agents that are not able to provide the required service – agent C3 in our
example. However, in practice the manager will send the call for proposal
372 C. Bădică and M. Scafeş
only to agents that are (at least in theory) able to provide the required ser-
vice by first querying a Yellow Pages server. Note that this does not mean
that contractors receiving the call for proposal will necessary propose some-
thing, because they might be too busy or they would not be able to propose
an appropriate deal that matches the negotiation subject (for example, the
manager might add to subject an extra issue that some agents might not be
aware of).
4 Related Work
Recently, a service negotiation framework has been proposed in paper [10].
That framework was applied to model Web service negotiations between in-
surance companies and car repair companies. The negotiation is implemented
as an interaction between web services and the participants’ preferences are
represented using an ontology. The framework is utilizing the CNET protocol
for negotiation, implemented as Web service method invocations. Note that
in our proposal we do not constrain the negotiation protocols to a particu-
lar interaction protocol. Rather, we define a set of generic negotiation steps
that are likely to be followed by many negotiation protocols. Our framework
is generic in the sense that it allows creation and integration of different
negotiation protocols. The approach of combining generic negotiation with
components for representation of negotiation participants’ deal space, util-
ities, and resources allows us to design and evaluate different negotiation
protocols useful for service negotiation in disaster management. Also, differ-
ent from [10] where the authors utilize a fixed set of subject issues specially
tailored for their scenario, in our approach we give the possibility to define
new subject issues with their properties that suit best the application in
hand. Additionally, contractor utilities have a different semantics inspired by
task allocation problems ( [13]) that is more appropriate for the collaborative
rather than competitive context for which our system is designed. However,
similarly to [10], we use additive weighted utility functions for managers that
take into account partial utility functions defined for each negotiation issue.
In [1] the authors discuss SAMIN – System for Analysis of Multi-Issue
Negotiation. SAMIN is an advanced Prolog-based software for analysis of
one-to-one, multi-issue, both automated and human negotiations. In our sys-
tem we are dealing with one-to-many service negotiations. We would like to
develop a formalization of negotiation processes similar to the work in [1], by
generalizing the results to one-to-many and many-to-many negotiations.
In [16] the authors discuss applications and improvements of the max-
sum algorithm for the coordination of mobile sensors acting in crisis situa-
tions. Our negotiation system differs from the coordination system in [16].
While the agents in our systems also need to coordinate, they are doing it
by establishing workflows of services. Agents that require services negotiate
with agents that are capable of providing those services using one-to-many
Conceptual Framework for Design of Service Negotiation 373
References
1. Bosse, T., Jonker, C.M., Meij, L., Robu, V., Treur, J.: A system for analysis of
multiissue negotiation. In: Software Agent-Based Applications, Platforms and
Development Kits, pp. 253–280. Birkhaeuser Publishing Company (2005)
2. Bădică, C., Scafeş, M.: One-to-many monotonic concession negotiation proto-
col. In: Proc. of 4th Balkan Conference in Informatics – BCI 2009, pp. 8–13.
IEEE Computer Society Press, Los Alamitos (2009)
3. Damen, D., Pavlin, G., Kooij, V.D., Bădică, C., Comes, T., Lilienthal, A.,
Fontaine, B., Schou-Jensen, L., Jensen, J.: Diadem environmental management
requirements document, issue 1.13.1 (2009)
4. Dobriceanu, A., Biscu, L., Bădică, A., Bădică, C.: The design and implementa-
tion of an agent-based auction service. International Journal of Agent-Oriented
Software Engineering (IJAOSE) 3(2/3), 116–134 (2009),
http://dx.doi.org/10.1504/IJAOSE.2009.023633
5. Faratin, P., Sierra, C., Jennings, N.R.: Negotiation decision functions for au-
tonomous agents. Int. Journal of Robotics and Autonomous Systems 24(3-4),
159–182 (1998), http://eprints.ecs.soton.ac.uk/2117/
6. Hattori, H., Klein, M., Ito, T.: A multi-phase protocol for negotiation with
interdependent issues. In: IAT 2007: Proceedings of the 2007 IEEE/WIC/ACM
International Conference on Intelligent Agent Technology, pp. 153–159. IEEE
Computer Society, Washington (2007)
7. Hemaissia, M., El Fallah Seghrouchni, A., Labreuche, C., Mattioli, J.: A mul-
tilateral multi-issue negotiation protocol. In: AAMAS 2007: Proceedings of the
6th international joint conference on Autonomous agents and multiagent sys-
tems, pp. 1–8. ACM Press, New York (2007)
8. Jennings, N.R., Faratin, P., Lomuscio, A.R., Parsons, S., Sierra, C., Wooldridge,
M.: Automated negotiation: Prospects, methods and challenges. International
Journal of Group Decision and Negotiation 10(2), 199–215 (2001),
http://eprints.ecs.soton.ac.uk/4231/
9. Manzoor, A., Truong, H.L., Dustdar, S.: On the evaluation of quality of context.
In: EuroSSC 2008: Proceedings of the 3rd European Conference on Smart Sens-
ing and Context. LNCS, vol. 4405, pp. 140–153. Springer, Heidelberg (2008)
10. Paurobally, S., Tamma, V., Wooldrdige, M.: A framework for web service ne-
gotiation. ACM Trans. Auton. Adapt. Syst. 2(4), 14 (2007),
http://doi.acm.org/10.1145/1293731.1293734
Conceptual Framework for Design of Service Negotiation 375
11. Pavlin, G., Kamermans, M., Scafeş, M.: Dynamic process integration frame-
work: Toward efcient information processing in complex distributed systems.
In: Papadopoulos, G., Bădică, C. (eds.) Intelligent Distributed Computing III.
SCI, vol. 237. Springer, Berlin (2009)
12. Pavlin, G., Wijngaards, N., Nieuwenhuis, K.: Towards a single information
space for environmental management through self- configuration of distributed
information processing systems. In: Hrebicek, J., Hradec, J., Pelikan, E.,
Mirovsky, O., Pillmann, W., Holoubek, I., Bandholtz, T. (eds.) Proceedings
of the TOWARDS eENVIRONMENT Opportunities of SEIS and SISE: Inte-
grating Environmental Knowledge in Europe, Masaryk University (2009)
13. Sandholm, T.W.: An implementation of the contract net protocol based on
marginal cost calculations. In: Proceedings of the 12th International Workshop
on Distributed Artificial Intelligence, pp. 295–308. Hidden Valley, Pennsylvania
(1993), citeseer.ist.psu.edu/sandholm93implementation.html
14. Scafeş, M., Bădică, C.: Service negotiation mechanisms in collaborative pro-
cesses for disaster management. In: Eleftherakis, G., Hannam, S., Kalyva, E.,
Psychogios, A. (eds.) Proceedings of the 4th South East European Doctoral
Student Conference (DSC 2009), Research Track 2: Information and Commu-
nication Technologies, pp. 407–417. SEERC, Thessaloniki (2009)
15. Smith, R.G.: The contract net protocol: High-level communication and con-
trol in a distributed problem solver. IEEE Trans. Comput. 29(12), 1104–1113
(1980), http://dx.doi.org/10.1109/TC.1980.1675516
16. Stranders, R., Farinelli, A., Rogers, A., Jennings, N.: Decentralised coordination
of mobile sensors using the max-sum algorithm. In: Proc 21st Int. Joint Conf
on AI, IJCAI (2009), http://eprints.ecs.soton.ac.uk/17268/
17. Vidal, J.M.: Fundamentals of Multiagent Systems: Using NetLogo Models
(2006) (unpublished), http://www.multiagent.com/fmas
An Optimized Solution for Multi-agent
Coordination Using Integrated
GA-Fuzzy Approach in Rescue
Simulation Environment
1 Introduction
The main objective in rescue simulation environment is to decrease the
amount of damages imposed to civilians and city by rescuing buried and
injured civilians and extinguishing the burning buildings as much as possi-
ble while a simulated earthquake happens. Three types of agents including
Fire brigades, Ambulances and Police forces act in this environment. In or-
der to reach high-performance coordination, agents should effectively share
a large amount of information which they percept from observed objects in
surrounding environment. Due to lack of time to process and limited capacity
Mohammad Goodarzi · Ashkan Radmand · Eslam Nazemi
Faculty of Electrical and Computer Engineering, Shahid Beheshti University, Iran
e-mail: {mo.goodarzi,a.radmand}@mail.sbu.ac.ir,nazemi@sbu.ac.ir
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 377–388.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
378 M. Goodarzi, A. Radmand, and E. Nazemi
2 Environment Categorizing
2.1 Data Categorizing
Figure 1 demonstrates a categorization of entities in rescue simulation envi-
ronment [7].
According to figure 1, each entity has a set of states that describes its
condition and agents can do some limited actions based on their type. At the
other hand, every change in rescue environment is the result of state-changing
or agents’ actions. Therefore, the set of all message types can be defined as
follows:
p
{M essageT ypes} = ∪m
i=1 ∪j=1 State(i, j) + ∪k=1 ∪l=1 Action(i, j)
n o
(1)
Where,
State(i, j) is j th state of the ith entity,
Action(k, l) is lth action of the k th agent,
m is number of the entities,
n is number of the ith entity’s states,
o is number of the actions
and p is number of the k th agent’s actions.
An Optimized Solution for Multi-agent Coordination 379
The above equation covers all the events in the environment according
to the entities and agents’ actions. Therefore, by using equation (1) for the
entities which demonstrated in figure 1 and compounding the messages which
express similar information, we can obtain the entire message types. Table 1
shows the most important message types generated through (1).
Some chromosomes that have the most fitness value will be selected to
make the next generation automatically. Rest of the needed chromosomes
will be extracted from the crossover and mutation operations.
Fig. 5 (a) Burning buildings parameter, (b) Blocked roads parameter, (c) Not
rescued civilians/agents parameter Membership Functions
4.3 Defuzzification
With defining fuzzy sets, membership functions and fuzzy if-then rules, the
system is ready to evaluate weight of message types for any given condition.
Figure 6 indicates the fuzzy system’s behaviour.
In defuzzification step, to achieve final weights for the given situation,
then parts of all rules should be combined considering their trigger value.
The method used for defuzzification step in this paper is Weighted Average
method.
5 Experimental Results
The performance of proposed method is directly related to GA’s output. An
iterative procedure is devised in order to reach the desirable performance.
The iteration is based on extracting training situations, trying to generalize
them and evaluate the final result and it continues until the given result
reaches the expected performance and become admissible.
Our goal in this paper is to reach the top score of about 20 maps which had
been used as communication-based maps in China2008 and IranOpen2009
RoboCup Rescue Agent Simulation competitions.
Based on our experience in RoboCup rescue simulation system, we assume
four partitions for each parameter in the fuzzy logic system as a start point
(by partition we mean the definition of partition introduced by Takagi-Sugeno
fuzzy modeling system). Thus, the first iteration of obtaining results has 43 =
64 training situations. Table 3 and figure 7 show the detailed average scores
and the final score progression respectively while the amount of training data
6 Conclusion
In this paper, we proposed an integrated solution based on GA and Fuzzy
logic in order to improve the agents’ coordination in multi-agent environ-
ments specifically in RoboCup rescue simulation system. We categorized all
the observed data in the environment to a set called Message Types. Using
these types, we produced a set of parameters that can fully describe the
environment on any possible communication-based situation. Afterward, we
utilized a GA-Fuzzy method to first generate a set of specific solutions to
some certain scenarios and then generalize them to the entire simulation sys-
tem. In an iterative procedure, we reached a satisfiable result that improved
agents’ coordination in any unknown scenarios.
7 Future Works
In all multi-agent environments with large number of effective parameters
finding an approximate optimal answer for a specific situation is a serious
problem. We used a method based on Genetic Algorithms in this paper and
obtained desirable results. However, Long time of execution for evaluating
the fitness values in RoboCup rescue simulation system can not be ignored.
We are looking to devise new faster method that can obtain suitable result
in much less time. We have also looking to improve the proposed method in
order to handle the scenarios which does not have any radio communications
between agents.
One of the research topics in information sharing methods is to select
a subset from the entire agents whom to send the messages according to
data characteristics [11]. We are also looking to research in this aspect of
information sharing methods and algorithms.
References
1. Al jaddan, O., Rajamani, L.: Improved Selection Operator for GA. Journal of
Theoretical and Applied Information Technology, 269–277 (2005)
2. Banzhaf, W., Nordin, P., Keller, R., Francone, F.: Genetic Programming - An
Introduction. Morgan Kaufmann, San Francisco (1998)
388 M. Goodarzi, A. Radmand, and E. Nazemi
3. Dubois, D., Prade, H.: Fuzzy Sets and Systems: Theory and Applications. Aca-
demic Press, New York (1980)
4. Goldberf, David, E.: Genetic Algorithms in Search, Optimization and Machine
Learning. Kluwer Academic Publishers, Boston (1989)
5. Holland, J.H.: Adaptation in Natural and Artificial Systems. University of
Michigan Press, Ann Arbor (1975)
6. Kruse, R., Gebhardt, J., Kalwonn, F.: Foundations of Fuzzy Systems. John
Wiley & Sons, New York (1994)
7. Morimoto, T.: How to Develop a RoboCup Rescue Agent. Rescue Technical
Committee (2002)
8. Radmand, A., Nazemi, E., Goodarzi, M.: Integrated Genetic Algorithmic and
Fuzzy Logic Approach for Decision Making of Police Force Agents in Rescue
Simulation Environment. In: Baltes, J., Lagoudakis, M.G., Naruse, T., Ghidary,
S.S. (eds.) RoboCup 2009: Robot Soccer World Cup XIII. LNCS, vol. 5949, pp.
288–295. Springer, Heidelberg (2010)
9. Siddhartha, H., Sarika, R., Karlapalem, K.: Score Vector: A New Evaluation
Scheme for RoboCup Rescue Simulation Competition 2009, Rescue Technical
Committee (2009)
10. Takagi, T., Sugeno, M.: Fuzzy Identification of Systems and Its Applications
to Modeling and Control. IEEE Transactions on systems, man and cybernetics
SMC-15(1) (January/February 1985)
11. Xu, Y., Lewis, M., Sycara, K., Scerri, P.: Information Sharing in Large Scale
Teams. In: AAMAS 2004 Workshop on Challenges in Coordination of Large
Scale MultiAgent Systems (2004)
12. Zimmermann, H.-J.: Fuzzy Set Theory and its Applications, 2nd edn. Kluwer
Academic Publishers, Boston (1991)
A Study of Map Data Influence on
Disaster and Rescue Simulation’s
Results
1 Introduction
In disasters, human behaviors range from rescue operations to evacuation ac-
tivities. We cannot do experiments on human evacuation or rescue behaviors
physically at the level of real scale. Agent-based simulation makes possible to
simulate the human behaviors and interactions between their rescue opera-
tions and disasters. The simulation results can be used to predict damages to
civil life, and to make plans to decrease the damages. The RoboCup Rescue
Simulation (RCRS) is an agent based simulation system and was designed
to simulate the rescue measures taken during the Hanshin-Awaji earthquake
disaster [5].
RCRS has been used in RoboCup competition since 2000, and has been
used in various situations [8]. Most of the usages come from the require-
ment to decrease the damage of disasters. Applying the simulation systems
to practical usage requires that the results of simulation are explainable to
Kei Sato · Tomoichi Takahashi
Meijo University
e-mail: e0428299@ccalumni.meijo-u.ac.jp,
ttaka@ccmfs.meijo-u.ac.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 389–402.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
390 K. Sato and T. Takahashi
Fig. 2 Architecture overview of RCRS. Components are plugged into kernel via
specified protocol.
Standards for Geographic Information (JSGI) 2.0 since 2002 [2]. From
2007, they have updated the JSGI as ISO 19100 compatible one and have
started GSI data available from the Internet [3]3 .
Ordnance Survey(OS): Ordnance Survey is Great Britain’s national map-
ping agency. It has provided digital data of OS Master Map. The digital
data are sold to business and individuals [6].
Open Street Map(OSM): OpenStreetMap creates and provides free geo-
graphic data. As the data are input by volunteer, the data do not cover
all area. But they can be used without legal or technical restrictions [7].
data start year coordinate system road network road property building data
OSM 2004 WGS84 ◦ ◦
√ √
OS Topo 2001 British National Grid
√
OS ITN 2001 British National Grid
√ √
GSI 2002 Tokyo Datum
GSI 2007 WGS84 ◦ ◦
◦ indicates the dates are available free.
OS ITN:OS Integrate Transport Network Layer [4].
OS Topo:OS Topography Layer [12].
Fig. 3 Road network map from OSM (left), building data from GSI (middle) and
composite map (right) of Tokorozawa, Japan
5
OSM and GSI(2007) do not cover the whole of Japan.
396 K. Sato and T. Takahashi
Time Step 1 Time Step 100 Time Step 200 Time Step 300
Time Step 1 Time Step 100 Time Step 200 Time Step 300
Time Step 1 Time Step 100 Time Step 200 Time Step 300
Table 3 Spread of fire and correlation between two maps created by Moritz pro-
gram and ours.
Fig. 9 Chronological change of unburned building area rate for map Tokorozawa1.
While the same numbers of ignition points are allocated in cases of S1 and
S2, more ignition points are allocated to bigger houses in cases S3 and S4 on
assumption that there are many origins of fire in a big building.
At Figure 6, 7 and 8, the ignition points are set within some area. Table 5
shows correlations between the simulations using maps created by Moritz and
our programs. For example,
• Column S1:S1 shows the correlation between the result of scenario1 at
Moritz map and ones of scenario1 at our program map.
• Column S2:S4 shows the correlation between the result of scenario2 at
Moritz map and ones of scenario4 at our program map.
The values are different from ones in Table 3 because the disaster situations
are different and rescue agents are included in the simulations.
400 K. Sato and T. Takahashi
Fig. 10 Chronological change of unburned building area rate for map Tokorozawa2.
Fig. 11 Chronological change of unburned building area rate for map Tokorozawa3.
1. The values in Table 5 are strongly positive, except the correlation of un-
burned building between S1:S1 in Tokorozawa2 and S2:S4 in Tokorozawa3.
Followings are thought to cause the two low values.
a. low correlation for S1:S1 in Tokorozawa2
Figure 9-11 show the chronological changes of unburned building area
rate for three maps. Fire spread stops from 90 time step in a case of S1
A Study of Map Data Influence 401
using our program map in Tokorozawa2. (Figure 10) The fire spread is
affected by size of building as described in subsection 4.2.
b. low correlation for S2:S4 in Tokorozawa3
Fire spread stops from 200 time step in a case of S2 using our program
map in Tokorozawa3. (Figure 11) On the other hand, fire spreads wider
from 200 time step in a case of S4 using Moritz map in Tokorozawa3,
because of large building begin to burn from about 180 time step.
2. The deviations of simulation results are different among scenarios and
maps. The difference is thought to come from fire fighting by fire brigade.
The deviation of unburned building of S2 using our program in Toko-
rozawa1 (Figure 9)) is the biggest. The fire spread stopped until 300 time
step for 3 cases.
3. The chronological change using our program maps looks like inverse pro-
portion. While fire spreads within a block, it seems difficult that fire
spreads over roads. This becomes noticeable when the sizes of buildings
are small and it is thought to cause the unburned rate looks like inversely
in our maps.
References
1. Göbelbecker, M., Dornhege, C.: Realistic cities in simulated environments - an
open street map to robocup rescue converter. In: Fourth International Work-
shop on Synthetic Simulation and Robotics to Mitigate Earthquake Disaster
(SRMED 2009) (2009)
2. http://zgate.gsi.go.jp/ch/jmp20/jmp20_eng.html
(September 5, 2009)
3. http://www.gsi.go.jp/ENGLISH/page_e30210.htm
(September 5, 2009)
4. http://www.ordnancesurvey.co.uk/oswebsite/products/
osmastermap/userguides/docs/OSMM_ITN_userguide_v1.0.pdf
(September 5, 2009)
5. Kitano, H., Tadokoro, S., Noda, I., Matsubara, H., Takahashi, T., Shinjou, A.,
Shimada, S.: Robocup rescue: Search and rescue in large-scale disasters as a
domain for autonomous agents research. In: IEEE International Conference on
System, Man, and Cybernetics (1999)
6. http://www.ordnancesurvey.co.uk/oswebsite/ (September 5, 2009)
7. http://wiki.openstreetmap.org/wiki/Main_Page
(September 5, 2009)
8. Takahashi, T.: RoboCup Rescue: Challenges and Lessons Learned, ch. 14, pp.
423–450. CRC Press, Boca Raton (2009)
9. Takahashi, T., Ito, N.: Preliminary study to use rescue simulation as check soft
of urban’s disasters. In: Workshop: Safety and Security in MAS (SASEMAS)
at AAMAS 2005, pp. 102–106 (2005)
10. Takahashi, T., Takeuchi, I., Matsuno, F., Tadokoro, S.: Rescue simulation
project and comprehensive disaster simulator architecture. In: IEEE/RSJ Inter-
national Conference in Intelligent Robots and Systems(IROS 2000), pp. 1894–
1899 (2000)
11. Takeuchi, I., Kakumoto, S., Goto, Y.: Towards an integrated earthquake dis-
aster simulation system. In: First International Workshop on Synthetic Sim-
ulation and Robotics to Mitigate Earthquake Disaster (2003), http://www.
dis.uniroma1.it/˜rescue/events/padova03/papers/index.html
12. http://www.ordnancesurvey.co.uk/oswebsite/products/
osmastermap/userguides/docs/OSMMTopoLayerUserGuide.pdf
(September 5, 2009)
13. de Walle, B.V.(edited), Tyroff, M.: Emergency response information systems:
Emerging trends and technologies. Communications of the ACM 50(3), 28–65
(2007)
Evaluation of Time Delay of Coping
Behaviors with Evacuation Simulator
1 Introduction
Terrorist acts using chemical and biological agents and radioactive (CBR)
mate-rials have typically been nonselective attacks on crowds in urban areas.
Tomohisa Yamashita
National Institute of Advanced Industrial Science and Technology (AIST),
2-3-26 Aomi, Koto-ku, Tokyo 135-0064 Japan
e-mail: tomohisa.yamashita@aist.go.jp
Shunsuke Soeda
National Institute of Advanced Industrial Science and Technology (AIST),
1-1-1 Umezono Tsukuba, Ibaraki 305-8568 Japan
e-mail: shunsuke.soeda@aist.go.jp
Itsuki Noda
National Institute of Advanced Industrial Science and Technology (AIST),
1-1-1 Umezono Tsukuba, Ibaraki 305-8568 Japan
e-mail: I.Noda@aist.go.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 403–414.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
404 T. Yamashita, S. Soeda, and I. Noda
Advancesoft. First, the prediction systems of indoor and outdoor gas dif-
fusion calculate the concentration of hazardous gases. The output of these
systems is a time series of gas concentration in designated areas. Then, the
pedestrian simulator calculates the evacuation time of all evacuees and the
cumulative exposure of each evacuee using a time series of gas concentration.
An dataflow of our system is presented in Fig. 1.
where Vf represents free flow speed of the pedestrians, which is the speed
of the pedestrian when not in a crowd, and di represents the density of the
pedestrians on link i.
Fig. 4 shows the relationship the speed of pedestrian and the density of
the link. The exception to this formula is the pedestrian on the head of a
crowd on the link. For this pedestrian, Vf is used regardless of how the link
is crowded. Note that when the density of a link exceeds 4 pedestrians/m2 ,
Evaluation of Time Delay of Coping Behaviors 407
(a)High-rise building
(b)Commercial complex
all the pedestrian on the link cannot move except for the one who is on the
head of the crowd.
2.1.3 Confluence
Confluences - where two or more paths meet together - slow down the speed of
pedestrian. To illustrate slow-down by confluence, we used a simple model of
limiting the number of the pedestrian who could enter a link. The maximum
number of the pedestrian entering link lout shown in Fig. 5 is determined
from the width of the link. When there are pedestrians on lout already, the
number of the pedestrians that could enter lout is decreased at some ratio.
408 T. Yamashita, S. Soeda, and I. Noda
When there are more than two links where the pedestrians are trying to enter
lout , this number is divided among the links depending on the number of the
pedestrians trying to enter lout . Also, in this case, the total number of the
pedestrians able to enter lout is also reduced.
3 Simulation
With our evacuation planning assist system, we investigated a simulated
chemical attack on a major rail station because of a request from the Fire
and Disaster Management Department of Kitakyushu City. In our simula-
tion, for enlightening the managers of the rail station, we showed a relation
between the time necessary to begin coping behaviors of the managers and
the damage to passengers.
amount of exposure
(mg · min/m3 ) 1 200 1,000 2,000 20,000
Influence in behavior Pain in the Nausea & Breathing Lethal dose Lethal dose
eye & throat headache trouble 50% 100%
Implementation in Decrease in Decrease in
pedestrian simulation speed (-40%) speed (-90%) Stop Stop Stop
Damage level Mild Moderate Severe Severe Severe
Table 2 Coping behaviors of the managers and the times required for them
Time required
to begin(min)
Manager Coping behavior quick slow
Conventional line 1 Detecting chemical attack 5 10
Reporting to the fire department
2 and the other managers 3 6
3 Shutting down the trains 3 6
4 Ordering an evacuation to the passengers 3 6
New bullet train line 5 Ordering an evacuation to the passengers 3 6
6 Shutting down the trains 3 6
Monorail station 7 Ordering an evacuation to the passengers 5 6
8 Shutting down the trains 3 6
Hotel 9 Ordering an evacuation to the guests 3 6
Fire and Disaster
Management Department 10 Rescuing insured passengers 20 30
and damage to about 9000 passengers are calculated using our pedestrian
simulator. The amount of exposure to chloropicrin of the passengers is cal-
culated as the product of the chloropicrin concentration and the time spent
in the area. The influence of chloro-picrin exposure [6, 8] on the passenger’s
behavior is described in the table. 1.
This rail station is a complex facility. It has facilities of four kinds: a con-
ventional line, a new bullet train line, a monorail, and a hotel. Each facility
has a manager. We assume the following 10 coping behaviors of the man-
agers and the times required to begin these coping behaviors described in
the table 2. The times required to begin these coping behaviors influence
the damage because the damage to passengers increases the beginning of
these coping behaviors are delayed. The sequence of the coping behaviors is
portrayed in Fig. 6.
For example, the manager of the conventional line has four coping behav-
iors: (1) detecting the chemical attack, (2) reporting to the fire station and
the other managers of the new bullet train line, monorail, and the hotel,
(3) shutting down conventional trains, and (4) ordering an evacuation of the
passengers of the conventional line. Each coping behavior is set to have the
time of two kinds. For example, if coping behavior 4 is begun quickly, the
Evaluation of Time Delay of Coping Behaviors 411
time necessary to begin is 3 min. Otherwise (begun slowly), the time required
is 6 min.
The evacuation scenarios are 1,024 because the number of combinations of
the times required to begin 10 coping behaviors is 210 (=1,024). We calculate
damage to approximately 9000 passengers in 1024 evacuation scenarios. In
our pedestrian simulation, each passenger walks around normally, from and
to outside areas of the station and from and to platforms, until the attack is
detected and an alarm is given. After ordering an evacuation of the passen-
gers, the passengers evacuate using a route directed by the station staff.
To assign a sequential serial number to each scenario according to whether
10 coping behaviors are begun quickly or slowly, we use a ten-digit number
in the binary system. For example, if coping behavior 1 (detecting chemical
attack) is begun quickly, then the first bit of the ten-digit number is set
as 0. This scenario is represented as 0000000000 in the binary system if all
coping behaviors are begun quickly. Then, the 10-digit number in the binary
system is transferred to a serial number in the decimal system. The scenario
represented as 0000000000 in the bi-nary system is assigned serial number
412 T. Yamashita, S. Soeda, and I. Noda
4 Conclusion
As described in this paper, we explained our evacuation planning assist sys-
tem consisting of a pedestrian simulator, a prediction system of indoor gas
diffusion, and a prediction system of outdoor gas diffusion. A chemical attack
at a major rail station in Japan was taken as an example of the practical use
of our system because of a request from the Fire and Disaster Management
414 T. Yamashita, S. Soeda, and I. Noda
References
1. Burmeister, B., Haddadi, A., Matylis, G.: Application of Multi-Agent Systems
in Traffic and Transportation. Software Engineering 144(1), 51–60 (2002)
2. Nara, M., Kato, S., Huang, H., Zhu, S.W.: Numerical analysis of fire on disasters
with “EVE SAYFA”, a fire simulator, and its validation – LES for thermal plume.
In: Summaries of Technical Papers of 2006 Annual Meeting of Architectural
Institute of Japan. pp. 317–318 (2006) (in Japanese)
3. Ohba, R., Kouchi, A., Hara, T.: Hazard Projection System of Intentional At-
tack in Urban Area. In: 11th Annual George Mason University, Conference on
Atmospheric Transport and Dispersion Modeling (2007) (Poster)
4. Ohba, R., Yamashita, T., Ukai, O., Kato, S.: Development of Hazard Prediction
System for Intentional Attacks in Urban Areas. In: Seventh International Sym-
posium on New Technologies for Urban Safety of Mega Cities in Asia (USMCA
2008), pp. 687–696 (2008)
5. Pan, X.: Computational Modelling of Human and Social Behaviours for Emer-
gency Egress Analysis. Ph.D. Dissertation, Department of Civil and Environ-
mental Engineering, Stanford University (2006)
6. http://en.wikipedia.org/wiki/Chloropicrin
7. http://venus.iis.u-tokyo.ac.jp/english/Research/field-e/
evesayfa/index.html
8. http://www.workcover.nsw.gov.au/Documents/Publications/
AlertsGuidesHazards/DangerousGoodsExplosivesFireworksPyrotechnics/
chloropicrin_fact_sheet_1371.pdf
Web-Based Sensor Network with
Flexible Management by an Agent
System
1 Introduction
To increase agricultural productivity and promote efficient management, it
is important to monitor crop growth, the field environment, and farm opera-
tions at field sites. One of the key technologies in field monitoring is a sensor
network constructed of many sensor nodes made up of small sensor units
with radio data links [2] [10]. A sensor network enables large-scale environ-
mental monitoring with high scalability and robustness, but it is difficult to
perform complicated operations and to treat large amounts of data because
of deficiencies in hardware performance. In the case of agriculture, a flexi-
ble monitoring system is required that collects various types of information,
including image data, changes measurement procedure according to users’
requests, and controls peripheral equipment from a remote site in response
to the situation.
Tokihiro Fukatsu · Masayuki Hirafuji · Takuji Kiura
National Agricultural Research Center, 3-1-1, Kannondai, Tsukuba,
Ibaraki 305-8666, Japan
e-mail: {fukatsu,hirafuji,kiura}@affrc.go.jp
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 415–424.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
416 T. Fukatsu, M. Hirafuji, and T. Kiura
Fig. 2 System architecture of the agent system, consisting of the agent program,
Profiles, and Web interface. The agent program has an HTTP access function, a
data processing function, a data output function, and an agent control function.
on. In a Web-based sensor network, our agent system, which acts like a user
based on an adaptive agent [12], accesses the Web pages of Field Servers,
extracts sensor data from the Web pages, analyzes the data by using Web
application tools, shows the resulting data on a Web page, and stores all data
in a database. All modules of the Web-based sensor network are accessed via
the Internet. The agent system seamlessly connects and uses them to man-
age Field Servers at a remote site, so users need not manage the system at
a field site except when deploying Field Servers and connecting them to the
Internet.
The agent system is constructed by an agent program written in Java,
program configuration files (“Profiles”) describing parameters of request op-
erations in an XML format, and Web interfaces for easy management of the
agent system (Fig. 2). The agent program can access all types of pages on
the Web and operate them flexibly based on the Profiles. It autonomously
executes the contents of Profiles with multi-thread and distributed functions.
Users can utilize the agent program and manage Field Servers to their satis-
faction by describing suitable Profiles.
The agent program is capable of not only basic functions such as accessing
Web pages, extracting available data, and outputting calculation results, but
also advanced rule-based control functions that enable the performance of
complicated operations. By using Web applications that analyze input data,
this program also provides versatile and easily expansible data processing
activities and makes it possible to distribute calculation tasks. In this section,
we show each function of the agent program.
418 T. Fukatsu, M. Hirafuji, and T. Kiura
Fig. 3 Examples of agent actions with the basic function to obtain target image
data according to a required procedure and to extract useful information from raw
data with automatic calculation.
Fig. 4 Examples of agent action with rule-based function. The left figure shows a
surveillance operation with gradual approach based on sensor and image data. The
right figure shows a priority control of running device and access interval based on
battery level.
In the case of priority control operations, the agent program suitably ac-
cesses Field Servers according to their condition and environmental situation.
If the battery level is low, the agent program adjusts the power consumption
of the Field Servers by switching the camera off and lengthening the access
interval. It can also provide appropriate management based on an expected
battery level from a Web weather forecast. A lot of IF-THEN rules act as
autonomous decision functions, so the agent program deals with complicated
and various requests.
420 T. Fukatsu, M. Hirafuji, and T. Kiura
Via the distributed data processing, the Web-based sensor network can
be used for various agricultural applications such as event detection in agri-
cultural fields, automatic counting of insects, and monitoring of farming op-
erations [9] [8]. Various types of data processing can be added simply by
developing or finding the appropriate Web applications, and users’ efforts
can be reduced by sharing Web applications. Moreover, dynamic changing
of data processing according to the situation can be realized by cooperating
with the rule-based function.
Web-Based Sensor Network with Flexible Management 421
3 Multi-Agent System
The agent system has Web interfaces that facilitate the control and moni-
toring of agent information such as checking operation status, creating and
changing Profiles, and searching for and extracting necessary information
from monitoring data. These functions allow the elements of the agent sys-
tem to operate via an http protocol, so the agent system itself can be treated
as just one of the nodes in our system. This also means that the operation of
the agent program can be dynamically changed and modified by accessing the
Web interfaces. By using a number of agent systems with Web interfaces, we
can construct a multi-agent system in which one agent system controls not
only Field Servers but also other agent systems. This structure is designed
to achieve high scalability and a robust system.
Fig. 6 shows two types of multi-agent functions, a backup coordination for
a robust system and task distribution for a scalable system. To establish a
robust system, giving priority to Profiles, in which several agent systems have
charge of several Field Servers redundantly with an access priority, is realized
to provide strongly reliable performance. A subordinate agent system with a
lower-priority Profile does not access the target Field Server, while a superior
agent system with a higher priority works normally. The subordinate agent
system periodically accesses the superior agent system to check its condition,
and it takes over the task only when the superior agent system is down. With
this method, an effective and robust backup function can be realized.
Fig. 6 Standard structure of a multi-agent system with robust and scalable man-
agement. The left figure shows a backup coordination of the system when one agent
system is down. The right figure shows a task distribution of the system to smooth
each task of the agent system.
422 T. Fukatsu, M. Hirafuji, and T. Kiura
References
1. Field Server Web Page, http://model.job.affrc.go.jp/FieldServer/
2. Akyildiz, I.F., Su, W., Sankarasubramaniam, Y., Cayirci, E.: Wireless Sensor
Networks: A Survey. Computer Networks 38(4), 393–422 (2002)
3. Brownston, L., Farrell, R., Kant, E., Martin, N.: Programming Expert Sys-
tems in OPS5: an Introduction to Rule-based Programming. Addison-Wesley
Longman Publishing Co., Inc., Boston (1985)
4. Etzioni, O., Weld, D.: A Softbot-based Interface to the Internet. Commun.
ACM 37(7), 72–76 (1994)
5. Flores-Mendez, R.A.: Towards a Standardization of Multi-agent System Frame-
work. Crossroads 5(4), 18–24 (1999)
6. Fukatsu, T., Hirafuji, M.: Field Monitoring Using Sensor-nodes with a Web
Server. J. Rob. Mechatron. 17(2), 164–172 (2005)
7. Fukatsu, T., Hirafuji, M., Kiura, T.: An Agent System for Operating Web-based
Sensor Nodes via the Internet. J. Rob. Mechatron. 18, 186–194 (2006)
8. Fukatsu, T., Nanseki, T.: Monitoring System for Farming Operations with
Wearable Devices Utilized Sensor Networks. Sensors 9, 6171–6184 (2009)
9. Fukatsu, T., Saito, Y., Suzuki, T., Kobayashi, K., Hirafuji, M.: A Long-term
Field Monitoring System with Field Servers at a Grape Farm. In: Proceedings
of International Symposium on Application of Precision Agriculture for Fruits
and Vegetables, pp. 183–190 (2008)
424 T. Fukatsu, M. Hirafuji, and T. Kiura
10. Kahn, J.M., Katz, R.H., Pister, K.S.J.: Next Century Challenges: Mobile Net-
working for “Smart Dust”. In: Proceedings of the 5th Annual ACM/IEEE
International Conference on Mobile Computing and Networking, pp. 271–278
(1999)
11. Lange, D., Oshima, M.: Programming and Deploying Java Mobile Agents
Aglets. Addison-Wesley Longman Publishing, Massachusetts (1998)
12. Lesser, V., Oritz, C., Tambe, M. (eds.): Distributed Sensor Networks: a Multi-
agent Perspective. Kluwer Academic Publishers, Dordrecht (2003)
13. Maes, P.: Agents that Reduce Work and Information Overload. Commun.
ACM 37(7), 30–40 (1994)
14. Russel, S., Norvig, P.: Artificial Intelligence: A Modern Approach. Prentice-
Hall, New Jersey (1995)
Sensor Network Architecture Based on
Web and Agent for Long-Term
Sustainable Observation in Open
Fields
1 Introduction
Long term observation data in open fields is very valuable for environmental
science, agriculture and so on. Observation in the fields should be continued
as long as possible, and data should be archived permanently. However life-
time of digital devices and data formats which depend on applications is not
long, since digital technologies are rapidly advancing. For example, digital in-
terfaces such as RS-232 had been popular, but currently RS-232C is treated
as “legacy devices” and we are using USB instead. Data recorded, 30 years
Masayuki Hirafuji · Tokihiro Fukatsu · Takuji Kiura · Haoming Hu · Hideo Yoichi ·
Kei Tanaka · Yugo Miki · Seishi Ninomiya
National Agricultural Research Center, Tsukuba, 305-8666, Japan
e-mail: {hirafuji,fukatsu,kiura,haominhu,hyoichi}@affrc.go.jp,
{tanaka.kei,mikiyugo,snino}@affrc.go.jp
Masayuki Hirafuji · Kei Tanaka · Seishi Ninomiya
University of Tsukuba, Tsukuba, 305-8666, Japan
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 425–433.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
426 M. Hirafuji et al.
ago, on 8-inch floppy diskettes were unreadable by recent PCs. Moreover data
files for legacy software such as VISICALC (the first spread sheet program)
also became to legacy.
Sustainable sensor network architecture as a kind of sustainable informa-
tion systems is indispensable to monitor natural phenomena. Here we propose
a sensor network architecture which can be maintained sustainably for long
term.
2 Architecture
Objectives of data collection in agronomy, ecology, meteorology, environmen-
tal science and so on are to analyze complex phenomena of nature based on
accumulated data for long term. A sustainable architecture for data format,
protocol, hardware and software were chosen as followings:
ordinal programmers can develop easily applications for the Web-based sensor
networks, and this advantage can sustain the Web-based sensor networks.
On 2001, the smallest Web server was a one-chip CPU (PIC microproces-
sor), which worked actually as ordinal Web servers (Fig. 1) [4]. This demon-
stration implied that all devices and appliances can be Web devices in low-cost,
and Web devices will be more functional using advanced technologies such as
more sensors and reconfigurable devices. We concluded, therefore, sensor net-
works should be constituted logically only by Web servers and we started a
project to develop desirable functions and devices to realize the web-based sen-
sor networks based on the proposed architecture (i.e. pure web).
The Web devices for Web-based sensor networks should be composed by:
428 M. Hirafuji et al.
(a) Prototype (b) 1st type (c) 2nd type (d) 3rd type
(2001) (2001) (2003) (2008)
were deployed at many sites in the world for testing of long-term operation
and experiments of durability (Fig. 5) [3]. Recent deployment was Himalaya.
Imja glacial lake and Namche bazaar have been monitored by 5 Field Servers
totally (Fig. 6), which equip cameras (Fig. 7) and sensors (air temperature,
relative humidity, solar radiation, CO2 concentration and water level).
and later two Field Servers under coconut trees were crushed by a falling
coconut. The Internet connections to Field Servers were frequently discon-
nected by same reasons, such as human errors, freezing gateway routers and
power failure, as in offices.
We could quickly find out the causes and improved the systems, since we
could obtain rich information on webs of the devices respectively. Sometime
we could restore the troubled sensor networks remotely from Japan only by
accessing webs of the devices. This manner is similar to remote maintenance
for space probes such as Mars surveyors.
Sensor networks which Field Servers constitute are “pull-type” sensor net-
work; contrarily conventional sensor networks such as smart dust [5], Mote2
and devices using ZigBee are “push-type”. “pull-type” is advantageous to:
• Stability
If enormous push-type sensor nodes send data onto a data storage server
all simultaneously, their accesses may damage the network and the storage
server like as DDoS (Distributed Denial of Service). On the contrary, pull-
type sensor nodes (Field Servers) wait for access of the agent. The agent
crawls over sensor nodes to collect data according to capacity of individual
network connection.
• Scalability
Even if too many pull-type sensor nodes are deployed for an agent, the
agent can crawl on them in series up to its maximum capability. Or only by
increasing number of PCs, capability of multiple agents can be reinforced.
• Qualia of connectivity
Qualia is subjective quality of conscious experience [1]. Users can access
pull-type sensor nodes manually using Web browser. Then the users can
sense capacity of connection to a sensor node, and delay of its response as
vivid feeling. Qualia in this sensation are very useful to estimate conditions
such as network capacity to deploy sensor nodes in unknown fields.
These points were useful to deploy sensor networks in short time completely.
Especially in extreme environment such as Himalaya, mission time for deploy-
ment in high altitude was limited for thin air and cold windy environment.
We could check devices and connectivity at once by feeling response from sen-
sor nodes. Later we could set up an agent to collect data from the deployed
Field Servers and optimize data sampling rate in Japan at ease. Moreover we
improved ventilation algorithm on the agent to protect inside devices against
low temperature and condensation.
Qualia of connectivity were indispensible for diagnosis, trouble shooting
and decision making in explorative deployments, where we must pay atten-
tion to enormous potential troubles. For example, only to choose the best an-
tenna we must consider many related factors such as communication range,
direction, antenna gain, tolerance against strong wind, depth of the snow and
2
http://www.xbow.com
432 M. Hirafuji et al.
so on. Only a failure of these factors causes serious troubles. In a case that a
device in a Field Server was disconnected at a site beside Imja Glacial Lake
in Himalaya, we could estimate causes based on Qualia of connectivity to
other live devices, and we repaired it there only by carrying tools prepared
in advance; kinds of carrying tools and parts are limited.
Currently we are trying to deploy Field Servers at an experimental field in
India, where air temperature is extremely high and almost other conditions
are unknown for us. Only connection using cellular phone (GPRS) is available
there, so agent box (agent installed in a Linux box) was deployed at the field
instead. Currently we have a lot anxiety about operation of the Field Servers,
since we cannot get Qualia of their operation in Japan.
All archived data collected by Field Servers are visualized by an applet
viewer as shown in Fig. 8. However this service and the agent service also are
not sustainable, since these services are only for scientific researches; these
services must terminate after all projects for developing Field Servers are
completed. Ideally such services should be operated in free as semi-permanent
business like as Google and Yahoo! to realize sustainable sensor network.
The agent service may create a new network business; the agents can collect
other numerical data provided on webs in the Internet by same manner as
agents crawl on Field Servers. Users can find hidden relationships between
time series data such as the price index of stocks and temperature in districts,
which are collected by the agent service.
6 Conclusions
We proposed sustainable sensor network architecture based on philosophy
of pure web: all sensor nodes and inside devices are web servers, and agents
Sensor Network Architecture Based on Web and Agent 433
crawl the web servers to collect data. We deployed sensor networks at many
sites for 8 years using Field Servers to examine that the proposed architecture
actually works well for long-term even in extreme environments. The pure web
architecture could work much better than we expected. The oldest prototype
Field Server is still on active service and accessible by the newest browser
such as Firefox and Safari. It will be kept on active services for long years
as hardware; however network services such as the agent service must be
sustainable as network business.
References
1. Chalmers, D.J.: The Conscious Mind: In Search of a Fundamental Theory (Phi-
losophy of Mind Series). Oxford Univ. Press, Oxford (1997)
2. Hirafuji, M.: Arbitrary Function Generating Circuit Using a Simple Operational
Element and an Encryption Method Thereof. US Patent: US6,898,616 B2 (2005)
3. Hirafuji, M., Ninomiya, S., Kiura, T., Fukatsu, T., Hu, H., Yuoichi, H., Tanaka,
K., Sugahara, K., Watanabe, T., Kameoka, T., Hashimoto, A., Ito, R., Ehsani,
R., Shimamura, H.: Field Server Projects. In: Proceedings of Practical Applica-
tions of Sensor Networking (SAINT 2007), pp. 15–19 (2007)
4. Ju, H.T., Choi, M.J., Hong, J.W.: An Efficient and Lightweight Embedded Web
Server for Web-based Network Element Management. International Journal of
Network Management 10(5), 261–275 (2000)
5. Kahn, J.M., Katz, R.H., Pister, K.S.J.: Next century challenges: Mobile net-
working for “smart dust”. In: Proceedings of the 5th Annual ACM/IEEE Inter-
national Conference on Mobile Computing and Networking, pp. 271–278. ACM,
New York (1999)
A Multi-Agent View of the Sensor
Web
1 Introduction
A large amount of times-series spatio-temporal data is collected from the
physical environment due to increase in the deployment of sensors. The de-
ployment ranges from satellite-based sensors [24] that monitor large scale
environmental and ecological systems, to smart meters [5] that monitor en-
ergy usages.
Meanwhile, there is a need for an open platform to allow users from dif-
ferent geographical location to share data for macro monitoring. The Open
Geospatial Consortium (OGC), an international standards organisation, has
a Sensor Web working group developing Sensor Web Enablement (SWE). The
goal of OGC-SWE is to develop interoperable data models and web service
Quan Bai · Siddeswara Mayura Guru · Daniel Smith · Qing Liu · Andrew Terhorst
Tasmanian ICT Centre, CSIRO, GPO Box 1538, Hobart, TAS, 7001, Australia
e-mail: {quan.bai,siddeswara.guru,daniel.v.smith}@csiro.au,
{q.liu,andrew.terhorst}@csiro.au
Q. Bai and N. Fukuta (Eds.): Advances in Practical Multi-Agent Systems, SCI 325, pp. 435–444.
springerlink.com
c Springer-Verlag Berlin Heidelberg 2010
436 Q. Bai et al.
SOSs and to get the offeringid (user has to know the offeringid to run any
GetObservation operations), then the GetObservation operation is invoked
to get the observation from the sensors.
These steps can be intimidating and complex to follow for a naive user.
It may be possible that some SOSs that fulfil the query could be missed.
Although there has been some work to develop cascading SOS to aggregate
or fuse the SOSs [9], the SOSs still need to be identified by the CSW. There
is a need for automation in the query execution. Intelligent coordinators
(agents) can interact with the services to pass messages, process requests,
submit requests and coordinate activities between the services. This infuses
intelligence into the framework and enables the introduction of application-
specific services. In the next section, we will outline some of the similarities
between MAS and the Sensor Web.
providers and service consumers are uni-directional. This creates a large com-
munication gap between the service consumers and providers. To overcome
these limitations, intelligent brokers are required to facilitate interactions and
enable bi-directional communications between the CSW, service consumers
and providers. With bi-directional communications, service consumers can
not only seek the required services from the CSW, but service providers can
have access to the consumer requirements.
There is a set of intelligent facilitators for supporting multi-agent inter-
actions. Two kinds of these multi-agent facilitators, the Yellow Pages Agent
and Blackboard Agent, are considered for use in the Sensor Web (see Fig. 3).
Yellow Pages Agent (YPA) is a multi-agent facilitator that processes
the advertisements of service providers and assists service consumers to find
suitable providers based upon their advertised capabilities [22].
In the Sensor Web, service providers could publish and update their meta-
data descriptions via the YPA to the CSW. The YPA classifies service
providers based upon their capabilities, identifying the offeringid before up-
dating the CSW. For a user executing the example hydrology query, the CSW
could provide an getObservation URI of SOSs directly, saving a couple of op-
erations (the getRecord of the CSW and getCapabilities of the SOS). This
could help to improve the automation of the query execution process and
reduce user involvement.
In some situations, the capabilities of a service provider cannot be fully
published to all the users due to security reasons, conflict goals, etc. In this
situation, we need to have a brokering service between the consumers and
providers to restrict the access of the consumers.
Blackboard Agent (BBA) is a multi-agent facilitator that can collect,
hold and publish requests from service consumers [3]. Service providers can
seek and select tasks that they can handle from the BBA, and interact with
the corresponding service consumers actively. The inclusion of the BBA en-
ables bi-directional interactions between service consumers and providers.
This allows service providers to protect their information.
References
1. Botts, M., Percivall, G., Reed, C., Davidson, J.: OGC Sensor Web Enablement:
Overview and High Level Architecture. In: Nittel, S., Labrinidis, A., Stefanidis,
A. (eds.) GSN 2006. LNCS, vol. 4540, pp. 175–190. Springer, Heidelberg (2008)
2. Catalog Services for the Web, http://www.opengeospatial.org/
standards/cat (last visited at September 2009)
3. Decker, K., Sycara, K., Williamson, M.: Middle-Agents for the Internet. In:
Proceedings of the Fifteenth International Joint Conference on Artificial Intel-
ligence, Nagoya, Japan, pp. 578–584 (1997)
4. Delin, K.A.: The Sensor Web: A Macro-Instrument for Coordinating Sensing.
Sensors 2, 270–285 (2002)
5. Derbel, F.: Trends in Smart Metering. In: Proceedings of the 6th International
Multi-Conference on Systems, Signals and Devices, pp. 1–4 (2009)
6. Gross, N.: 21 Ideas for the 21st Century: 14 Internet: The Earth will
don an Electronic Skin, http://www.businessweek.com/1999/99_35/
b3644024.htm (last visited at September 2009)
7. Guru, S.M., Peters, C., Smith, D., Pratt, A., Bai, Q., Terhorst, A.: Sensor
web and scientific workflow technology to manage and observe hydrological
phenomenon. In: Abstrtact accepted for 9th International Conference on Hy-
droInformatics (2010)
444 Q. Bai et al.
8. Guru, S.M., Taylor, P., Neuhaus, H., Shu, Y., Smith, D., Terhorst, A.: Hy-
drological sensor web for the south esk catchment in the tasmanian state of
australia. In: Fourth IEEE International Conference on eScience [poster], pp.
432–433 (2008)
9. Havlik, D., Bleier, T., Schimak, G.: Sharing sensor data with sensorsa and
cascading sensor observation service. Sensors 9(7), 5493–5502 (2009)
10. Jirka, S., Broring, A., Stasch, C.: Discovery mechanisms for the sensor web.
Sensor, 2661–2681 (2009)
11. Moodley, D., Simonis, I.: New Architecture for the Sensor Web: the SWAP
Framework. In: Cruz, I., Decker, S., Allemang, D., Preist, C., Schwabe, D.,
Mika, P., Uschold, M., Aroyo, L.M. (eds.) ISWC 2006. LNCS, vol. 4273,
Springer, Heidelberg (2006),
http://www.ict.csiro.au/ssn06/data/swap.pdf
12. Mui, L., Mohtashemi, M., Halberstadt, A.: Notions of Reputation in Multi-
Agents Systems: a Review. In: Proceedings of the First International Joint Con-
ference on Autonomous Agents and Multiagent Systems, pp. 280–287 (2002)
13. Open Geospatial Consortium: Observation Measurement Specification, docu-
ment 07-022r1, http://www.opengeospatial.org/standards/om (last
visited at September 2009)
14. Open Geospatial Consortium: Web Notification Service, OpenGIS Project Doc-
ument: OGC 03-008r2 Version 0.1.0., http://portal.opengeospatial.
org/files/?artifact_id=1367 (2003), (last visited at September 2009)
15. Open Geospatial Consortium: OGC Sensor Web Enablement: Overview and
High Level Architecture. Tech. Rep. OGC 07-165 (2007)
16. OpenGIS Sensor Model Language (SensorML),
http://www.opengeospatial.org/standards/sensorml (last visited
at September 2009)
17. Request for Comment on Sensor Alert Service,
http://www.opengeospatial.org/standards/requests/44 (last vis-
ited at September 2009)
18. Sabater, J., Sierra, C.: Reputation and Social Network Analysis in Multi-Agent
Systems. In: Proceedings of the First International Joint Conference on Au-
tonomous Agents and Multiagent Systems, pp. 475–482 (2002)
19. Sensor Observation Service Specification, http://www.opengeospatial.
org/standards/sos (last visited at September 2009)
20. Sensor Planning Service Specification, http://www.opengeospatial.org/
standards/sps (last visited at September 2009)
21. Subrahmanian, V., Bonatti, P., Dix, J., Eiter, T., Kraus, S., Ozcan, F., Ross,
R.: Heterogeneous Agent Systems. The MIT Press, Cambridge (2000)
22. Sycara, K., Klusch, M., Widoff, S., Lu, J.: Dynamic Service Matchmaking
Among Agents in Open Information Environments. SIGMOD Record 28(1),
47–53 (1999)
23. SouthEsk Hydrological Sensor Web, http://152.83.132.197/au.csiro.
OgcThinClient/site/OgcThinClient.html (last visited at September
2009)
24. Ye, W., Silva, F., DeSchon, A., Bhatt, S.: Architecture of a Satellite-based Sen-
sor Network for Environment Observation. In: Proceedings of NASA Earth Sci-
ence Technology Conference (ESTC 2008) (2008) http://www.esto.nasa.
gov/conferences/estc2008/papers/Ye_Wei_A8P2.pdf
Author Index