Abstract
The area of agent modeling deals with the task of observing other agents and modeling their behavior, in order to predict their future behavior, coordinate with them, assist them, or counter their actions. Typically, agent modeling techniques assume the availability of a plan- or behavior-library, which encodes the full repertoire of expected observed behavior. However, recent applications areas of agent modeling raise challenges to the assumption of such a library, as agent modeling systems are increasingly used in open and/or adversarial settings, where the behavioral repertoire of the observed agents is unknown at design time. This paper focuses on the challenge of the unsupervised autonomous learning of the sequential behaviors of agents, from observations of their behavior. The techniques we present translate observations of the dynamic, complex, continuous multi-variate world state into a time-series of recognized atomic behaviors. This time-series is then analyzed to find repeating subsequences characterizing each team. We compare two alternative approaches to extracting such characteristic sequences, based on frequency counts and statistical dependencies. Our results indicate that both techniques are able to extract meaningful sequences, and do significantly better than random predictions. However, the statistical dependency approach is able to correctly reject sequences that are frequent, but are due to random co-occurrence of behaviors, rather than to a true sequential dependency between them.
Chapter PDF
Similar content being viewed by others
Keywords
- Statistical Dependency
- Frequency Count
- Agent Modeling
- Atomic Behavior
- Defense Advance Research Project Agency
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Lesh, N., Rich, C., Sidner, C.L.: Using plan recognition in human-computer collaboration. In: Proceedings of the International Conference on User Modelling (UM 1999), Banff, Canada (1999)
Davidson, B., Hirsh, H.: Probabilistic online action prediction. In: Proceedings of the 1998 AAAI Spring Symposium on Intelligent Environments (1998)
Rickel, J., Johnson, W.L.: Animated agents for procedural training in virtual reality: Perception, cognition, and motor control. Applied Artificial Intelligence 13, 343–382 (1999)
Tambe, M., Rosenbloom, P.S.: RESC: An approach to agent tracking in a realtime, dynamic environment. In: Proceedings of the International Joint Conference on Artificial Intelligence (1995)
Riley, P., Veloso, M., Kaminka, G.: An empirical study of coaching. In: Proceedings of Distributed Autonomous Robotic Systems, vol. 6. Springer, Heidelberg (2002) (to appear)
Bauer, M.: From interaction data to plan libraries: A clustering approach. In: Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, vol. 2, pp. 962–967. Morgan-Kaufman Publishers, Inc., San Francisco (1999)
Knuth, D.E.: Sorting and Searching. The art of computer programming, vol. 3. Addison-Wesley, Reading (1973)
Howe, A.E., Cohen, P.R.: Understanding planner behavior. Artificial Intelligence 76, 125–166 (1995)
Zaki, M., Lesh, N., Ogihara, M.: Planmine: Sequence mining for plan failures. In: Proceedings of the 4th International Conference on Knowledge Discovery and Data Mining, pp. 369–374. AAAI Press, Menlo Park (1998)
Howe, A.E., Somlo, G.L.: Modeling intelligent system execution as state-transition diagrams to support debugging. In: Proceedings of the Second International Workshop on Automated Debugging (1997)
Wendler, J., Kaminka, G.A., Veloso, M.: Automatically improving team cooperation by applying coordination models. In: The AAAI Fall symposium on Intent Inference for Collaborative Tasks. AAAI Press, Menlo Park (2001)
Han, K., Veloso, M.: Automated robot behavior recognition applied to robotic soccer. In: Proceedings of the IJCAI 1999 Workshop on Team Behavior and Plan-Recognition (1999); Also appears in Proceedings of the 9th International Symposium of Robotics Research (ISSR 1999)
Ando, T.: Refinement of soccer agents’ positions using reinforcement learning. In: Kitano, H. (ed.) RoboCup-97: Robot soccer world cup I. LNCS (LNAI), vol. 1395, pp. 373–388. Springer, Heidelberg (1998)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kaminka, G.A., Fidanboylu, M., Chang, A., Veloso, M.M. (2003). Learning the Sequential Coordinated Behavior of Teams from Observations. In: Kaminka, G.A., Lima, P.U., Rojas, R. (eds) RoboCup 2002: Robot Soccer World Cup VI. RoboCup 2002. Lecture Notes in Computer Science(), vol 2752. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-45135-8_9
Download citation
DOI: https://doi.org/10.1007/978-3-540-45135-8_9
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40666-2
Online ISBN: 978-3-540-45135-8
eBook Packages: Springer Book Archive