Efficient evaluation functions for evolving coordination

被引:45
作者
Agogino, A. [1 ]
Tumer, K. [2 ]
机构
[1] NASA, UCSC, Ames Res Ctr, Moffett Field, CA 94035 USA
[2] Oregon State Univ, Corvallis, OR 97331 USA
关键词
evolution strategies; distributed control; fitness evaluation;
D O I
10.1162/evco.2008.16.2.257
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents fitness evaluation functions that efficiently evolve coordination in large multi-component systems. In particular, we focus on evolving distributed control policies that are applicable to dynamic and stochastic environments. While it is appealing to evolve such policies directly for an entire system, the search space is prohibitively large in most cases to allow such an approach to provide satisfactory results. Instead, we present an approach based on evolving system components individually where each component aims to maximize its own fitness function. Though this approach sidesteps the exploding state space concern, it introduces two new issues: (1) how to create component evaluation functions that are aligned with the global evaluation function; and (2) how to create component evaluation functions that are sensitive to the fitness changes of that component, while remaining relatively insensitive to the fitness changes of other components in the system. If the first issue is not addressed, the resulting system becomes uncoordinated; if the second issue is not addressed, the evolutionary process becomes either slow to converge or worse, incapable of converging to good solutions. This paper shows how to construct evaluation functions that promote coordination by satisfying these two properties. We apply these evaluation functions to the distributed control problem of coordinating multiple rovers to maximize aggregate information collected. We focus on environments that are highly dynamic (changing points of interest), noisy (sensor and actuator faults), and communication limited (both for observation of other rovers and points of interest) forcing the rovers to evolve generalized solutions. On this difficult coordination problem, the control policy evolved using aligned and component-sensitive evaluation functions outperforms global evaluation functions by up to 400%. More notably, the performance improvements increase when the problems become more difficult (larger, noisier, less communication). In addition we provide an analysis of the results by quantifying the two characteristics (alignment and sensitivity discussed above) leading to a systematic study of the presented fitness functions.
引用
收藏
页码:257 / 288
页数:32
相关论文
共 38 条
  • [1] A genetic algorithm-based controller for decentralized multi-agent robotic systems
    Agah, A
    Bekey, GA
    [J]. 1996 IEEE INTERNATIONAL CONFERENCE ON EVOLUTIONARY COMPUTATION (ICEC '96), PROCEEDINGS OF, 1996, : 431 - 436
  • [2] Agogino A, 2004, LECT NOTES COMPUT SC, V3102, P1
  • [3] Online interactive neuro-evolution
    Agogino, A
    Stanley, K
    Miikkulainen, R
    [J]. NEURAL PROCESSING LETTERS, 2000, 11 (01) : 29 - 37
  • [4] Agogino A.K., 2005, P 4 INT JOINT C AUT, P81
  • [5] Agogino A. K., 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems-Volume, P980
  • [6] Agogino A, 2006, GECCO 2006: GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, VOL 1 AND 2, P1079
  • [7] Agogino A, 2005, GECCO 2005: GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, VOLS 1 AND 2, P1309
  • [8] [Anonymous], IEEE INTELLIGENT SYS
  • [9] Dorigo M, 2004, ANT COLONY OPTIMIZATION, P1
  • [10] FARRITOR S, 2002, ASME, V124, P698