Distributed Learning for Planning Under Uncertainty Problems with Heterogeneous Teams

被引:0
作者
Ure, N. Kemal [1 ]
Chowdhary, Girish [2 ]
Chen, Yu Fan [1 ]
How, Jonathan P. [1 ]
Vian, John [3 ]
机构
[1] MIT, Cambridge, MA 02139 USA
[2] Oklahoma State Univ, Stillwater, OK 74074 USA
[3] Boeing Res & Technol, Seattle, WA USA
关键词
Distributed learning; Planning under uncertainty; Unmanned aerial systems; MULTIAGENT; SYSTEMS;
D O I
10.1007/s10846-013-9980-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper considers the problem of multiagent sequential decision making under uncertainty and incomplete knowledge of the state transition model. A distributed learning framework, where each agent learns an individual model and shares the results with the team, is proposed. The challenges associated with this approach include choosing the model representation for each agent and how to effectively share these representations under limited communication. A decentralized extension of the model learning scheme based on the Incremental Feature Dependency Discovery (Dec-iFDD) is presented to address the distributed learning problem. The representation selection problem is solved by leveraging iFDD's property of adjusting the model complexity based on the observed data. The model sharing problem is addressed by having each agent rank the features of their representation based on the model reduction error and broadcast the most relevant features to their teammates. The algorithm is tested on the multiagent block building and the persistent search and track missions. The results show that the proposed distributed learning scheme is particularly useful in heterogeneous learning setting, where each agent learns significantly different models. We show through large-scale planning under uncertainty simulations and flight experiments with state-dependent actuator and fuel-burn-rate uncertainty that our planning approach can outperform planners that do not account for heterogeneity between agents.
引用
收藏
页码:529 / 544
页数:16
相关论文
共 32 条
  • [1] [Anonymous], EUR C MACH LEARN ECM
  • [2] Bertsekas D. P., 1995, Dynamic programming and optimal control, V1
  • [3] Bethke B., 2008, AIAA GUID NAV CONTR
  • [4] Busoniu L, 2010, AUTOM CONTROL ENG SE, P1, DOI 10.1201/9781439821091-f
  • [5] A comprehensive survey of multiagent reinforcement learning
    Busoniu, Lucian
    Babuska, Robert
    De Schutter, Bart
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02): : 156 - 172
  • [6] Consensus-Based Decentralized Auctions for Robust Task Allocation
    Choi, Han-Lim
    Brunet, Luc
    How, Jonathan P.
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2009, 25 (04) : 912 - 926
  • [7] Distributed Bayesian Learning in Multiagent Systems Improving our understanding of its capabilities and limitations
    Djuric, Petar M.
    Wang, Yunlong
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2012, 29 (02) : 65 - 76
  • [8] Geramifard A., 2011, P INT C MACHINE LEAR, P881
  • [9] Real-time indoor Autonomous Vehicle test Environment
    How, Jonathan P.
    Bethke, Brett
    Frank, Adrian
    Dale, Daniel
    Vian, John
    [J]. IEEE CONTROL SYSTEMS MAGAZINE, 2008, 28 (02): : 51 - 64
  • [10] Quickest Time Detection and Constrained Optimal Social Learning with Variance Penalty
    Krishnamurthy, Vikram
    [J]. 49TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2010, : 1102 - 1107