Variational Hierarchical Mixtures for Probabilistic Learning of Inverse Dynamics

被引:0
作者
Abdulsamad, Hany [1 ]
Nickl, Peter [2 ]
Klink, Pascal [3 ]
Peters, Jan [3 ]
机构
[1] Aalto Univ, Dept Elect Engn & Automat, Espoo 02150, Finland
[2] RIKEN Ctr Adv Intelligence Project, Chuo City 1030027, Japan
[3] Tech Univ Darmstadt, Dept Comp Sci, D-64289 Darmstadt, Germany
基金
欧盟地平线“2020”;
关键词
Bayes methods; Data models; Computational modeling; Uncertainty; Mixture models; Manipulator dynamics; Neural networks; Dirichlet process mixtures; generative models; hierarchical local regression; inverse dynamics control; SAMPLING METHODS; INFERENCE; NETWORKS; MODELS;
D O I
10.1109/TPAMI.2023.3314670
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Well-calibrated probabilistic regression models are a crucial learning component in robotics applications as datasets grow rapidly and tasks become more complex. Unfortunately, classical regression models are usually either probabilistic kernel machines with a flexible structure that does not scale gracefully with data or deterministic and vastly scalable automata, albeit with a restrictive parametric form and poor regularization. In this paper, we consider a probabilistic hierarchical modeling paradigm that combines the benefits of both worlds to deliver computationally efficient representations with inherent complexity regularization. The presented approaches are probabilistic interpretations of local regression techniques that approximate nonlinear functions through a set of local linear or polynomial units. Importantly, we rely on principles from Bayesian nonparametrics to formulate flexible models that adapt their complexity to the data and can potentially encompass an infinite number of components. We derive two efficient variational inference techniques to learn these representations and highlight the advantages of hierarchical infinite local regression models, such as dealing with non-smooth functions, mitigating catastrophic forgetting, and enabling parameter sharing and fast predictions. Finally, we validate this approach on large inverse dynamics datasets and test the learned models in real-world control scenarios.
引用
收藏
页码:1950 / 1963
页数:14
相关论文
共 50 条
[31]   A Hierarchical Incentive Mechanism for Federated Learning [J].
Huang, Jiwei ;
Ma, Bowen ;
Wu, Yuan ;
Chen, Ying ;
Shen, Xuemin .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (12) :12731-12747
[32]   Entropy-Based Incremental Variational Bayes Learning of Gaussian Mixtures [J].
Penalver, Antonio ;
Escolano, Francisco .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2012, 23 (03) :534-540
[33]   Brain networks for confidence weighting and hierarchical inference during probabilistic learning [J].
Meyniel, Florent ;
Dehaene, Stanislas .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2017, 114 (19) :E3859-E3868
[34]   Learning the Inverse Dynamics of Robotic Manipulators in Structured Reproducing Kernel Hilbert Space [J].
Cheng, Ching-An ;
Huang, Han-Pang ;
Hsu, Huan-Kun ;
Lai, Wei-Zh ;
Cheng, Chih-Chun .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (07) :1691-1703
[35]   Mixtures of Variational Autoencoders [J].
Ye, Fei ;
Bors, Adrian G. .
2020 TENTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), 2020,
[36]   Learning Nonlinear Mixtures: Identifiability and Algorithm [J].
Yang, Bo ;
Fu, Xiao ;
Sidiropoulos, Nicholas D. ;
Huang, Kejun .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 :2857-2869
[37]   Variational Approximation for Mixtures of Linear Mixed Models [J].
Tan, Siew Li ;
Nott, David J. .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2014, 23 (02) :564-585
[38]   VASE: Variational Assorted Surprise Exploration for Reinforcement Learning [J].
Xu, Haitao ;
Szymanski, Lech ;
McCane, Brendan .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (03) :1243-1252
[39]   A low-cost variational-Bayes technique for merging mixtures of probabilistic principal component analyzers [J].
Bruneau, Pierrick ;
Gelgon, Marc ;
Picarougne, Fabien .
INFORMATION FUSION, 2013, 14 (03) :268-280
[40]   Probabilistic Deep Learning Based on Bayes by Backprop for Remaining Useful Life Prognostics of Consumer Electronics [J].
Wang, Guochao ;
Wang, Yu ;
Li, Baotong ;
Zhang, Bin .
IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2025, 71 (01) :839-848