Learning Probabilistic Logical Control Networks: From Data to Controllability and Observability

被引:2
作者
Lin, Lin [1 ]
Lam, James [1 ]
Shi, Peng [2 ,3 ]
Ng, Michael K. [4 ]
Lam, Hak-Keung [5 ]
机构
[1] Univ Hong Kong, Dept Mech Engn, Pokfulam, Hong Kong, Peoples R China
[2] Univ Adelaide, Sch Elect & Mech Engn, Adelaide, SA 5005, Australia
[3] Obuda Univ, Res & Innovat Ctr, H-1034 Budapest, Hungary
[4] Hong Kong Baptist Univ, Dept Math, Kowloon Tong, Hong Kong, Peoples R China
[5] Kings Coll London, Dept Engn, London WC2R 2LS, England
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Observability; Controllability; Probabilistic logic; Mathematical models; Matrix converters; Dynamic programming; Biological system modeling; Q-learning; Proteins; Optimal control; observability; probabilistic logical networks (PLNs); reinforcement learning (RL); semitensor product (STP); BOOLEAN CONTROL NETWORKS; STABILIZATION; AUTOMATA; SYSTEMS; MODELS;
D O I
10.1109/TAC.2024.3524241
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article studies controllability and observability problems for a class of mixed-valued probabilistic logical control networks (PLCNs). First, PLCN is transformed into the algebraic state-space representation (ASSR)-form by resorting to the semitensor product method. Then, the formulas are presented to calculate the lower and upper bounds of the transition probability matrix, which further derive the controllability and observability criteria. Furthermore, the ASSR-form of a PLCN can be regarded as a Markov decision process. Using the latter framework, we prove the equivalence between the controllability probability and the optimal state-value function, which is an iteration equation. Besides, the parallel extension technique transforms the observability of PLCNs into the set stabilization of an augmented system. The correspondence between observability probability and optimal state-value function is also derived. Afterward, based on the state-value function, the algorithms via the Q-learning technique are exploited to estimate the controllability and observability probabilities along with obtaining the corresponding optimal control sequences. Finally, all the theoretical results are elaborated via a genetic regulatory p53-Mdm2 network.
引用
收藏
页码:3889 / 3904
页数:16
相关论文
共 45 条
[1]   From structure to dynamics: Frequency tuning in the p53-Mdm2 network I. Logical approach [J].
Abou-Jaoude, Wassim ;
Ouattara, Djomangan A. ;
Kaufman, Marcelle .
JOURNAL OF THEORETICAL BIOLOGY, 2009, 258 (04) :561-577
[2]   Reinforcement Learning Approach to Feedback Stabilization Problem of Probabilitic Boolcan Control Networks [J].
Acernese, Antonio ;
Yerudkar, Amol ;
Glielmo, Luigi ;
Del Vecchio, Carmen Del .
IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (01) :337-342
[3]   Control of Boolean networks: Hardness results and algorithms for tree structured networks [J].
Akutsu, Tatsuya ;
Hayashida, Morihiro ;
Ching, Wai-Ki ;
Ng, Michael K. .
JOURNAL OF THEORETICAL BIOLOGY, 2007, 244 (04) :670-679
[4]   Self-triggered control of probabilistic Boolean control networks: A reinforcement learning approach [J].
Bajaria, Pratik ;
Yerudkar, Amol ;
Glielmo, Luigi ;
Del Vecchio, Carmen ;
Wu, Yuhu .
JOURNAL OF THE FRANKLIN INSTITUTE, 2022, 359 (12) :6173-6195
[5]   A MARKOVIAN DECISION PROCESS [J].
BELLMAN, R .
JOURNAL OF MATHEMATICS AND MECHANICS, 1957, 6 (05) :679-684
[6]   Boolean network models of cellular regulation: prospects and limitations [J].
Bornholdt, Stefan .
JOURNAL OF THE ROYAL SOCIETY INTERFACE, 2008, 5 (SUPPL. 1) :S85-S94
[7]  
Chaves Madalena, 2009, 2009 European Control Conference (ECC), P671
[8]  
Chen Liming, 2008, International Journal of Assistive Robotics and Systems, V9, P20
[9]   Analysis and control of general logical networks - An algebraic approach [J].
Cheng, Daizhan ;
Qi, Hongsheng ;
Zhao, Yin .
ANNUAL REVIEWS IN CONTROL, 2012, 36 (01) :11-25
[10]   Strategy Optimization with Its Application to Dynamic Games [J].
Cheng, Daizhan ;
Zhao, Yin ;
Mu, Yifen .
49TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2010, :5822-5827