TY - JOUR
T1 - Learning Probabilistic Logical Control Networks
T2 - From Data to Controllability and Observability
AU - Lin, Lin
AU - Lam, James
AU - Shi, Peng
AU - Ng, Michael K.
AU - Lam, Hak Keung
N1 - Funding Information:
The work was partially supported by the National Natural Science Foundation of China under Grant 62273286, GRF 17200124, and the Australian Research Council (DP240101140).
Publisher Copyright:
© 2024 IEEE.
PY - 2024/12/30
Y1 - 2024/12/30
N2 - This paper studies controllability and observability problems for a class of mixed-valued probabilistic logical control networks (PLCNs). First, PLCN is transformed into the algebraic state space representation (ASSR)-form by resorting to the semi-tensor product method. Then, the formulas are presented to calculate the lower and upper bounds of the transition probability matrix, which further derive the controllability and observability criteria. Furthermore, the ASSR-form of a PLCN can be regarded as a Markov decision process. Using the latter framework, we prove the equivalence between the controllability probability and the optimal state-value function, which is an iteration equation. Besides, the parallel extension technique transforms the observability of PLCNs into the set stabilization of an augmented system. The correspondence between observability probability and optimal state-value function is also derived. Afterward, based on the state-value function, the algorithms via the Q-learning technique are exploited to estimate the controllability and observability probabilities along with obtaining the corresponding optimal control sequences. Finally, all the theoretical results are elaborated via a genetic regulatory p53-Mdm2 network.
AB - This paper studies controllability and observability problems for a class of mixed-valued probabilistic logical control networks (PLCNs). First, PLCN is transformed into the algebraic state space representation (ASSR)-form by resorting to the semi-tensor product method. Then, the formulas are presented to calculate the lower and upper bounds of the transition probability matrix, which further derive the controllability and observability criteria. Furthermore, the ASSR-form of a PLCN can be regarded as a Markov decision process. Using the latter framework, we prove the equivalence between the controllability probability and the optimal state-value function, which is an iteration equation. Besides, the parallel extension technique transforms the observability of PLCNs into the set stabilization of an augmented system. The correspondence between observability probability and optimal state-value function is also derived. Afterward, based on the state-value function, the algorithms via the Q-learning technique are exploited to estimate the controllability and observability probabilities along with obtaining the corresponding optimal control sequences. Finally, all the theoretical results are elaborated via a genetic regulatory p53-Mdm2 network.
KW - Controllability
KW - Observability
KW - Probabilistic Logical Networks
KW - Reinforcement Learning
KW - Semi-Tensor Product
UR - http://www.scopus.com/inward/record.url?scp=85214297336&partnerID=8YFLogxK
U2 - 10.1109/TAC.2024.3524241
DO - 10.1109/TAC.2024.3524241
M3 - Journal article
AN - SCOPUS:85214297336
SN - 0018-9286
SP - 1
EP - 16
JO - IEEE Transactions on Automatic Control
JF - IEEE Transactions on Automatic Control
ER -