TY - GEN
T1 - Deep inverse reinforcement learning for sepsis treatment
AU - Yu, Chao
AU - Ren, Guoqi
AU - LIU, Jiming
N1 - Publisher Copyright:
© 2019 IEEE.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2019/6
Y1 - 2019/6
N2 - Sepsis is a leading cause of mortality in hospitals, but its optimal treatment strategy still remains unclear. Recent years have witnessed several successful applications of Reinforcement Learning (RL) approaches in sepsis treatment, achieving far more efficient strategies than those by clinicians. To ensure such applications, an explicit reward function encoding medical domain knowledge should be specified beforehand to indicate the goal of learning. However, due to the paucity of clear understanding of sepsis itself, there is still considerable inconsistency in the formulation of reward functions for sepsis treatment. In this poster, we address the reward learning problem in RL for treatment of sepsis, which has been largely neglected by previous studies. A deep inverse RL with Mini-Tree (DIRL-MT) model is proposed to infer the best reward functions from a set of presumably optimal treatment trajectories using retrospective real medical data. In the model, the MT component learns the factors that are most important in influencing the mortality during sepsis treatment, while the DIRL component infers the complete reward function in terms of weights of those factors. Our work shows that PaO2 and PT can play a vital role and should be paid more attention in the design of more efficient treatment strategies for sepsis in the future.
AB - Sepsis is a leading cause of mortality in hospitals, but its optimal treatment strategy still remains unclear. Recent years have witnessed several successful applications of Reinforcement Learning (RL) approaches in sepsis treatment, achieving far more efficient strategies than those by clinicians. To ensure such applications, an explicit reward function encoding medical domain knowledge should be specified beforehand to indicate the goal of learning. However, due to the paucity of clear understanding of sepsis itself, there is still considerable inconsistency in the formulation of reward functions for sepsis treatment. In this poster, we address the reward learning problem in RL for treatment of sepsis, which has been largely neglected by previous studies. A deep inverse RL with Mini-Tree (DIRL-MT) model is proposed to infer the best reward functions from a set of presumably optimal treatment trajectories using retrospective real medical data. In the model, the MT component learns the factors that are most important in influencing the mortality during sepsis treatment, while the DIRL component infers the complete reward function in terms of weights of those factors. Our work shows that PaO2 and PT can play a vital role and should be paid more attention in the design of more efficient treatment strategies for sepsis in the future.
KW - Deep Learning
KW - Intravenous
KW - Inverse Learning
KW - Reinforcement Learning
KW - Sepsis
KW - Vasopressor
UR - http://www.scopus.com/inward/record.url?scp=85075938691&partnerID=8YFLogxK
U2 - 10.1109/ICHI.2019.8904645
DO - 10.1109/ICHI.2019.8904645
M3 - Conference proceeding
AN - SCOPUS:85075938691
T3 - 2019 IEEE International Conference on Healthcare Informatics, ICHI 2019
BT - 2019 IEEE International Conference on Healthcare Informatics, ICHI 2019
PB - IEEE
T2 - 7th IEEE International Conference on Healthcare Informatics, ICHI 2019
Y2 - 10 June 2019 through 13 June 2019
ER -