Machine learning guided prediction of warfarin blood levels for personalized medicine based on clinical longitudinal data from cardiac surgery patients: a prospective observational study
View/ Open
Date
2024-10Author
Xue, Ling
He, Shan
Singla, Rajeev K.
Qin, Qiong
Ding, Yinglong
Liu, Linsheng
Ding, Xiaoliang
Bediaga Bañeres, Harbil
Arrasate Gil, Sonia
Durado Sánchez, Aliuska
Zhang, Yuzhen
Shen, Zhenya
Shen, Bairong
Miao, Liyan
González Díaz, Humberto
Metadata
Show full item record
International Journal of Surgery 110(10) : 6528-6540 (2024)
Abstract
Background: Warfarin is a common oral anticoagulant, and its effects vary widely among individuals. Numerous dose-prediction
algorithms have been reported based on cross-sectional data generated via multiple linear regression or machine learning. This study
aimed to construct an information fusion perturbation theory and machine-learning prediction model of warfarin blood levels based
on clinical longitudinal data from cardiac surgery patients.
Methods and material: The data of 246 patients were obtained from electronic medical records. Continuous variables were
processed by calculating the distance of the raw data with the moving average (MA Δvki(sj)), and categorical variables in different
attribute groupswere processed using Euclidean distance (ED ǁΔvk(sj)ǁ). Regression and classification analyses were performed on the
raw data, MA Δvki(sj), and ED ǁΔvk(sj)ǁ. Different machine-learning algorithms were chosen for the STATISTICA and WEKA software.
Results: The random forest (RF) algorithm was the best for predicting continuous outputs using the raw data. The correlation
coefficients of the RF algorithm were 0.978 and 0.595 for the training and validation sets, respectively, and the mean absolute errors were
0.135 and 0.362 for the training and validation sets, respectively. The proportion of ideal predictions of the RF algorithm was 59.0%.
General discriminant analysis (GDA) was the best algorithm for predicting the categorical outputs using the MA Δvki(sj) data. The GDA
algorithm’s total true positive rate (TPR) was 95.4% and 95.6% for the training and validation sets, respectively, with MA Δvki(sj) data.
Conclusions: An information fusion perturbation theory and machine-learning model for predicting warfarin blood levels was
established. A model based on the RF algorithm could be used to predict the target international normalized ratio (INR), and a model
based on the GDA algorithm could be used to predict the probability of being within the target INR range under different clinical scenarios.
Collections
Except where otherwise noted, this item's license is described as © 2024 The Author(s). Published by Wolters Kluwer Health, Inc. This is an open access article distributed under the Creative Commons Attribution License 4.0
(CCBY), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited