Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Neural Network Prediction of ICU Length of Stay Following Cardiac Surgery Based on Pre-Incision Variables

  • Rocco J. LaFaro ,

    Contributed equally to this work with: Rocco J. LaFaro, Suryanarayana Pothula, Keshar Paul Kubal, Mario Emil Inchiosa, Mario Anthony Inchiosa Jr.

    Affiliation Department of Surgery, New York Medical College, Valhalla, New York, United States of America

  • Suryanarayana Pothula ,

    Contributed equally to this work with: Rocco J. LaFaro, Suryanarayana Pothula, Keshar Paul Kubal, Mario Emil Inchiosa, Mario Anthony Inchiosa Jr.

    Affiliation Department of Anesthesiology, New York Medical College, Valhalla, New York, United States of America

  • Keshar Paul Kubal ,

    Contributed equally to this work with: Rocco J. LaFaro, Suryanarayana Pothula, Keshar Paul Kubal, Mario Emil Inchiosa, Mario Anthony Inchiosa Jr.

    Affiliation Department of Pharmacology, New York Medical College, Valhalla, New York, United States of America

  • Mario Emil Inchiosa ,

    Contributed equally to this work with: Rocco J. LaFaro, Suryanarayana Pothula, Keshar Paul Kubal, Mario Emil Inchiosa, Mario Anthony Inchiosa Jr.

    Current address: Information Management and Machine Learning, Microsoft Corporation, Mountain View, California, United States of America

    Affiliation Revolution Analytics, Inc., Mountain View, California, United States of America

  • Venu M. Pothula ,

    Current address: Department of Psychiatry, Mt Sinai Hospital, New York, New York, United States of America

    ‡ These authors also contributed equally to this work.

    Affiliation Department of Pharmacology, New York Medical College, Valhalla, New York, United States of America

  • Stanley C. Yuan ,

    Current address: Department of Anesthesiology, Virginia Mason Medical Center, Seattle, Washington, United States of America

    ‡ These authors also contributed equally to this work.

    Affiliation Department of Anesthesiology, New York Medical College, Valhalla, New York, United States of America

  • David A. Maerz ,

    Current address: Department of Anesthesiology, Mt Sinai Hospital, New York, New York, United States of America

    ‡ These authors also contributed equally to this work.

    Affiliation Department of Pharmacology, New York Medical College, Valhalla, New York, United States of America

  • Lucresia Montes ,

    Current address: Department of Anesthesiology, St. Luke’s Roosevelt Hospital, New York, New York, United States of America

    ‡ These authors also contributed equally to this work.

    Affiliation Department of Pharmacology, New York Medical College, Valhalla, New York, United States of America

  • Stephen M. Oleszkiewicz ,

    Current address: Department of Orthopedic Surgery, Indiana University Health, Indianapolis, Indiana, United States of America

    ‡ These authors also contributed equally to this work.

    Affiliation Department of Pharmacology, New York Medical College, Valhalla, New York, United States of America

  • Albert Yusupov ,

    ‡ These authors also contributed equally to this work.

    Affiliation Department of Anesthesiology, New York Medical College, Valhalla, New York, United States of America

  • Richard Perline ,

    ‡ These authors also contributed equally to this work.

    Affiliation The SAS Institute, Cary, North Carolina, United States of America

  • Mario Anthony Inchiosa Jr.

    Contributed equally to this work with: Rocco J. LaFaro, Suryanarayana Pothula, Keshar Paul Kubal, Mario Emil Inchiosa, Mario Anthony Inchiosa Jr.

    mario_inchiosa@nymc.edu

    Affiliation Department of Pharmacology, New York Medical College, Valhalla, New York, United States of America

Abstract

Background

Advanced predictive analytical techniques are being increasingly applied to clinical risk assessment. This study compared a neural network model to several other models in predicting the length of stay (LOS) in the cardiac surgical intensive care unit (ICU) based on pre-incision patient characteristics.

Methods

Thirty six variables collected from 185 cardiac surgical patients were analyzed for contribution to ICU LOS. The Automatic Linear Modeling (ALM) module of IBM-SPSS software identified 8 factors with statistically significant associations with ICU LOS; these factors were also analyzed with the Artificial Neural Network (ANN) module of the same software. The weighted contributions of each factor (“trained” data) were then applied to data for a “new” patient to predict ICU LOS for that individual.

Results

Factors identified in the ALM model were: use of an intra-aortic balloon pump; O2 delivery index; age; use of positive cardiac inotropic agents; hematocrit; serum creatinine ≥ 1.3 mg/deciliter; gender; arterial pCO2. The r2 value for ALM prediction of ICU LOS in the initial (training) model was 0.356, p <0.0001. Cross validation in prediction of a “new” patient yielded r2 = 0.200, p <0.0001. The same 8 factors analyzed with ANN yielded a training prediction r2 of 0.535 (p <0.0001) and a cross validation prediction r2 of 0.410, p <0.0001. Two additional predictive algorithms were studied, but they had lower prediction accuracies. Our validated neural network model identified the upper quartile of ICU LOS with an odds ratio of 9.8(p <0.0001).

Conclusions

ANN demonstrated a 2-fold greater accuracy than ALM in prediction of observed ICU LOS. This greater accuracy would be presumed to result from the capacity of ANN to capture nonlinear effects and higher order interactions. Predictive modeling may be of value in early anticipation of risks of post-operative morbidity and utilization of ICU facilities.

Introduction

The availability of advanced statistical software for predictive analytical investigations has led to many applications in clinical outcomes research. Artificial neural networks have been used in attempts to weight the contribution of various factors, and their interactions, to a particular medical diagnosis, classification, or clinical outcome. The networks use algorithms that are derived from preexisting data to arrive at the smallest prediction errors when applied to the same combination of factors in newly acquired data. Amato et al. [1] have reviewed the rather extensive literature that has focused on neural network applications in medical diagnosis. They also provide an excellent introduction to the mathematical foundation and design of neural networks, and how they are suggested to simulate the “learning” and “generalization” properties of human neural networks. For example, such networks have been applied to diagnosis of coronary artery syndromes [2,3], interpretation of electrocardiograms [4,5] classification of hemodynamic states in pregnancy [6], prediction of intensive care unit length of stay (ICU LOS) in trauma patients[7], and prediction of an adverse outcome in cardiac surgery [8].

Tu and Guerriere [9] appear to be among the first to apply a neural network model to predict ICU LOS following cardiac surgery. They evaluated 15 pre-operative factors in relation to a binary measure, i.e., greater or less than 48 hours of ICU care. The area under the receiver operator curve (AUC) for their training set was 0.7094, S.E. 0.0224. They tested the trained network on data from an independent set of patients; the AUC for the test set was 0.6960, S.E. 0.0227.

Buchman et al. [10] also evaluated prediction accuracy for a binary outcome measure of ICU LOS. They compared several neural network models with a multiple logistic regression model. Of patients in a general surgical ICU that were surviving on the third postoperative day, they evaluated the prediction of those that would be discharged between days 3 through 6 and those that would require more than 7 days of ICU care. Data for the 11 factors that were included in the prediction models were collected through ICU Day 3. The maximum sensitivity and specificity obtained with a neural network model in these predictions was 97% and 83%, respectively; r2 = 0.57. The multiple logistic regression model achieved a sensitivity and specificity of 72% and 62%, respectively; r2 = 0.26.

Rowan et al. [11] used artificial neural networks to study the influence of 32 demographic factors, medical and surgical histories, and clinical measures in relation to the need for an extended LOS in the post-cardiac surgery ICU. The factors encompassed the pre-operative (15 variables), intra-operative (2 variables), and post-operative (15 variables) periods. The immediate post-operative factors included 15 hemodynamic, respiratory and renal parameters, cognitive state, and blood chemistries and blood gas analyses. They evaluated a binary outcome of less than or greater than 24 hours LOS in the ICU. Ensembles of neural network models were found to strengthen predictive accuracies; the strongest model resulted in an AUC of 0.902 (sensitivity of 91%, specificity of 78%).

We have compared the predictive strength of a linear model to that of an artificial neural network in relation to LOS in the post-cardiac surgical ICU utilizing only pre-incision (i.e. prior to initiation of surgery) variables; the same variables were evaluated in both models. The studies noted above were designed to predict a binary outcome of a shorter or longer period of ICU stay. We have attempted to predict a continuous outcome in hours of ICU LOS. Also, some of the previous studies were not limited to pre-operative/pre-incision variables in their prediction models, but extended to factors derived from the intra-operative, post-operative and early ICU periods.

The primary objective of this study was to demonstrate the predictive strength that may be possible from pre-operative/pre-incision patient characteristics using a neural network model, and its superiority over a linear model; both models were generated with a commercially available statistical software package. It may be noted that statistical strengths could be identified even with a relatively modest patient sample size. It has also been our objective to qualify the perception that neural networks constitute “black boxes;” we have included details of the algorithms that allow each input variable to be tracked and interpreted for its positive, negative, synergistic or neutralizing contribution to the outcome variable.

We also carried out secondary comparative analyses of our data with two additional predictive analytical models, Decision Tree and Random Forest. These models have wide interest because of their inherent transparency in model development.

The odds ratio of our final neural network model to identify patients with the highest risk of prolonged ICU stays is also presented.

Methods

This study was carried out with New York Medical College IRB approval and a Westchester Medical Center (Valhalla, New York) HIPAA waiver to collect medical record data, retrospectively. Data were collected with complete confidentiality and plan for de-identification from a total of 185 patients who underwent coronary artery bypass graft (CABG) or heart valve surgeries, or a combination of both. All patients underwent surgery with cardiopulmonary bypass. A total of 7 patients (3.8%) were excluded as follows: Expiration in the ICU, 1; unplanned return to the operating room from the ICU, 1; additional aortic surgery not including the valve, 3; one renal transplant patient that had remained intubated in the ICU for 7 days before cardiac surgery; one patient with data inadvertently collected twice. The “Automatic Linear Modeling” (ALM) module and the “Artificial Neural Network” (ANN) module of IBM SPSS 21 statistical software (New Orchard Road, Armonk, NY 10504) were used for the primary analyses. Secondary analyses used the “Decision Tree” module of IBM SPSS 22 software, and open source R software [12] was used for “Random Forest” analyses.

Automatic Linear Modeling

A set of 36 pre-incision/pre-operative demographic, anthropomorphic, hemodynamic, medical and cardiac surgical histories, and clinical interventions (Fig 1) were analyzed as input variables with the ALM module; hours of ICU LOS was the continuous output (target) variable. In our application, the model identified 8 pre-incision factors that were statistically associated with ICU LOS: Use of an intra-aortic balloon pump (IABP); O2 delivery index (ml/min/m2); age; use of positive cardiac inotropic agents; blood hematocrit level (%) (HCT); serum creatinine ≥ 1.3 mg/deciliter (dL); gender; arterial pCO2 (mm Hg). The software module preserves a file of the weighted contributions of each of the factors incorporated in the model. These weights represent a “trained” data set; these weights can then be applied to data for a new patient (“untrained data”) to predict ICU LOS for that individual.

thumbnail
Fig 1. Pre-incision variables chosen for predictive analytical modeling from a data base of 185 cardiac surgical patients.

https://doi.org/10.1371/journal.pone.0145395.g001

The predictive accuracy of the trained weights on untrained data was evaluated by the “cross-validation” approach [13]. For this analysis, the data from 3 patients were excluded in training the entire remaining data; the weights were then applied to predict the ICU LOS for the 3 untrained patients. The data for those 3 patients were returned to the data set, and the process was repeated, excluding a new set of 3 patients each time. Thus, a total of approximately 60 analyses were carried out to get predicted ICU LOS for each patient.

Neural Network Modeling

The same 8 variables that were identified by linear modeling to be statistically associated with ICU LOS were used for neural network modeling. Although it might have been possible to identify other variables that would give stronger predictive results for ICU LOS, we considered that a comparison with linear modeling would be best served by using the same variables in this demonstration study.

A considerable number of trials were run in our attempts to optimize the operator options in the design of a neural network with the IBM-SPSS software. Specific settings were as follows:

Initial mode choice

Multilayer Perceptron

Variable entry window

Nominal variables entered as, “Factors”

Ordinal variables entered as, “Factors”

Continuous variables entered as, “Covariates”

Outcome target variable entered as, “Dependent variable”

Partitioning of the cases

All training of the model and cross validation were carried out with 90% of the cases assigned to training and 10% of the cases dedicated to testing (these are randomly assigned by the software with each run). The inclusion of test cases is essential to prevent “overtraining” of the model. The test cases are repeatedly used in the iterative process to optimize the predictive strength of the model.

Rescaling of covariate

Standardized option: the mean is subtracted and the result is divided by the standard deviation; (x-mean)/s

Architecture:

Minimum number of nodes in hidden layer, 1

Maximum number of nodes in hidden layer, 50

Training criteria

Type of training: batch

Optimization algorithm: scaled conjugate gradient

Initial Lambda: 0.0000005

Initial Sigma: 0.00005

Interval center: 0

Interval offset: 0.5

User missing values

Exclude

Stopping rules

Maximum steps without a decrease in error: 5

Default options were used for any other choices.

As with the ALM module, the neural network module preserves a file of the weighted contributions of each of the factors incorporated in the model. These weights represent the trained data set, and can be applied to new untrained data. The cross-validation approach, as described above, was used with repeated sets of 3 patients to arrive at predicted values for ICU LOS for “untrained” data. There is a major difference between ALM and neural networks in regard to the generation of ICU LOS predictions in both the training output and the output in the cross-validation analyses. The linear modeling produces the identical prediction with repeated runs. In comparison, the neural network has an inherent instability (subtle variations) in the output predictions. The iterative analytical process is much more complex in the neural networks. A common approach to accommodate this instability is to use an ensemble of training sets to generate an average output prediction [14,15]. We used an ensemble of 10 trained sets in all predictions of ICU LOS. This procedure does not introduce any time restraints because of the speed of the computer processing of the algorithms. Our analyses were carried out with a common PC laptop computer with a 32-bit operating system, 2.10 GHz microprocessor and 3.00 GB of RAM. Individual training and cross-validation machine runs were typically completed in approximately 1.0 s with our size data sets.

Results

Automatic Linear Modeling

As noted above, 36 pre-incision variables (Fig 1) were analyzed with the ALM module. Eight variables were found to have a statistically significant (or close to statistically significant) correlation with ICU LOS. These included: Pre-incision use of an intra-aortic balloon pump; O2 delivery index; age; pre-incision use of positive cardiac inotropes; hematocrit; serum creatinine ≥ 1.3; gender; arterial pCO2. Their weighted coefficients to the final model, t statistic and p values for linear correlations, and fractional contributions to the final correlation with ICU LOS are presented in Table 1.

thumbnail
Table 1. ALM Fractional Contributions to the Final Correlation with ICU LOS.

https://doi.org/10.1371/journal.pone.0145395.t001

The correlation between the t statistic and the fractional importance in the ALM model is shown in Fig 2. The application of the weighted coefficients to determination of ICU LOS is presented in S1 Appendix.

thumbnail
Fig 2. The correlation between the t statistic and the fractional importance of the 8 pre-incision factors that had the strongest associations with ICU length of stay.

https://doi.org/10.1371/journal.pone.0145395.g002

Of the 36 variables, only the 8 listed above were used to generate the correlation with LOS in the training set (Fig 3A), and their weighted contributions were preserved for the cross-validation of untrained data. The r2 value for the training set was 0.356, p <0.0001. The cross-validation results with untrained data are presented in Fig 3B. The r2 value for the final prediction was 0.200, p <0.0001.

thumbnail
Fig 3. A. ALM Training Model for Prediction of ICU LOS. The regression relationship between the predicted ICU length of stay from 8 pre-incision factors and the observed ICU length of stay from the training model generated by the SPSS Automatic Linear Modeling module. B. ALM Cross-Validated Predictions for ICU LOS.

The cross-validated result of the training model when applied to new ("untrained") patient data.

https://doi.org/10.1371/journal.pone.0145395.g003

Neural Network Modeling

The same 8 variables identified by linear modeling were entered in the neural network module.

The typical model network generated by the software iterations is presented in Fig 4; 4 hidden nodes were typically employed by the software in the optimization process. The neural network predicted output for ICU LOS for the training set is presented in Fig 5A; the r2 for the correlation was 0.535, p <0.0001. As noted above, ensembles of trained data sets were used in cross-validation. The cross-validation results with untrained data with neural network analysis are presented in Fig 5B. The r2 value for the final prediction was 0.410, p <0.0001.

thumbnail
Fig 4. The architecture of the typical neural network utilized in the SPSS Artificial Neural Network module for the 8 pre-incision factors.

https://doi.org/10.1371/journal.pone.0145395.g004

thumbnail
Fig 5. A. ANN Training Model for Prediction of ICU LOS. The regression relationship between the predicted ICU length of stay from 8 pre-incision factors and the observed ICU length of stay from the training model generated by the SPSS Artificial Neural Network module. B. ANN Prediction Results for ICU LOS for Untrained Data.

The cross-validated result of the training model when applied to new ("untrained") patient data.

https://doi.org/10.1371/journal.pone.0145395.g005

A set of typical parameter weights preserved for the cross-validation analysis are presented in Fig 6. The formulas for the application of the parameter weights to the prediction of ICU LOS are presented in S2 Appendix. It should be noted that this represents the algorithm for the neural network.

thumbnail
Fig 6. ANN Parameter Weights.

The prediction weights generated by the neural network for each interaction among the 8 pre-incision factors ("input layer") and the 4 nodes ("hidden layer"), and the output weights of each node to the prediction of ICU length of stay; bias weights are also contributed from the input layer and the hidden layer.

https://doi.org/10.1371/journal.pone.0145395.g006

The typical relative importance of the variables in the neural network modeling for this data set is presented in Fig 7; the algorithm for determination of predictor importance, also termed “sensitivity,” is presented in S3 Appendix.

thumbnail
Fig 7. The relative importance of the 8 pre-incision factors to the anticipation of ICU length of stay (generated from the Artificial Neural Network algorithm).

https://doi.org/10.1371/journal.pone.0145395.g007

We calculated the predicted odds ratio for a patient falling in the upper quartile of the observed ICU LOS vs. the lower 75th percentile from both the ANN training model and the cross-validated model. The threshold for demarcation of the upper quartile for observed ICU LOS for our data is 74.6 hours. Statistically strong odds ratios were found for predicting patients that would be in the upper quartile of ICU LOS for both the training and cross-validated models (Table 2).

thumbnail
Table 2. Odds Ratios for Prediction that a Patient would have ICU LOS in the Upper Quartile of Observed Data from both the Trained and Cross-Validated ANN Analyses.

https://doi.org/10.1371/journal.pone.0145395.t002

We also carried out secondary analyses of our data with the SPSS “Decision Tree” algorithm and with “Random Forest” modeled in R. The decision tree analysis was initiated with the original 36 input variables and was assessed for strength of the prediction accuracy by cross-validation of successive levels of branching. We found that four levels of branching produced the greatest prediction accuracy. This tree is presented in Fig 8. It utilized 7 of the input variables at this optimal training level. It shared only 3 of the same variables that were modeled with ALM and ANN: hematocrit (HCT); oxygen delivery index; use of inotropic agents. The prediction accuracy for decision tree in the training algorithm, r2 = 0.502, p <0.0001, (Fig 9A) was similar to that for ANN; however, cross-validation with the same “leave-three-out” procedure as described above produced a considerably weaker prediction accuracy (r2 = 0.113, p <0.0001; Fig 9B) than that found with ANN.

thumbnail
Fig 8. The Final Decision Tree Model that was Assessed for Prediction of ICU LOS.

https://doi.org/10.1371/journal.pone.0145395.g008

thumbnail
Fig 9. A. Decision Tree Training Model for Prediction of ICU LOS. The regression relationship between the predicted ICU length of stay from 7 pre-incision factors and the observed ICU length of stay from the training model generated by the SPSS Decision Tree module. B. Cross-Validated Predictions for ICU LOS.

The cross-validated result of the training model when applied to new ("untrained") patient data.

https://doi.org/10.1371/journal.pone.0145395.g009

Finally, for the Random Forest modeling, we started with the 8 input variables modeled with ALM and ANN. We used default values for the Random Forest modeling function [12] running in the R software environment [16], yielding a Random Forest model consisting of 500 decision trees. The Random Forest model fit the data with r2 = 0.836, p <0.0001 (Fig 10A). The prediction accuracy, evaluated using leave-one-out cross-validation was weaker (r2 = 0.303, p <0.0001; Fig 10B) than that found with ANN.

thumbnail
Fig 10. A. Random Forest Training Model for Prediction of ICU LOS. The regression relationship between the predicted ICU length of stay from 8 pre-incision factors and the observed ICU length of stay from the training model. B. Cross-Validated Predictions for ICU LOS.

The cross-validated result of the training model when applied to new ("untrained") patient data.

https://doi.org/10.1371/journal.pone.0145395.g010

A comparison of the prediction accuracies that were found with the four models that were studied is presented in Table 3.

thumbnail
Table 3. Comparison of the Accuracies for the Four Models used to Predict ICU LOS.

https://doi.org/10.1371/journal.pone.0145395.t003

Discussion and Conclusions

We have attempted to develop a neural network to model a continuous measure of hours of ICU LOS following cardiac surgery. It must be noted that non-neural network models have been successfully applied to prediction of ICU-LOS following cardiac surgery. A large prospective study by Barili et al. [17], utilizing both pre-operative variables and surgical data, identified four factors by regression modeling that significantly increased the risk of a prolonged ICU stay following cardiac surgery: Critical preoperative state; surgical emergency; poor left-ventricular function; and elevated serum creatinine. A critical/emergency preoperative state (represented by the need for an intra-aortic balloon pump and preoperative use of positive inotropic agents in our results) as well as elevated serum creatinine were also identified in our analyses. Najafi and Goodarzynejad [18] analyzed factors that would predict ICU-LOS of greater or less than 48 hours. They applied multivariate analysis to pre-operative, intra-operative, and post-operative factors, extending to 6 hours post-operatively for arterial gas analyses and 24 hours for insulin intake. Six factors were independent predictors of ICU-LOS greater than 48 hours: Intra-aortic balloon pump; New York Heart Association functional class; post-operative arrhythmia; 24-hour average insulin intake; mean 6-hour base excess; and a surgeon category.

In contrast to some of the neural network studies referenced above [911], we focused on the earliest factors that could be modeled for prediction accuracy. Thus, we limited the variables to those that could be obtained pre-incision. Efforts to identify a set of factors that might be expected intuitively to be useful as predictors of our outcome measure proved difficult. Instead, we subjected our data set of 36 variables (Fig 1) to analysis by the Automatic Linear Modeling module of the IBM-SPSS 21 software. Eight factors were identified that had statistically significant (or approaching significant) associations with ICU LOS (Table 1). It was found that the t statistic and the fractional importance in the ALM model were closely correlated (Fig 2). (The application of the weighted coefficients to determination of ICU LOS is presented in S1 Appendix.)

The trained ALM model for prediction of ICU LOS is presented in Fig 3A (r2 = 0.356, p <0.0001). As discussed above, we used a cross-validation approach to test the accuracy of predictions on “untrained” patient data. In this case, repeated sets of data from 3 patients were tested with the trained model obtained from all but those 3 patients. The accuracy of the predictions on “untrained” data was r2 = 0.200, p <0.0001 (Fig 3B).

The same 8 factors identified by linear modeling were used in the neural network modeling. It is possible that the inclusion of additional variables might produce even stronger predictive neural network models than we obtained, however, we felt that a direct comparison with the linear modeling would be of interest at this stage of our evaluation of the two models. The neural network trained the data set with an r2 value of 0.535, p <0.0001 (Fig 5A). Cross-validation of an ensemble of trained models on repeated sets of 3 untrained patient data, as noted above, produced a predicted r2 value of 0.410, p <0.0001 (Fig 5B). The more than twice stronger prediction accuracy of the neural network over the linear model (r2 0.410 vs 0.200) would appear to be related to the contribution of the interactions among the variables to the final predictions. In addition, the rank order of the relative importance of the 8 pre-incision variables to ICU LOS was not identical (Table 4). Presumably, this is again related to the neural network’s capability to capture nonlinear effects, including higher order interactions; linear modeling only considers the strengths of the correlations of the individual factors with the target outcome, ICU LOS.

thumbnail
Table 4. Fractional Importance differences between ALM and ANN.

https://doi.org/10.1371/journal.pone.0145395.t004

Analysis of the 36 input variables with the use of Decision Tree or Random Forest demonstrated that these models could produce training algorithms that had approximately the same or greater predictive accuracy as ANN (Figs 9A and 10A); however, they cross-validated with less accuracy (Figs 9B and 10B; Table 3). It is not suggested that Decision Tree or Random Forest are weaker predictive analytical tools; it is known that performance may vary with the characteristics of the database [19].

It should be emphasized that the predictions of ICU LOS from the neural network models are not calculated from the fractional importance presented in Fig 7 or Table 4. The hours of ICU LOS for the ANN modeling is based on parameter weights (e.g., as in Fig 6) and the algorithm presented in S2 Appendix. The fractional importance measures of Fig 7 or Table 4, also termed “sensitivity” measures, (S3 Appendix) may be useful in identifying factors that warrant clinical consideration in terms of strengths or risks.

It would be expected that a larger data base would yield still stronger prediction accuracy. As new patients are added to the data base, the ICU LOS for each new patient can be predicted from ensembles of trained models from our entire data set. Also, each new patient will next become part of the training set. In principle, each new patient should add to the training experiences such that it becomes increasingly more likely that data from a new untrained patient will already have a close counterpart in the trained data base. At a clinical practice level, for example, data for the 8 input variables can be entered in the software (manually or from an electronic record) and the predicted ICU LOS can be scored from the average output of an ensemble of models generated from the entire dataset. As noted above, we include 10% testing in training the models. This represents our “final” ANN model at this point. The inherent speed of the calculations would provide an almost immediate prediction.

In regard to the possible potential value of neural networks to anticipate high risk patients, our cross-validated network stratified the patients in the upper quartile of observed ICU LOS with an odds ratio of 9.8, p <0.0001 (Table 2).

In conclusion, we believe that our experience with neural network modeling with a relatively small data base demonstrates the potential strength of this approach for identifying and weighting prognosticators of outcomes. The results demonstrate that 41% (the r2 value from the neural network cross-validation) of the variation in ICU LOS among these cardiac surgical patients was influenced by 8 pre-incision factors. In one sense, this is quite substantial considering the possible contributions from the surgery itself, the surgical team personnel (surgeons, anesthesiologists, perfusionists, nurses), and the ICU environment and staff. In terms of our outcome measure of ICU LOS, it may have value in identifying the risk of postoperative morbidity. Such analyses may also have economic implications in relation to expected costs and utilization of ICU facilities.

Limitations

The most serious limitation of this study is the relatively small data base that could be collected during the period that we had IRB approval in place and patient files that had not yet been designated for off-campus storage. However, as noted above, the results give encouragement that modeling of even modest size data bases, which are common in pilot studies, can give indications of possibly worthwhile further investigation. It is also obvious that these studies did not approach the development of new paradigms of predictive modeling, such as those of Zhang et al. [20] and Deng et al. [21]. The paper by Zhang and coworkers presented the evidence that their advanced neural network model, “Extreme Machine Learning,” generally outperformed the more conventional neural network model represented in this present study in classification of several cancer diagnoses from microarray gene expression data [20]. A most recent study by Deng et al. employs still more original algorithm development to classify positive and negative diagnoses in epilepsy, breast cancer and heart disease [21]. Our work is limited to the application of open source software in identifying possibly important associations between patient characteristics and clinical outcomes; however, such studies are often useful in identifying risk factors, guiding treatments, and focusing further investigations.

Supporting Information

S1 Appendix. Application of the coefficients generated in the Automatic Linear Modeling module to calculate the predicted ICU length of stay.

https://doi.org/10.1371/journal.pone.0145395.s001

(TIF)

S2 Appendix. Algorithm for application of the parameter weights generated in the Artificial Neural Network modeling module to calculate the predicted ICU length of stay.

https://doi.org/10.1371/journal.pone.0145395.s002

(TIF)

S3 Appendix. Algorithm for calculation of the relative importance of the input variable (Predictor Importance) in Artificial Neural Network modeling.

https://doi.org/10.1371/journal.pone.0145395.s003

(TIF)

S4 Appendix. Complete database for these studies.

https://doi.org/10.1371/journal.pone.0145395.s004

(CSV)

Author Contributions

Conceived and designed the experiments: MAI KPK RJL SP. Analyzed the data: RJL SP KPK MEI VMP SCY DAM LM SMO AY RP MAI. Contributed reagents/materials/analysis tools: MAI KPK MEI RP DAM VMP. Wrote the paper: MAI KPK RJL SP.

References

  1. 1. Amato F, Lobez A, Pena-Mendez EM, Vanhara P, Hampl A, Havel J. Artificial Neural Networks in Medical Diagnosis. Journal of Applied Biomedicine. 2013;11: 47–58.
  2. 2. Atkov OY, Gorokhova SG, Sboev AG, Generozov EV, Muraseyeva EV, Moroshkina SY et al. Coronary heart disease diagnosis by artificial neural networks including genetic polymorphisms and clinical parameters. J Cardiol. 2012;59: 190–194. pmid:22218324
  3. 3. Harrison RF, Kennedy RL. Artificial neural network models for prediction of acute coronary syndromes using clinical data from the time of presentation. Ann Emerg Med. 2005;46: 431–439. pmid:16271675
  4. 4. Atoui H, Fayn J, Rubel P. A novel neural-network model for deriving standard 12-lead ECGs from serial three-lead ECGs: application to self-care. IEEE Trans Inf Technol Biomed. 2010;14: 883–890. pmid:20378474
  5. 5. Mateo J, Rieta JJ. Application of artificial neural networks for versatile preprocessing of electrocardiogram recordings. J Med Eng Technol. 2012;36: 90–101. pmid:22268996
  6. 6. Tejera E, Jose AM, Rodrigues A, Ramoa A, Manuel Nieto-Villar J, Rebelo I. Artificial neural network for normal, hypertensive, and preeclamptic pregnancy classification using maternal heart rate variability indexes. J Matern Fetal Neonatal Med. 2011;24: 1147–1151. pmid:21250912
  7. 7. Gholipour C, Rahim F, Fakhree A, Ziapour B. Using an Artificial Neural Networks (ANNs) Model for Prediction of Intensive Care Unit (ICU) Outcome and Length of Stay at Hospital in Traumatic Patients. J Clin Diagn Res. 2015;9: OC19–OC23.
  8. 8. Peng SY, Peng SK. Predicting adverse outcomes of cardiac surgery with the application of artificial neural networks. Anaesthesia. 2008;63: 705–713. pmid:18582255
  9. 9. Tu JV, Guerriere MR. Use of a neural network as a predictive instrument for length of stay in the intensive care unit following cardiac surgery. Comput Biomed Res. 1993;26: 220–229. pmid:8325002
  10. 10. Buchman TG, Kubos KL, Seidler AJ, Siegforth MJ. A comparison of statistical and connectionist models for the prediction of chronicity in a surgical intensive care unit. Crit Care Med. 1994;22: 750–762. pmid:8181282
  11. 11. Rowan M, Ryan T, Hegarty F, O'Hare N. The use of artificial neural networks to stratify the length of stay of cardiac patients based on preoperative and initial postoperative factors. Artif Intell Med. 2007;40: 211–221. pmid:17580112
  12. 12. Liaw A, Wiener M (2002) Classification and Regression by randomForest. 2: 18–22.
  13. 13. Geisser S (1993) Predictive Inference: An Introduction. Chapman & Hall.
  14. 14. Hashem S. Optimal Linear Combinations of Neural Networks. Neural Netw. 1997;10: 599–614. pmid:12662858
  15. 15. Walsh P, Cunningham P, Rothenberg SJ, O'Doherty S, Hoey H, Healy R. An artificial neural network ensemble to predict disposition and length of stay in children presenting with bronchiolitis. Eur J Emerg Med. 2004;11: 259–264. pmid:15359198
  16. 16. R Core Team (2015). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. Available: https://www.R-project.org/.
  17. 17. Barili F, Barzaghi N, Cheema FH, Capo A, Jiang J, Ardemagni E et al. An original model to predict Intensive Care Unit length-of stay after cardiac surgery in a competing risk framework. Int J Cardiol. 2013;168: 219–225. pmid:23040076
  18. 18. Najafi M, Goodarzynejad H. Determinants of length of stay in surgical ward after coronary bypass surgery: glycosylated hemoglobin as a predictor in all patients, diabetic or non-diabetic. J Tehran Heart Cent. 2012;7: 170–176. pmid:23323078
  19. 19. Hall LO, Liu X, Bowyer KW, Banfield R (2003) Why are neural networks sometimes much more accurate than decision trees: an analysis on a bioinformatics problem. 3: 2851–2856.
  20. 20. Zhang R, Huang GB, Sundararajan N, Saratchandran P. Multi-category classification using an Extreme Learning Machine for microarray gene expression cancer diagnosis. IEEE/ACM Trans Comput Biol Bioinform. 2007;4: 485–495. pmid:17666768
  21. 21. Deng Z, Cao L, Jiang Y, Wang S. Minimax Probability TSK Fuzzy System Classifier: A More Transparent and Highly Interpretable Classification Model. IEEE Transactions on Fuzzy Systems. 2015;23: 813–825.