Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Index Blood Tests and National Early Warning Scores within 24 Hours of Emergency Admission Can Predict the Risk of In-Hospital Mortality: A Model Development and Validation Study

  • Mohammed A. Mohammed ,

    m.a.mohammed@bham.ac.uk

    Affiliation Primary Care Clinical Sciences, University of Birmingham, Edgbaston, Birmingham, United Kingdom

  • Gavin Rudge,

    Affiliation Unit of Public Health, Epidemiology and Biostatistics, University of Birmingham, Edgbaston, Birmingham, United Kingdom

  • Duncan Watson,

    Affiliation University Hospital Coventry and Warwickshire, Coventry, United Kingdom

  • Gordon Wood,

    Affiliation George Eliot Hospital, Nuneaton, Warwickshire, United Kingdom

  • Gary B. Smith,

    Affiliation Centre of Postgraduate Medical Research and Education, School of Health and Social Care, Bournemouth University, Royal London House, Bournemouth, Dorset, United Kingdom

  • David R. Prytherch,

    Affiliations Portsmouth Hospitals NHS Trust, Portsmouth, United Kingdom, Centre for Healthcare Modelling, School of Computing, University of Portsmouth, Portsmouth, United Kingdom

  • Alan Girling,

    Affiliation Primary Care Clinical Sciences, University of Birmingham, Edgbaston, Birmingham, United Kingdom

  • Andrew Stevens

    Affiliation Unit of Public Health, Epidemiology and Biostatistics, University of Birmingham, Edgbaston, Birmingham, United Kingdom

Abstract

Background

We explored the use of routine blood tests and national early warning scores (NEWS) reported within ±24 hours of admission to predict in-hospital mortality in emergency admissions, using empirical decision Tree models because they are intuitive and may ultimately be used to support clinical decision making.

Methodology

A retrospective analysis of adult emergency admissions to a large acute hospital during April 2009 to March 2010 in the West Midlands, England, with a full set of index blood tests results (albumin, creatinine, haemoglobin, potassium, sodium, urea, white cell count and an index NEWS undertaken within ±24 hours of admission). We developed a Tree model by randomly splitting the admissions into a training (50%) and validation dataset (50%) and assessed its accuracy using the concordance (c-) statistic. Emergency admissions (about 30%) did not have a full set of index blood tests and/or NEWS and so were not included in our analysis.

Results

There were 23248 emergency admissions with a full set of blood tests and NEWS with an in-hospital mortality of 5.69%. The Tree model identified age, NEWS, albumin, sodium, white cell count and urea as significant (p<0.001) predictors of death, which described 17 homogeneous subgroups of admissions with mortality ranging from 0.2% to 60%. The c-statistic for the training model was 0.864 (95%CI 0.852 to 0.87) and when applied to the testing data set this was 0.853 (95%CI 0.840 to 0.866).

Conclusions

An easy to interpret validated risk adjustment Tree model using blood test and NEWS taken within ±24 hours of admission provides good discrimination and offers a novel approach to risk adjustment which may potentially support clinical decision making. Given the nature of the clinical data, the results are likely to be generalisable but further research is required to investigate this promising approach.

Introduction

There is considerable interest in developing statistical models which can adjust for patient case-mix and predict the risk of death in hospital. Whilst there are several different models in widespread use [1], [2], [3], [4], [5], [6], [7] a common feature of existing approaches is that they generally rely on routinely collected clinically coded administrative databases [8]. Whilst such data are readily obtained [9] it is important to recognise that clinical coding is not an integral part of the clinical decision making process and this has serious implications (which ultimately can undermine risk adjustment schemes), viz:- (1) Differences in the clinical coding processes at different hospitals (or even amongst coders in the same hospital) can produce materially different primary and secondary diagnoses codes for the same patient episode. (2) The era of case-mix adjusted hospital mortality ratios has seen changes, some of which amount to gaming, in the clinical coding process usually aimed at reducing the headline mortality ratio, thereby giving the possibly misleading impression that mortality has fallen because of quality of care [10]. It is therefore unsurprising to learn that there is evidence that hospital mortality ratios may sometimes reflect differences in clinical coding practices as opposed to genuine differences in quality of care [11], [12], [13]. (3) The clinically coded data frame is completed after the patient has been discharged and so restricts case-mix adjustment models that rely on coded data to retrospective use with no prospect of real-time decision support for clinical decision making.

Given these limitations we sought risk adjustment covariates from other routinely collected data sources which could be used to predict the risk of patients dying in hospital. Notwithstanding the use of such data to produce hospital mortality ratios, our primary focus is on developing tools which may ultimately support clinical decision making in real time. Candidate covariates, therefore, need to be routinely available on, or near, admission, be integral to the clinical decision making process and be measured with high quality without being susceptible to gaming. As most deaths in hospital occur in patients admitted as emergencies, we focused on these and noted that very early into their care, almost all of these patients will have a routine blood test and, at least in the National Health Service (NHS), an early warning score (EWS) derived from a range of vital signs (e.g. respiration rate, blood pressure, heart rate, temperature) measurements [14]. These data are not subject to many of the shortcomings of administrative clinically coded data. Laboratory tests are subject to strict quality control, and their results are clinically meaningful, have face-validity and are collected as part of the process of care. Similarly, EWS values generated from physiological measurements are part of the care process and are unlikely to be intentionally altered, although their accuracy depends upon correct equipment calibration, measurement technique and the arithmetical skills of staff, although it has been shown that electronic aids can improve accuracy [15], [16], [17].

Since our focus is on supporting clinical decision making, we used Classification and Regression Trees, (CART), which is a statistical data mining technique for constructing empirical decision Trees by recursively splitting or partitioning patients into homogenous subgroups [18]. Tree models have been used to support medical decision making [19], [20], [21].Although their use is still somewhat novel, Tree models, unlike regression models, are intuitive to interpret because they have a simple flowchart type presentation which starts by identifying the most important predictor variables, naturally incorporate interaction effects, identify cut-offs for continuous covariates, are distribution free and can handle non-linear relationships. Some of these characteristics reflect human decision making processes.

We therefore sought to determine if the combination of routine blood test results and electronically collected EWS values could be used in a Tree model to predict in-hospital mortality with a reasonable degree of accuracy.

Methods

Setting and Data

Our cohort of emergency admissions is from University Hospitals Coventry and Warwickshire NHS Trust, one of the largest acute teaching hospitals in the UK with 1250 beds. All spells following emergency admission within the period April 2009 to March 2010 were included. For each admission we obtained the following: patient’s age, gender, admission date/time, discharge date/time and discharge status (alive/dead). The following were excluded - patients aged less than 16 years of age, admissions to the maternity unit or any admissions with missing or invalid data. Using a pseudonymised, unique patient identifier, we obtained linked admission spells with the database of EWS values collected from patients’ vital signs using commercially available, personal digital assistants running specifically designed software (VitalPAC™, The Learning Clinic Ltd, London) [22]. VitalPAC™ has a series of robust real-time data validation checks to ensure that valid physiological measurements are entered at the point of care and accurate EWS are automatically generated [22]. The index EWS was defined as the first recorded value within a ±24 hour window either side of the admission date/time. We converted the local EWS into the recently published National EWS (NEWS) [23] for the NHS to aid generalisability. This conversion from EWS to NEWS was straightforward because EWS and NEWS have the same seven underlying physiological variables (respiration rate, oxygen saturations, any supplemental oxygen (yes/no), temperature, systolic blood pressure, heart rate and level of consciousness (Alert, Voice, Pain, Unresponsive). The accuracy of this conversion was assured by using the actual physiological measurements that was entered in VitalPAC™ and mapping these onto the NEWS scoring system [23].

These emergency admissions data were linked using unique (but de-identified) patient identifiers to the hospital laboratory computer system to determine the first blood test reported within a ±24 hour window (either side of the admission date/time). We included tests and EWS up to 24 hours before the admission date because it is not unusual for patients to have these available in the Accident and Emergency (A&E) department just before being formally admitted to the hospital. Blood tests outside this ±24 hour window were not regarded as index blood tests and were excluded. Emergency admissions without a full set of data were excluded from this analysis.

We considered the following commonly undertaken seven blood tests:- albumin (g/L), creatinine (µmol/L), haemoglobin (g/dL), potassium (mmol/L), sodium (mmol/L), urea (mmol/L) and white cell count (WCC) (109 cells/L) because they are routinely undertaken. Like most hospitals, the blood tests results in UHCW are subject to day-to-day internal and external quality assurance [24]. The NEWS ranged from 0 (indicating the lowest severity of illness) to 19 (the maximum NEWS value possible is 20), although we capped this at 8, because only 346 (1.49%) admissions had higher scores.

Statistical Analyses

We began by undertaking exploratory analysis of the NEWS and the blood test results. To avoid undue influence of extreme outliers to the modelling process, very high (> = 99.9% centile) or very low (< = 1% centile), blood test results were truncated to the limit of the inequality where necessary. We produced scatter plots showing the relationship between mortality and NEWS and the blood test results (grouped into sextiles).

We modelled the risk of death using Classification and Regression Trees. Our Tree modelling strategy involved the following covariates – age, gender, albumin, creatinine, haemoglobin, potassium, sodium, urea, WCC and NEWS. When first developed, CARTs could lead to quite large Tree models, but recent work has incorporated p-value based Tree modelling, known as conditional Trees, which yield smaller Tree models whilst simultaneously controlling for multiple testing, (Bonferroni adjustment, based on p≤0.001). They are available in the Party Package [25] in R. The Tree models produce a flowchart type of output which is intuitive. The Tree models have nodes (shown as ovals) and branches (shown as lines) and end up locating homogeneous subgroups in terminal nodes which are represented by a rectangular box at the bottom of the Tree with the sample size indicated by n =  and the risk of death as y = . Tree models can also be summarised in nested tabular format to which we added 95% binomial confidence intervals using the exactci package [26] in R [27]. All analyses were undertaken in R.

We randomly divided our dataset into a training set (n = 11624) and a testing set (n = 11624) for model validation [28]. In assessing the Tree model we used the concordance measure of discrimination – the c-statistic. For a binary outcome, the c-statistic is the area under the Receiver Operating Characteristics, (ROC) [29] curve. The ROC curve is a plot of the sensitivity, (true positive rate), versus 1-specificity, (false positive rate), for consecutive predicted risks. The area under the ROC curve is interpreted as the probability that a deceased patient has a higher risk of death than a randomly chosen non-deceased patient. A c-statistic of 0.5 is no better than tossing a coin, whilst a perfect model has a c-statistic of 1. The higher the c-statistic the better the model. In general, values less than 0.7 are considered to show poor discrimination, values of 0.7 to 0.8 can be described as reasonable, and values above 0.8 suggest good discrimination. The 95% confidence interval for the c-statistic was derived using DeLong’s method as implemented in the pROC library [30] in R [31].

Ethical Approval

The lead author (MAM) sought advice from chair of the Birmingham research ethics committee and was advised that formal ethical approval was not necessary because this is an audit/service evaluation of existing hospital data. However, a co-author (DP) has obtained ethical approval from the Isle of Wight, Portsmouth & South East Hampshire Research Ethics Committee (Reference No: 08/02/1394) for desktop analysis of routinely collected NHS data. All data was de-identified prior to analysis. The datasets used in our study are routinely collected as part of the process of care whilst patients stay in hospital and therefore do not require specific consent.

Results

There were 23248 emergency admissions from April 2009 to March 2010, with a mean age of 61.54 years (SD 21.59), a female to male ratio of 1.09 and accompanying hospital mortality of 5.69% (1323/23248). Table 1 shows the mean and standard deviationsfor the individual blood test results and index NEWS values in admissions that were discharged alive and dead. In general, patients who died were older, had higher index NEWS, lower albumin, higher creatinine, lower haemoglobin, higher potassium, higher urea and higher WCC. Little difference was seen in the mean sodium values. Figure 1 shows the relationship between these variables and mortality. All variables showed a mostly non-linear relationship with mortality.

thumbnail
Figure 1. Scatter plots showing observed risk of death with NEWS and seven blood tests results.

https://doi.org/10.1371/journal.pone.0064340.g001

thumbnail
Table 1. Mean and (standard deviation) for the continuous covariates in the Tree model.

https://doi.org/10.1371/journal.pone.0064340.t001

We developed a Tree model (see figure 2) with age, NEWS and the seven blood test results. The c-statistic for the training model was 0.864 (95%CI 0.852 to 0.876) and when applied to the testing data set this was 0.853 (95%CI 0.840 to 0.866).

The Tree identified age, NEWS, albumin, WCC and urea as significant predictors of death. The appearance of some of these covariates (eg age, albumin) in more than one branch of the Tree model reflects the presence of significant interaction effects, wherein the relationship between a given covariate and risk of death changes depending on its context with respect to other covariates in the Tree. The Tree model involved 33 nodes of which 17 were terminal nodes (grey rectangles at bottom of Tree). For convenience, the 17 terminal nodes (ie the 17 homogenous subgroups) are summarised in table 2 along with 95% confidence intervals. The highest risk was seen in terminal node 33 which involved 25 emergency admissions with a 60% risk of death. These admissions were characterised by albumin >34.5, NEWS >5 and urea>26.3. The lowest risk of death (0.2%) was seen in terminal node 18 which had the largest sample size (n = 4909) and characterised emergency admissions with age ≤76, NEWS ≤3, albumin >37, urea ≤8.9 and WCC ≤15.52.

thumbnail
Table 2. Shows the characteristics of admission in the seventeen terminal nodes (in descending risk) of the Tree model.

https://doi.org/10.1371/journal.pone.0064340.t002

Discussion

We have used a simple set of clinical variables based on the index routine blood test results and NEWS to develop a validated risk adjustment model for in-hospital mortality following emergency admission. Unlike case-mix adjustment schemes that rely on administrative datasets, our approach has employed covariates that are routinely measured, routinely quality assured/audited, are an integral part of the process of care, are not susceptible to gaming, have clinical face-validity and are an established clinical currency within and between hospitals. These features of the covariate set are important because they help to protect against two major issues in case-mix adjustment – the constant risk fallacy [32] (where statistical adjustment increases bias) and the case-mix adjustment fallacy [33] (where differences in adjusted mortality rates does not reflect quality of care). The constant risk fallacy can occur, for example, when two identical patients have differing covariates (eg two identical unconscious stroke patients, but one gets a primary diagnosis label of syncope and the other as stroke; risk adjustment based on these diagnostic labels may now increase, not decrease, bias). In our approach it is not usually possible for two identical patients to have different blood test results even if different hospitals use different reference ranges. Our approach is immune to this because we use the actual values and ignore the reference ranges. It is however possible, especially using pen and paper methods, that two identical patients may have different NEWS, although the use of electronic aids appears to have addressed such measurement problems [17], [34]. The case-mix adjustment fallacy operates when it is presumed that the residual unexplained variation in risk-adjusted mortality reflects quality of care. In risk-adjustment schemes that use administrative data sets, difference in adjusted rates can often simply reflect differences in clinical coding practices as opposed to quality of care [11]. In our approach there is much less scope for identifying problems with the underlying data and hence more scope for identifying clinical issues and this is essential for any tool aimed at supporting the clinical decision making process.

Several previous studies have suggested the potential role of blood test results [35], [36], [37], [38], [39], [40], [41], [42] or patient physiology [43], [44] but few studies have combined these two data sources [45], [46], [47]. In our case, this was more feasible because UHCW uses an electronic system for EWS calculation [18]. In some respects the finding that a valid model can be obtained using the patient’s age, routine blood tests results and NEWS, is unsurprising because these variables have been an integral part of hospital medicine for decades, and as the scatter plots (figure 2) show, they have an empirical relationship with the risk of death. Although the Tree model did not identify all seven blood tests, it would be incorrect to infer that the blood tests (eg potassium, haemoglobin) that were not included are therefore of little clinical use. The usefulness or otherwise of a blood test is a much broader question involving differential diagnoses, whereas here we are concerned with developing a parsimonious model to predict mortality, with the potential for consideration as an index of quality of care/hospital performance.

The data used in our study comes from one of the largest acute hospitals in the NHS. These data form integral components of patient care and clinical decision making in a modern NHS hospital and suggest that our approach is transferable to other hospitals. Generalisability is further enhanced by our use of the recently introduced NEWS [19] for the National Health Service. However, since we analysed data from patients who have both a full set of index routine blood tests results and an index EWS within ±24 hours of admission, our findings are only valid for such emergency admissions (about 70%). We necessarily excluded a considerable proportion (about 30%) of emergency admissions (typically with a median hospital stay of zero days) because of missing data. Whilst these admissions merit further study (eg some may be too sick to undergo such assessments whilst others may be too well), it seems reasonable to presume that in general, they are not representative of those patients who go through a pathway of care which accrues an index blood test and an EWS.

Mandating that these emergency admissions should also have an index blood test and EWS is difficult to justify because a sophisticated mortality risk assessment tool is likely to be of minimal use in these patients. We also excluded elective admissions from our analysis because mortality is a rare event in the elective setting, although future studies could incorporate elective admissions, subject to the availability of the necessary data items. In addition, diagnoses and comorbidity labels are absent in the Tree model because they are not necessarily known or recorded (electronically) with a high degree of belief within ±24 hours of admission. Likewise, it is worth emphasising that administrative data sets also, in general, omit blood test results and EWS. O’Sullivan et al found that including comorbidity did not improve the accuracy of a laboratory data model to predict mortality in emergency admissions [48] suggesting that the omission of comorbidity in our approach may not be a major limitation.

The validated Tree model reported here has two potential uses. Retrospective use could take the form of clinical audit and review of deceased patients in a specific terminal node (eg node 18) with a very low risk of death (to check for suboptimal care) or with a high risk of death (eg node 33, to review the quality of end-of-life care perhaps). The Tree model can also be used to prospectively estimate the risk of death for patients within ±24 hours. This is akin to the practice of using risk models for managing critical care patients [49]. This estimated risk of death can be communicated to carers, patients and relatives to influence and inform the subsequent process of care. Since blood tests and EWS are in widespread use in hospital medicine, our approach could be developed (and properly evaluated) in any hospital. However, although our approach could also be used to develop case-mix adjusted hospital mortality ratios, we are cautious about this use because of unresolved methodological issues [2], [8], [12], [13], [14], [50]. Finally, it is worth emphasizing that ultimately the usefulness of a risk prediction model, irrespective of any impressive desktop statistical properties, is determined by the extent to which it enhances the delivery of high quality of care and reduces avoidable mortality [51], [52]. Our approach offers a promising start because it builds on established current clinical practice and has the potential to be integrated into the clinical decision making process but needs further research.

Acknowledgments

Thanks are due to Dr P Barker, Dr R Kennedy and the UHCW hospital mortality committee for supporting the study. We are grateful to K O’Sullivan-Smith for help with data extraction at UHCW and The Learning Clinic for their assistance with data extraction. G Rudge is funded by the National Institute for Health Research (NIHR) through the Collaborations for Leadership in Applied Health Research and Care for Birmingham and Black Country (CLAHRC-BBC) programme. The views expressed in this publication are not necessarily those of the NIHR, the Department of Health, University of Birmingham or the CLAHRC-BBC Theme Management/Steering Group.

Data sharing: No further data is available.

Author Contributions

Conceived and designed the experiments: MAM GBS DRP. Performed the experiments: GR MAM DRP. Analyzed the data: MAM GR. Contributed reagents/materials/analysis tools: GR. Wrote the paper: MAM GBS DRP AG AS GW DW. Clinical perspective provided by: GBS GW DW.

References

  1. 1. Jarman B, Gault S, Alves B, Hider A, Dolan S, et al. (1999) Explaining differences in English hospital death rates using routinely collected data. BMJ 318(15): 15–20.
  2. 2. Shahian DM, Wolf R, Iezzoni L, Kirle L, Normand S-LT (2010) Variability in the measurement of hospital-wide mortality rates. N Eng J Med 363(26): 2530–2539.
  3. 3. Canadian Institute for Health Information. (2007) HSMR: a new approach for measuring hospital mortality trends in Canada. CIHI: Ottawa.
  4. 4. Heijink R, Koolman X, Pieter D, van der Veen A, Jarman B, et al.. (2008) Measuring and explaining mortality in Dutch hospitals; The Hospital Standardized Mortality Rate between 2003 and 2005. BMC Health Serv Res (8)73.
  5. 5. Köster M, Jürgensen U, Spetz C-L, Rutberg H (2008) »Standardiserad sjukhusdödlighet« som kvalitetsmått i hälsooch sjukvård. Läkartidningens 105(19): 1391–1396.
  6. 6. Miyata H, Hashimoto H, Horiguchi H, Matsuda S, Motomura N, et al. (2008) Performance of in-hospital mortality prediction models for acute hospitalization: Hospital Standardized Mortality Ratio in Japan. BMC Health Serv Res (8): 229 DOI:https://doi.org/ 10.1186/1472-6963-8-229.
  7. 7. Campbell MJ, Jacques RM, Fotheringham J, Maheswaran R, Nicholl J (2012) Developing a summary hospital mortality index: retrospective analysis in English hospitals over five years. BMJ (344): e1001.
  8. 8. Aylin P, Bottle A, Majeed A (2007) Use of administrative data or clinical databases as predictors of risk of death in hospital: comparison of models. BMJ. (334): 1044.
  9. 9. Mohammed MA, Stevens AJ (2007) The value of administrative databases. BMJ (334): 1014–5.
  10. 10. Hawkes N (2010) Patient coding and the ratings game. BMJ (340): 950–952.
  11. 11. Mohammed M, Stevens AJ, Lilford RJ (2009) Evidence of methodological bias in hospital standardised mortality ratios: retrospective database study of English hospitals. BMJ (338): b780.
  12. 12. Black N (2010) Assessing the quality of hospitals: Hospital standardised mortality ratios should be abandoned. BMJ (340): 933.
  13. 13. Lilford RJ, Pronovost P (2010) Using hospital mortality rates to judge hospital performance: a bad idea that just won’t go away. BMJ (340): 955–957.
  14. 14. National Institute for Health and Clinical Excellence. (2007) Acutely ill patients in hospital: recognition of and response to acute illness in adults in hospital. NICE clinical guideline No. 50. London.
  15. 15. Smith GB, Prytherch DR (2011) The Afferent Limb of Rapid Response Systems. In DeVita MA, Hillman K, Bellomo R. Textbook of Rapid Response Systems: Concept and Implementation, Springer.
  16. 16. Smith GB (2011) Measuring pulse and breathing rates - Simple, yet complex (editorial). Resuscitation (82): 1367–1368.
  17. 17. Mohammed MA, Hayton R, Clements G, Smith G, Prytherc D (2009) Improving accuracy and efficiency of early warning scores in acute care. Br J Nurs, 18, (1).
  18. 18. Breiman L, Friedman JH, Olshen RA, Stone CJ (1984) Classification and regression Trees. Monterey, CA: Wadsworth & Brooks/Cole Advanced Books & Software.
  19. 19. Steyerberg EW (2009) Clinical Prediction Models. A practical approach to development, validation and updating. Springer.
  20. 20. Harper PR (2005) A review and comparison of classification algorithms for medical decision making. Health Policy 71: 315–31.
  21. 21. Podgorelec V, Kokol P, Stiglic B, Rozman I (2002) Decision Trees: An Overview and Their Use in Medicine. J Med Syst 26(5): 455–63.
  22. 22. Smith GB, Prytherch DR, Schmidt P, Featherstone PI, Knight D, et al. (2006) Hospital-wide physiological surveillance – a new approach to the early identification and management of the sick patient. Resuscitation (71): 19–29.
  23. 23. National Early Warning Score (NEWS). (2012) Standardising the assessment of acute-illness severity in the NHS’. Report of a Royal College of Physicians working party, Royal College of Physicians.
  24. 24. University Hospitals Coventry and Warwickshire. Biochemistry & Immunology. Available: http://www.uhcw.nhs.uk/our-services/a-z-of-services?sID=1. Accessed March 2013.
  25. 25. Hothorn T, Hornik K, Zeileis A (2006) Unbiased Recursive Partitioning: A Conditional Inference Framework. J Comput Graph Stat 15(3): 651–674.
  26. 26. Fay MP (2010) Two-sided Exact Tests and Matching Confidence Intervals for Discrete Data. R Journal 2(1): 53–58.
  27. 27. R Development Core Team (2011) R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. Available: http://www.R-project.org.
  28. 28. Steyerberg EW (2008) Clinical Prediction Models. A practical approach to development, validation and updating. Springer.
  29. 29. Hanley JA, McNeil BJ (1982) The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143: 29–36.
  30. 30. Robin X, Turck N, Hainard A, Tiberti N, Lisacek F (2011) pROC: an open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics, 12, p. 77. DOI: 10.1186/1471-2105-12-77.
  31. 31. R Development Core Team (2011) R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. ISBN 3-900051-07-0. Available: http://www.R-project.org/.
  32. 32. Nicholl J (2007) Case-mix adjustment in non-randomised observational evaluations: the constant risk fallacy. J Epidemiol Community Health 61: 1010–3.
  33. 33. Lilford R, Mohammed MA, Spiegelhalter D, Thomson R (2004) Use and misuse of process and outcome data in managing performance of acute medical care: avoiding institutional stigma. Lancet (363): 1147–54.
  34. 34. Prytherch DR, Smith GB, Schmidt P, Featherstone PI, Knight D, et al. (2006) Calculating early warning scores – classroom comparison of pen and paper and hand-held computer methods. Resuscitation 70(2): 173–8.
  35. 35. Prytherch DR, Sirl JS, Weaver PC, Schimdt P, Higgins B, et al. (2003) Towards a national clinical minimum data set for general surgery. Br J Surg (90): 1300–5.
  36. 36. Prytherch DR, Sirl JS, Schmidt P, Featherstone PI, Weaver PC, et al. (2005) The use of routine laboratory data to predict in-hospital death in medical admissions. Resuscitation (66): 203–7.
  37. 37. Pine M, Jones B, Lou Y-B (1998) Laboratory values improve predictions of hospital mortality. Int J Qual Health Care (10): 491–501.
  38. 38. Vroonhof K, van Solinge WW, Rovers MM, Huisman A (2005) Differences in mortality on the basis of laboratory parameters in an unselected population at the Emergency Department. Clin Chem Lab Med (43): 536–41.
  39. 39. Hucker TR, Mitchell GP, Blake LD, Cheek E, Bewick V, et al. (2005) Identifying the sick: can biochemical measurements be used to aid decision making on presentation to the accident and emergency department. Br J Anaesth (94): 735–41.
  40. 40. Froom P, Shimoni S (2006) Prediction of Hospital Mortality Rates by Admission Laboratory Tests. Clinical Chemistry (52): 325–8.
  41. 41. Asadollahi K, Hastings M, Gill GV, Beeching NJ (2011) Prediction of hospital mortality from admission laboratory data and patient age: A simple model. Emerg Med Australas (23): 354–363.
  42. 42. Boekel E, Vroonhof K, Huisman A, van Kampen C, de Kieviet W, et al. (2006) Clinical laboratory findings associated with in-hospital mortality. Clin Chim Acta (372): 1–13.
  43. 43. Prytherch D, Smith GB, Schmidt PE, Featherstone PI (2010) ViEWS - towards a national Early Warning Score for detecting adult inpatient deterioration. Resuscitation (81): 932–937.
  44. 44. Duckitt RW, Buxton-Thomas R, Walker J, Cheek E, Beiwet V, et al. (2007) Worthing physiological scoring system: derivation and validation of a physiological early-warning system for medical admissions. An observational, population-based single-centre study. Br J Anaesth (98): 769–774.
  45. 45. Hodgetts TJ, Kenward G, Vlachonikolis IG, Payne S, Castle N (2002) The identification of risk factors for cardiac arrest and formulation of activation criteria to alert a medical emergency team. Resuscitation (54): 125–131.
  46. 46. Prytherch DR, Briggs JS, Weaver PC, Schmidt P, Smith GB (2005) Measuring clinical performance using routinely collected clinical data. Med Inform Internet Med (30): 151–156.
  47. 47. Silke B, Kellet J, Rooney T, Bennett K, O’Riordan D (2010) An improved medical admissions risk system using multivariable fractional polynomial logistic regression modelling. QJ Med (103): 23–32.
  48. 48. O’Sullivan E, Callely E, O’Riordan D, Bennett K, Silke B (2012) Predicting outcome in emergency medical admissions – role of laboratory data and co-morbidity. Acute Medicine 11(2): 59–65.
  49. 49. Harrison DA, Parry GJ, Carpenter JR, Short A, Rowan K (2007) A new risk prediction model for critical care: The Intensive Care National Audit & Research Centre (ICNARC) model. Critical Care Medicine (35): 1091–8.
  50. 50. Van den Bosch W, Spreeuwenberg P, Wagner C (2012) Variations in hospital standardised mortality ratios (HSMR) as a result of frequent readmissions. BMC Health Serv Res (12): 91.
  51. 51. Zegers M, de Bruijne MC, Wagner C, Hoonhout NH, Waaijman R, et al. (2009) Adverse events and potentially preventable deaths in Dutch hospitals: results of a retrospective patient record review study. Qual Saf Health Care (18): 297–302.
  52. 52. Hogan H, Healey F, Neale G, Thomson R, Vincent C, et al.. (2012) Preventable deaths due to problems in care in English acute hospitals: a retrospective case record review study. BMJ Qual Saf doi:10.1136/bmjqs-2012-001159.