Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Comparison of Epithor clinical national database and medico-administrative database to identify the influence of case-mix on the estimation of hospital outliers

  • Alain Bernard ,

    Roles Conceptualization, Methodology, Writing – review & editing

    alain.bernard@chu-dijon.fr

    Affiliation Department of Thoracic Surgery, Dijon University Hospital, Dijon, France

  • Pierre-Emmanuel Falcoz,

    Roles Validation, Writing – original draft, Writing – review & editing

    Affiliation Department of Thoracic Surgery, CHRU Strasbourg, Strasbourg, France

  • Pascal Antoine Thomas,

    Roles Validation

    Affiliation Department of Thoracic Surgery, Hopital-Nord-APHM, Aix-Marseille University, Marseille, France

  • Caroline Rivera,

    Roles Validation, Writing – original draft, Writing – review & editing

    Affiliation Department of Thoracic Surgery, Bayonne Hospital, Bayonne, France

  • Laurent Brouchet,

    Roles Validation

    Affiliation Department of Thoracic Surgery, Hopital Larrey, CHU Toulouse, Toulouse, France

  • Jean Marc Baste,

    Roles Validation, Writing – original draft

    Affiliation Department of Thoracic Surgery, CHU Rouen, Rouen, France

  • Marc Puyraveau,

    Roles Formal analysis, Methodology

    Affiliation Department of Biostatistics and Epidemiology CHU Besançon, Besançon, France

  • Catherine Quantin,

    Roles Formal analysis, Methodology

    Affiliations Department of Biostatistics and Medical Informatics, Dijon University Hospital, Dijon, France, INSERM, CIC 1432, Clinical Investigation Center, clinical epidemiology/clinical trials unit, Dijon University Hospital, University of Burgundy, Dijon, France

  • Pierre Benoit Pages,

    Roles Writing – original draft, Writing – review & editing

    Affiliations Department of Thoracic Surgery, Dijon University Hospital, Dijon, France, INSERM UMR 866, Dijon University Hospital, University of Burgundy, Dijon, France

  • Marcel Dahan

    Roles Validation

    Affiliation Department of Thoracic Surgery, Hopital Larrey, CHU Toulouse, Toulouse, France

Abstract

Background

The national Epithor database was initiated in 2003 in France. Fifteen years on, a quality assessment of the recorded data seemed necessary. This study examines the completeness of the data recorded in Epithor through a comparison with the French PMSI database, which is the national medico-administrative reference database. The aim of this study was to demonstrate the influence of data quality with respect to identifying 30-day mortality hospital outliers.

Methods

We used each hospital’s individual FINESS code to compare the number of pulmonary resections and deaths recorded in Epithor to the figures found in the PMSI. Centers were classified into either the good-quality data (GQD) group or the low-quality data (LQD) group. To demonstrate the influence of case-mix quality on the ranking of centers with low-quality data, we used 2 methods to estimate the standardized mortality rate (SMR). For the first (SMR1), the expected number of deaths per hospital was estimated with risk-adjustment models fitted with low-quality data. For the second (SMR2), the expected number of deaths per hospital was estimated with a linear predictor for the LQD group using the coefficients of a logistic regression model developed from the GQD group.

Results

Of the hospitals that use Epithor, 25 were classified in the GQD group and 75 in the LQD group. The 30-day mortality rate was 2.8% (n = 300) in the GQD group vs. 1.9% (n = 181) in the LQD group (P <0.0001). The between-hospital differences in SMR1 appeared substantial (interquartile range (IQR) 0–1.036), and they were even higher in SMR2 (IQR 0–1.19). SMR1 identified 7 hospitals as high-mortality outliers. SMR2 identified 4 hospitals as high-mortality outliers. Some hospitals went from non-outlier to high mortality and vice-versa. Kappa values were roughly 0.46 and indicated moderate agreement.

Conclusion

We found that most hospitals provided Epithor with high-quality data, but other hospitals needed to improve the quality of the information provided. Quality control is essential for this type of database and necessary for the unbiased adjustment of regression models.

Introduction

Epithor, a French national database for thoracic surgery, has been in operation since 2003. Participation is on a voluntary basis. It has led to the publication of several research articles [16]) and, most notably, Falcoz et al. [7] used it to develop the Thoracoscore, a predictive score that is widely used by European surgeons according to the latest European recommendations [8].A number of existing publications have highlighted the importance of data quality in medical databases [9,10], particularly since missing or biased data can lead to erroneous conclusions regarding hospital quality [11,12]. At this stage of development, it seems to us that the quality of the data within the Epithor database needs to be assessed. We therefore used the "Programme de Medicalisation des Systèmes d’information" (PMSI), a French national medico-administrative database that collects exhaustive data regarding pulmonary resection for lung cancer, as a point of reference [13]. Comparing the two databases will make it possible to identify the hospitals participating in Epithor that provide insufficient data concerning the lung cancer patients treated in their facility. The aim of this study was to rank hospitals according to the completeness of the data and to estimate the influence of data quality with respect to identifying 30-day mortality outliers.

Materials and methods

National medico-administrative database

All the data for patients who underwent pulmonary resection for lung cancer (LC) from January 2016 to December 2017 were collected from the French national medico-administrative database. The completeness and validity of PMSI data have already been assessed [13]. Routinely collected medical information includes the principal diagnosis, secondary diagnoses and the procedure performed on the patient. Diagnoses identified during the hospital stay are coded according to the International Classification of Diseases, tenth revision (ICD-10) [14]. We selected patients with a principal discharge diagnosis of primary lung cancer (codes C34, C34.0, C34.1, C34.2, C34.3, and C34.9). Procedures were coded according to the CCAM (Classification Commune des Actes Médicaux). For all patients, LC was confirmed by pathology analyses according to the 2004 World Health Organization classification of LC [15]. Surgery-related variables included the surgical approach (thoracotomy or video assisted thoracic surgery), and the type of resection (limited resection, lobectomy, bi-lobectomy or pneumonectomy). In this study, data access was based on special permissions given by the National Health Insurance. All data were completely anonymous. Patients consent is not required. Ethics approval, use of this database was approved by the National Commission for Data protection (CNIL No 1576793) and this study adhered to the tenets of the declaration of Helsinsksi.

The Epithor national clinical database

Epithor was modified in 2016 so that each surgeon could upload their patient data directly to a website called “web Epithor”. Between 1st January 2016 and 31st December 2017, 100 centers that operated patients for lung cancer provided data to Epithor. Baseline demographic data included sex, age, body mass index (BMI), performance status, American Society of Anesthesiologists (ASA) score, forced expiratory volume (FEV), dyspnea score, and smoking status. The comorbidities selected for our analysis were smoking status, chronic bronchitis, arrhythmia, chronic heart failure, peripheral artery disease, alcoholism, cirrhosis, stroke, diabetes, coagulopathy, hematologic disease, history of neoplasm, surgical history, severe malnutrition, pulmonary embolism, valvulopathy, neurological history, psychiatric history, asthma, respiratory failure, infectious disease, cardiac malformation, endocrine disorder, anemia, immunosuppression, and steroid treatment. The details of the surgery included surgical approach (open thoracotomy or video-assisted thoracoscopy) and type of surgery (segmentectomy, lobectomy, bilobectomy, or pneumonectomy). For clinical database, all data were completely anonymous. Patients consent is not required. Ethics approval, use of this database was approved by the National Commission for Data protection (CNIL No 809833) and this study adhered to the tenets of the declaration of Helsinki.

Outcome definition

In-hospital mortality was defined as the death of a patient within 30 days of the surgery or at a later time but during the same hospital stay.

Comparison of databases

The PMSI database was used to classify the hospitals that participated in data collection for the Epithor database. We used hospital FINESS codes to compare the number of pulmonary resections and related deaths recorded in Epithor to the figures recorded in the PMSI during the same period. We calculated the ratio of the number of lung resections in the PMSI to the number of lung resections in Epithor. For deaths, the calculation of the ratio was the same as for the number of pulmonary resections. The hospitals in the Epithor database were then divided into two groups according to the ratios obtained for pulmonary resections and deaths. Group 1was considered good-quality data (GQD) and included hospitals with ratios between 1 and 0.7 for both measures (75th percentile). Group 2 was defined as low-quality data (LQD) and included the hospitals with at least one of the two ratios below 0.7.

Risk-adjustment models

We applied a multiple imputation framework to compensate for missing FEV data. For the variable TNM stage, we created a category for all cases with missing TNM data. We developed a logistic regression model from the centers in the GQD group. Variables with a level of significance of ≤0.1 in univariate analyses were included in multivariate analyses by means of logistic regression. Continuous or ranked variables were tested to ensure conformity with the linear gradient by using the likelihood-ratio chi-squared test. Interaction effects were sought for all variables included in the model. All models were constructed using backward stepwise variable selection. A step-down variable selection using Akaike’s information criterion was used as a stopping rule. The area under the receiver operating characteristic (ROC) curve, and the R2 value were used to measure the discriminatory ability of the model. The reliability of the model was assessed with the Hosmer Lemeshow goodness-of-fit test [16].

Identification of quality outliers

The standardized mortality ratio (SMR) can be used as an indicator of quality for each hospital. SMR is defined as the ratio of O, number of deaths observed, to E, expected number of deaths per hospital estimated with the case-mix adjustment models. The estimated standard error of SMR is calculated with by Faris’ method [17]. A hospital whose SMR was significantly below 1 was considered a low mortality outlier, and when the SMR was significantly above 1 it was a high mortality outlier. We used the outlier detection method based on the test statistic [18].

To demonstrate the influence of case-mix quality on the classification of hospitals with low quality data (LQD group), we estimated SMR in two ways. For SMR1, the expected number (E1) of deaths per hospital was estimated with the risk-adjustment models fitted with low-quality data (LQD group). For SMR2, the expected number (E2) of deaths per hospital was estimated with the linear predictor for the low-quality data using the coefficients of a logistic regression model developed from the GQD group, as described by EW Steyerberg [16] (S1 Appendix). The effect of the different models on the overall between-hospital variation was quantified by calculating the interquartile ranges of SMR1 and SMR2. We used the kappa (k) statistic to assess the level of agreement between methods for outliers. The statistic measures the proportion of observed-to-expected agreement, and we adopted the principle that k>0.8 indicates excellent agreement, k = 0.6–0.8 indicates satisfactory agreement, k = 0.41–0.6 indicates moderate agreement, and k = 0.21–0.4 indicates low agreement [19].

Calculations were done with STATA 14 statistical software (StataCorp, College Station, Tex) and R statistical software for which we used Harrell’s Design library.

Results

Comparison of databases

Twenty-five hospitals participating in Epithor were included in the good-quality data (GQD) group and 75 were classified in the low-quality data (LGD) group (Figs 1 and 2) (S1 Table). Hospital characteristics are presented in Table 1.

thumbnail
Fig 1. Ratio of the number of procedures in the Epithor National clinical database in comparison with the Medico-administrative database (PMSI).

https://doi.org/10.1371/journal.pone.0219672.g001

thumbnail
Fig 2. Ratio of the number of in-hospital deaths in the Epithor National clinical database in comparison with the Medico-administrative database (PMSI).

https://doi.org/10.1371/journal.pone.0219672.g002

thumbnail
Table 1. Comparison hospital characteristics participating in Epithor with ratio of the number of pulmonary resections and the number of deaths greater than 0.7 vs hospital with ratio less than 0.7.

https://doi.org/10.1371/journal.pone.0219672.t001

Patient characteristics

The 30-day mortality rate was 2.8% (n = 300) in the GQD group vs. 1.9% (n = 181) in the LQD group (p <0.0001). The comorbidities are compared in Table 2. In the GQD group, significantly more patients had a performance status of 1 and the dyspnea score was higher on average (Table 2). The TNM stage was missing significantly more often in the LQD group (Table 2).

thumbnail
Table 2. Comparison of patient characteristics in hospitals with good-quality data and low-quality data (Epithor database).

https://doi.org/10.1371/journal.pone.0219672.t002

Risk-adjustment models

ASA scores were used as the linear variable. Performance status and dyspnea scores were categorized into 2 classes, and 3 classes were defined for BMI. Comorbidities were then selected in logistic regression using forward variable stepwise selection. The model was first developed from the GQD group and then applied to the LQD group (Tables 3 and S2). Variables such as performance status and dyspnea score had very different coefficients in the logistic model applied to the LQD group. The same can be said for comorbidities. The Hosmer-Lemeshow goodness-of-fit test was non-significant (Table 3). The C-statistic compared the discriminatory ability, which was found to be good for both models (HQD group 0.78 and LQD group 0.8) (Table 3).

thumbnail
Table 3. Logistic regression coefficients for the good-quality data group and low quality group for 30-day mortality (Epithor database).

https://doi.org/10.1371/journal.pone.0219672.t003

Effects of case-mix

The between-hospital differences were substantial in SMR1 (interquartile range (IQR) 0–1.036), and they were even higher in SMR2 (IQR 0–1.19). SMR1 and SMR2 for hospitals classified as LQD are compared in Table 4. Seven hospitals were identified as high-mortality outliers with SMR1. The methods used for SMR2 identified 4 hospitals as high-mortality outliers (Table 4). Some hospitals changed from non-outlier to high-mortality and vice-versa (Table 4). Kappa values were roughly 0.46 and indicated moderate agreement.

thumbnail
Table 4. Classification of hospitals with low-quality data according to standardized mortality rate (Epithor database).

https://doi.org/10.1371/journal.pone.0219672.t004

Discussion

This study demonstrates the limitations of a clinical database with voluntary participation. We found that comorbidities were generally under-coded in the low-quality data group of hospitals and that the underestimation of observed mortality influenced the construction of the risk-adjustment model. These elements can lead to hospital misclassification [2023]. However, the number of patients, specifically the number of patients who have a given event per hospital, have a much greater impact on accuracy [22]. The between-hospital difference in SMR was substantial (interquartile range) for both estimates. Our work shows the ability of the risk-adjustment model for identifying high-mortality outliers and low-mortality outliers with marginal agreement between the two methods.

Some work on the comparison of medico-administrative databases and clinical databases can be found in the literature; these studies emphasize that both types of database have their limits [2124]. The observed mortality rates vary between databases for the same types of surgery (25). Co-morbidities can also vary significantly in two databases that include the same type of surgery [25]. These differences, which can be the result of non-standardized end-points or the misclassification of cases [25], have a direct influence on the construction of risk-adjustment models for estimating the standardized mortality ratio. One published study compared the European and the North American general thoracic surgery databases [26]. The authors revealed considerable disparities in the rates of certain comorbidities, including coronary artery disease. This poses a problem in the interpretation of results seeing as this discrepancy can hardly be explained by patient characteristics. Two potential reasons could be put forward: firstly, that the European database under-codes certain comorbidities, or, secondly, that the databases do not use the same definitions. Further in the article, missing data for the TNM stage are reported. Missing data were more numerous in the European database, corroborating the theory that certain variables were under-reported [26].

The definition of an outcome such as mortality, which is the most commonly used indicator in surgery, can vary across databases. For some, mortality means hospital mortality, defined as any death occurring during the same hospital stay as the surgery, but mortality can also be 30-day mortality. The reported rate depends on the definition used, leading to discrepancies in the logistic model [20,22,27,28].

Our work is different from other publications comparing databases. Contrary to one study that used the clinical database as a reference to validate the medico-administrative database, we used the medico-administrative database to validate the completeness of the data from the centers participating in Epithor [25]. We used the PMSI, the exhaustive medico-administrative database, as the gold standard against which we compared the number of pulmonary resections and deaths recorded in Epithor. All private and public centers in France are required to use the PMSI database to finance their activity. Death is systematically reported as one of reasons for patient discharge from the hospital. In addition, it is possible to follow up on patients to see whether they died later on in another hospital.

This work raises the essential question of data quality[23,25,27,28,29]. We showed that the estimation of logistic model coefficients differs considerably with the quality of the data. The differences between logistic regression models have an effect on the determination of hospital outliers as demonstrated by previous studies using different methodologies [11,20,30]. We used two logistic models to estimate the expected number of deaths in the low-quality data group. The first model, estimated from the LQD data, was responsible for the over-fitting of coefficients. In the second model, using the method proposed by Steyerberg, we applied the coefficients from logistic regression estimated from the data provided by the GQD group [16]. We showed that this changes the estimate of the standard mortality rate (SMR) (appendix) and consequently the determination of the outliers. This difference in case-mix can influence the comparison of standardized mortality ratios as demonstrated Manktelow et al. [11]. In addition, it has been shown that the death rate and the volume of activity have a significant influence on whether the hospital is considered a high mortality or low mortality outlier [11,20,30]. The observed mortality rate in the LQD group was significantly lower than that of the GQD group; this is explained by an under-reporting of events (as shown in Fig 2 where these hospitals had a ratio significantly lower than 0.7).

The main limit of our study was the use of the FINESS code of each hospital to link the two databases [31]. This code was used to identify the hospital in the PMSI data and then manually match it with the Epithor database. This method was a reliable means to compare the number of lung resections and the number of deaths. In France it is very difficult to link two databases using patient identifiers because personal data is strictly protected by the CNIL [32]. Our assessment of the Epithor database is limited, but it remains enlightening regarding the need to apply enhanced measures for quality control. Recently, dashboards have been put into place so that each center can see its fill rate for different items compared to a national reference. Future on-site audits are planned, similar to what is currently in place for the STS database [33].

This study was essential for improving the quality of the Epithor database and it has been used to select the hospitals and surgeons providing the most complete data. These hospitals will serve as references for teams with lower quality data. In parallel, on-site audits are being implemented with the aim of continuously improving the quality of the Epithor database.

Conclusion

Epithor is an essential clinical database for measuring quality of care in France. There is a real interest in providing quality data, though some hospitals need to work on upgrading their participation. The systematic recording of data is an essential step in quality measurement, and it is necessary for the unbiased adjustment of regression models.

Supporting information

S1 Table. Data that was used to make Figs 1 and 2.

https://doi.org/10.1371/journal.pone.0219672.s002

(PDF)

S2 Table. Data used to perform the risk-adjustments models.

https://doi.org/10.1371/journal.pone.0219672.s003

(PDF)

Acknowledgments

We are grateful to all the French thoracic surgeons who participated in this study for their essential collaboration: Dr Michel ALAUZEN (Montpellier), Dr Jean-François ANDRO (Quimper), Dr Maxime AUBERT (Grenoble), Dr Jean Philippe AVARO (Marseille), Pr Jacques AZORIN (Bobigny), Dr Patrick BAGAN (Argenteuil), Dr François BELLENOT (Cergy Pontoise), Dr Vincent BLIN (Vannes), Dr Philippe BOITET (Harfleur), Dr Laurent BORDIGONI (Toulon), Pr Pierre-Yves BRICHON (Grenoble), Dr Gilles CARDOT (Boulogne-sur-Mer), Dr Jean Michel CARRIE (Saint Jean), Pr Pierre CORBI (Poitiers), Dr Michel DEBAERT (Lille), Dr Bertrand DEBRUERES (Ploemeur), Dr Jean DUBREZ (Bayonne), Dr Xavier DUCROCQ (Strasbourg), Dr Antoine DUJON (Bois Guillaume), Pr Pascal DUMONT (Tours), Dr Philippe FERNOUX (Chalon sur Saône), Pr Marc FILAIRE (Clermont-Ferrand), Dr Eric FRASSINETTI (Chambéry), Dr Gil FREY (Saint Etienne), Dr Dominique GOSSOT (Paris), Pr Gilles GROSDIDIER (Nancy), Dr Benoit GUIBERT (Lyon), Dr Olivier HAGRY (Chalon sur Saône), Dr Sophie JAILLARD (Lille), Dr Jean-Marc JARRY (Aix en Provence), Dr David KACZMAREK (Saint Etienne), Dr Yves LABORDE (Pau), Dr Bernard LENOT (Saint Brieuc), Dr Francis LEVY (Bordeaux), Dr Laurent LOMBART (Béziers), Dr Eric MARCADE (Saint Grégoire), Dr Jean Paul MARCADE (La Rochelle), Dr Jean MARZELLE (Créteil), Pr Gilbert MASSARD (Strasbourg), Dr Florence MAZERES (Bayonne), Dr Eric MENSIER (Lille), Dr David METOIS (Orléans), Dr JL MICHAUD/Edouard PARIS (Nantes), Dr Philippe MONDINE (Brest), Dr Michel MONTEAU (Reims), Dr Jean-Michel MOREAU (Nantes), Pr Jérôme MOUROUX (Nice), Dr Antoine MUGNIOT (Nantes), Dr Pierre MULSANT (Lyon), Dr Nidal NAFFAA (Avignon), Dr Pierre NEVEU (Talant), Dr Gérard PAVY (Arras), Pr Christophe PEILLON (Rouen), Pr François PONS (Percy), Pr Henri PORTE (Lille), Pr Jean-François REGNARD (Paris), Pr Marc RIQUET (Paris), Dr Babak SADEGHI LOOYEH (Morlaix), Pr Pascal THOMAS (Marseille), Pr Olivier TIFFET (Saint Etienne), Dr Bruno TREMBLAY (Meaux), Dr Jean VALLA (Charenton le Pont), Pr J JOUGON (Bordeaux), Dr Bernard WACK (Metz), Dr Jean-Didier WAGNER (Colmar), Dr Didier WOELFFE (Valenciennes).

The authors thank Suzanne Rankin who corrected the English.

References

  1. 1. Bernard A., Rivera C., Bt Pages P, Falcoz PE, Vicaut E, Dahan M Risk model of in-hospital mortality after pulmonary resection for cancer: a national database of the French Society of Thoracic and Cardiovascular Surgery (Epithor). J Thorac Cardiovasc Surg 2011; 141:449–458 pmid:20692003
  2. 2. Rivera C., Falcoz PE, Bernard A., Thomas P., Dahan M. Surgical management and outcomes of elderly patients with early stage of non-small cell lung cancer: a nested case-control study. Chest 2011; 140: 874–880 pmid:21436251
  3. 3. Rivera C., Bernard A., Falcoz PE, Thomas P, Schimdt A, Benard S, et al Characterization and prediction of prolonged air leak after pulmonary resection: a nation-wide study setting up the index of prolonged air leak. Ann Thorac Surg 2011; 92: 1062–1068 pmid:21871301
  4. 4. Thomas PA, Berbis J, Falcoz PE, Le Pimpec-Barthes F, Bernard A, Jougon J, et al. National perioperative outcomes of pulmonary lobectomy for cancer: the influence of nutritional status: Eur J CardioVasc Surg 2013; 1–8 pmid:24062351
  5. 5. Thomas PA, Falcoz PE, Bernard A, Le Pimpec-Barthes F, Jougon J, Brouchet L, et al for the EPITHOR group Bilobectomy for lung cancer: contemporary national early morbidity and mortality outcomes. Eur J CardioVasc Surg 2015: 1–6
  6. 6. Pagès PB, Delpy JP, Orsini B, Gossot D, Baste JM, Thomas PA, et al; Epithor project (French Society of Thoracic and Cardiovascular Surgery) Propensity Score Analysis Comparing Videothoracoscopic Lobectomy With Thoracotomy: A French Nationwide Study. Ann Thorac Surg. 2016 Feb 9. [Epub ahead of print]
  7. 7. Falcoz PE, Conti M, Brouchet L, Chocron S, Puyraveau M, Mercier M, et al. The Thoracic Surgery Scoring System (Thoracoscore): Risk model for in-hospital death in 15,183 patients requiring thoracic surgery. J Thorac Cardiovasc Surg 2007;133:325–32. pmid:17258556
  8. 8. Guidelines on the radical management of patients with lung cancer. Lim E, Baldwin D, Beckles M,Duffy J, Entwisle J, Faivre-Finn C, et al British Thoracic Society and the Society for Cardiothoracic Surgery in Great Britain and Ireland Thorax 2010;65(Suppl III):iii1eiii27. pmid:20940263
  9. 9. Walraven CV, Austin P. Administrative database research has unique characteristics that can risk biased results. J Clin Epidemiol 2012; 65: 126–131. pmid:22075111
  10. 10. Helgeland J, Kristoffersen DT, Skyrud KD, Lindman AS. Variation between hospitals with to diagnostic practice, coding accuracy, and case_mix. A retrospective validation study of administrative data versus medical records for estimating 30-day mortality after hip fracture. Plos ONE 2016; 11: e0156075 pmid:27203243
  11. 11. Manktelow BN, Evans TA, Draper ES. Differences in case-mix can influence the comparison of standardized mortality ratios even with optimal risk adjustment: an analysis of data from paediatric intensive care. BMJ Qual Saf 2014; 23: 782–788. pmid:24840239
  12. 12. Austin PC, Tu JV, Alter DA, Naylor CD. The impact of under coding of cardiac severity and comorbid diseases on the accuracy of hospital report cards. Med care 2005; 43: 801–809. pmid:16034294
  13. 13. Pagès PB, Cottenet J, Mariet AS, Bernard A, Quantin C In-hospital mortality following lung cancer resection: nationwide administrative database. Eur Respir J. 2016 Jun;47(6):1809–17. Epub 2016 Mar 10. pmid:26965293
  14. 14. Iezzoni LI. Assessing quality using administrative data. Ann Intern Med 1997; 127: 666–674. pmid:9382378
  15. 15. World Health Organization. International Statistical Classification of Diseases and Related Health Problems 10thRevision. http://apps.who.int/classifications/icd10/browse/2016/en Date last accessed: March 1, 2016. Date last updated:2016.
  16. 16. Steyerberg EW. Clinical prediction models. A practical approach to development, validation and updating. New York: springer; 2009
  17. 17. Faris PD, Ghali WA, Brant R. Bias in estimated of confidence intervals for health outcome report cards. Journal of epidemiology 2003; 56: 553–558
  18. 18. Shahian DM, Normand SL. What is a performance outlier? BMJ quality & safety 2015; 24: 95–99
  19. 19. Cohen J. A Coefficient of agreement for nominal scales. Educ Psychol Meas 1960; 20: 37–46
  20. 20. Fischer C, Lingsma HF, van Leersum N, Tollenaar RAEM, Wouters MW, Steyerberg EW. Comparing colon cancer outcomes: the impact of low hospital case volume and case-mix adjustment. EJSO 2015; 41: 1045–1053.
  21. 21. Welke KF, Diggs BS, Karamlou T, Ungerleider RM. Comparison of pediatric cardiac surgical mortality rates from national administrative data to contemporary clinical standards. Ann Thorac Surg 2009; 87: 216–223. pmid:19101301
  22. 22. Anema HA, van der Veer SN, Kievit J, Krol-Warmerdam E, Fischer C, Steyerberg EW, et al. Influences of definition ambiguity on hospital performance indicator scores: examples from the Netherlands. European Journal of public Health 2013; 24: 73–78 pmid:23543677
  23. 23. Salati M, Brunelli A, Dahan M, Rocco G, Van Raemdonck DEM, Varela G, on behalf of the European Society of thoracic surgeons database committee. Task-independent metrics to assess the data quality of medical registries using the European society of thoracic surgeons (ESTS) database. European journal of cardio-thoracic surgery 2011; 40: 91–98 pmid:21168341
  24. 24. Lawson EH, Louie R, Zingmond DS, Brook RH, Hall BL, Han L, et al. A comparison of clinical registry versus administrative claims data for reporting of 30-day surgical complications. Ann Surg 2012; 256: 973–981. pmid:23095667
  25. 25. Shahian DM, Silverstein T, Lovett AF, Wolf RE, Normand SLT. Comparison of clinical and administrative data sources for hospital coronary artery bypass graft surgery report cards. Circulation 2007; 115: 1518–1527. pmid:17353447
  26. 26. Seder CW, Salati M, Kozower BD, Wright CD, Falcoz PE, Brunelli A, et al. Variation in pulmonary resection practices between the society of thoracic surgeons and the European society of thoracic surgeons general thoracic surgery databases. Ann Thorac Surg 2016; 101: 2077–2084 pmid:27021033
  27. 27. Prasad A, Helder MR, Brown DA, Schaff HV. Understanding differences in administrative and audited patient data in cardiac surgery: comparison of the university healthsystem consortium and society of thoracic surgeons databases. J Am Coll Surg 2016; 223: 551–558. pmid:27457251
  28. 28. Welke KF, Peterson ED Vaughan-sarrazin MS, O’Brien SM, Rosenthal GE, Shook GJ, et al. Comparison of cardiac surgery volumes and mortality rates between the society of thoracic surgeons and medicare databases from 1993 through 2001. Ann Thorac Surg 2007; 84:1538–1547. pmid:17954059
  29. 29. Aylin P, Bottle A, Majeed A. Use of administrative data or clinical databases as predictors of risk of death in hospital: comparison of models. BMJ 2007; https://doi.org/10.1136/bmj.39168.496366.55
  30. 30. Kristoffersen DT, Helgeland J, Clench-Aas J, Laake P, Veiered MB. Observed to expected or logistic regression to indentify hospitals with high or low 30-day mortality? PLos ONE 2018; https://doi.org/10.1317/journal.pone0195248
  31. 31. FITNESS code. http://finess.sante.gouv.fr/fininter/jsp/rechercheAvancee.jsp
  32. 32. Commission National Informatique et Libertés (CNIL). https://www.cnil.fr/
  33. 33. Magee MJ, Wright CD, McDonald D, Fernandez FG, Kozower BD. External validation of the society of thoracic surgeons general thoracic surgery database. Ann Thorac Surg 2013; 96: 1734–1739 pmid:23998406