Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Identification of suicidal behavior among psychiatrically hospitalized adolescents using natural language processing and machine learning of electronic health records

  • Nicholas J. Carson ,

    Roles Conceptualization, Funding acquisition, Methodology, Project administration, Supervision, Writing – original draft, Writing – review & editing

    ncarson@cha.harvard.edu

    Affiliations Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America, Department of Psychiatry, Harvard Medical School, Boston, MA, United States of America

  • Brian Mullin,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Writing – original draft, Writing – review & editing

    Affiliation Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America

  • Maria Jose Sanchez,

    Roles Conceptualization, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    Affiliations Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America, Prevention and Community Health Department, Milken School of Public Health, George Washington University, Washington, D.C., United States of America

  • Frederick Lu,

    Roles Conceptualization, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America

  • Kelly Yang,

    Roles Conceptualization, Writing – review & editing

    Affiliations Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America, Department of Psychiatry, Albert Einstein College of Medicine, Bronx, NY, United States of America

  • Michelle Menezes,

    Roles Conceptualization, Writing – review & editing

    Affiliations Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America, University of Virginia, Charlottesville, VA, United States of America

  • Benjamin Lê Cook

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing

    Affiliations Health Equity Research Lab, Cambridge Health Alliance, Cambridge, MA, United States of America, Department of Psychiatry, Harvard Medical School, Boston, MA, United States of America

Abstract

Objective

The rapid proliferation of machine learning research using electronic health records to classify healthcare outcomes offers an opportunity to address the pressing public health problem of adolescent suicidal behavior. We describe the development and evaluation of a machine learning algorithm using natural language processing of electronic health records to identify suicidal behavior among psychiatrically hospitalized adolescents.

Methods

Adolescents hospitalized on a psychiatric inpatient unit in a community health system in the northeastern United States were surveyed for history of suicide attempt in the past 12 months. A total of 73 respondents had electronic health records available prior to the index psychiatric admission. Unstructured clinical notes were downloaded from the year preceding the index inpatient admission. Natural language processing identified phrases from the notes associated with the suicide attempt outcome. We enriched this group of phrases with a clinically focused list of terms representing known risk and protective factors for suicide attempt in adolescents. We then applied the random forest machine learning algorithm to develop a classification model. The model performance was evaluated using sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and accuracy.

Results

The final model had a sensitivity of 0.83, specificity of 0.22, AUC of 0.68, a PPV of 0.42, NPV of 0.67, and an accuracy of 0.47. The terms mostly highly associated with suicide attempt clustered around terms related to suicide, family members, psychiatric disorders, and psychotropic medications.

Conclusion

This analysis demonstrates modest success of a natural language processing and machine learning approach to identifying suicide attempt among a small sample of hospitalized adolescents in a psychiatric setting.

Introduction

In 2017, 17.2% of U.S. high school students reported having seriously considered attempting suicide and 7.4% reported having attempted suicide in the past year [1]. Rates of completed suicide among adolescents continue to increase [2]. Lifetime risk factors for suicide are well-established for adolescents who receive inpatient psychiatric treatment [35]. Prior suicide attempt is an especially potent risk factor for future suicide attempts [6], but some adolescents may not be willing to share such history with clinicians or family during the inpatient stay [79]. This reluctance to disclose might be due to the repercussions of disclosing to clinicians or family, including feelings of shame, loss of privacy or privileges, mistrust of healthcare providers, fears of further restrictions (e.g. a longer hospital stay), or under-estimating the severity of past behaviors [10]. In addition, adolescents hospitalized after acute suicidal behavior remain at high risk in the initial months following discharge [5, 1114].

Existing clinical tools for assessment of suicide risk or prior attempt can be time-intensive, costly, and might require clinician administration [1517]. Therefore, a computerized algorithm developed from clinical notes and integrated into a hospital’s electronic health record could function as an innovative and efficient complement to the judgment of the clinical team to classify whether a hospitalized patient, often whom the clinical team is meeting for the first time, has a history of suicide attempt. The National Action Alliance for Suicide Prevention’s Research Prioritization Task Force has identified the development of systems using healthcare data as a promising approach to the “retrospective examination of pathways leading to suicide events” [18].

Natural language processing (NLP) and machine learning (ML) have the potential to complement clinical practice by categorizing and analyzing data from clinical notes [19]. NLP is a computerized process that analyzes and codes human language into text that [20] ML algorithms can analyze and use to predict outcomes [21]. ML approaches have been used to predict suicidality in research using clinical notes (accuracy > 65%) [22], patient text messages (sensitivity = 0.56) [23], healthcare administrative data (AUC = 0.84) [24], and from structured data found in adolescent’s EHR [25]. Research applying NLP and machine learning to classification of suicidal behavior among psychiatrically hospitalized adolescents is limited but needed because such youth have high clinical severity, a greater frequency of past self-harm, and a higher propensity for future self-harm [3, 26, 27].

Using NLP to codify pre-admission electronic health record (EHR) notes to detect suicidal behavior is a promising approach [28]. EHR clinician notes are likely to capture important correlates of suicidal behavior in aggregate over time since mental health clinicians are trained in biopsychosocial mental health evaluation, including risk assessment [29]. Further, diagnostic codes for suicidal behavior are proving inadequate for identifying research cohorts of youth with suicidal behavior. A study using NLP to examine EHR notes of suicidal patients found that only a small proportion of patients with suicidal ideation or attempt documented in EHR notes had a corresponding diagnostic code recorded in the EHR [30]. A study of adolescents with autism spectrum disorder developed an NLP classification tool with a recall of 0.91 for detection of suicidality in EHR notes [31]. However, such studies may fail to capture suicide attempts that are not overtly mentioned in the medical record. This approach has yet to be applied to studying suicide among psychiatrically hospitalized adolescents and may prove valuable, as adolescents have been found to be more likely to report suicide attempts under conditions of anonymity [32].

An algorithm based on unstructured notes may capture a broader spectrum of patients than one using structured diagnostic codes, which may not include all patients with a history of suicide attempt. EHR notes are also relatively frequent in this population (e.g. weekly psychotherapy and monthly medication visits, in addition to primary care, emergency, and inpatient treatment, and documentation of collateral contacts), which is important for detecting low prevalence outcomes like suicide attempt. Finally, NLP of clinical notes can allow for detection of novel variables that are specific to the particular health system under study.

To address this gap in the research literature, we describe the development of a machine learning algorithm that generates classification models from codes developed by NLP analysis of EHR notes in order to categorize adolescents by history of suicide attempt. We describe in detail the process and method used to capture codes from the clinical notes and the development and refinement of the machine learning algorithm. The algorithm yielded a modest classification capability, and a method for developing and refining similar algorithms in other psychiatric settings.

Methods

Sample

The IRB of the Cambridge Health Alliance approved this study (CHA-IRB-0886/01/12). Adolescents gave in person assent to participate in the survey and their legal guardian provided in person or telephone (audio-recorded) consent to participate and for analysis of electronic health record data. Participants aged 12 to 20 years old were recruited from an inpatient psychiatric unit of a community hospital in Massachusetts from February 2012 to September 2016. They were invited to complete a confidential self-report survey that assessed mental health and risk-taking behaviors, including history of suicidal thoughts and attempts. This was a research survey and neither clinicians nor parents were privy to the patients’ responses. Parents or guardians provided consent either in-person or by phone (audio recorded), and the adolescent patients provided in-person assent. Human subjects approval to survey patients and to analyze notes from their electronic health records was obtained from the institutional review board of the health system. The survey sample included a total of 241 respondents.

Sample dataset

Of the total survey sample, 73 youth had at least one EHR documentation available for treatment visits in the year prior to the index psychiatric admission in the same health system. Patients in this sample are described in Table 1. EHR documentation included outpatient, inpatient, or emergency room clinical encounters across mental health and primary care. A total of 9415 notes were identified, ranging from 1 to 876 notes per patient, with a mean of 129 and median of 70 notes prior to admission. These clinical notes generally included clinician progress notes and clinician documentation of contacts with family members, other members of the treatment team, and relevant systems (e.g. schools, community service agencies). Notes were written by a variety of providers, including physicians, psychologists, nurses, case managers, social workers, and medical assistants.

thumbnail
Table 1. Sociodemographic and clinical descriptors of adolescent sample.

https://doi.org/10.1371/journal.pone.0211116.t001

Outcome variable of interest

The outcome variable of interest was any past year suicide attempt captured by an item from the Youth Risk Behavior Survey: “During the past 12 months, how many times did you actually attempt suicide?” The research literature supports the use of a self-report suicide history variable in inpatient settings. Adolescents are much more likely to self-report suicide attempt under conditions of anonymity [32], which our survey was. A survey of outpatient and inpatient adolescents in the UK showed that 20% reported at least one episode of self-harm on the questionnaire that was not recorded in the clinical record [33]. A study among adults showed 83% agreement between self-report of self-harm and therapist notes and, further, all medically treated episodes were reported by participants [34]. On the other hand, a study of adolescents found under-reporting of “self-harm” on a questionnaire when compared to hospital admission data [35]. Based on this evidence, we determined that self-report was an adequate gold standard variable to use for this study, and further chart review to augment self-report was not conducted.

Responses were dichotomized as zero or at least once. A total of 27 (37%) survey respondents reported at least one suicide attempt. Differences in demographic and clinical characteristics between participants with attempt and no attempt were compared using chi-square and t tests.

Natural language processing of electronic health records

A dataset was created of all clinical notes for survey participants with EHR documentation for one year prior to the index admission (where the survey was completed). The NLP analysis used Invenio software [36] to codify the unstructured text of EHR records during that time period. Invenio is based on the open source Apache cTAKES system [37] and analyzes the unstructured free text of clinical notes to generate Concept Unique Identifiers (CUIs), which are alphanumeric codes representing specific items in the Unified Medical Language System (UMLS) [38]. For example, the CUI “C0424000” represents “Feeling suicidal (finding)”. Similar to the cTAKES platform, Invenio uses features such as a sentence boundary detector, tokenizer, normalizer, part-of-speech tagger, shallow parser, and named entity recognition annotator to convert free text into UMLS CUIs. Invenio also captures negation in the context of EHR sentences, such as “no suicide attempts”. A total of 11806 CUIs, including their negations, were extracted from the clinical notes of the total sample and used as data for the machine learning algorithm. This list of CUIs served as the library of eligible CUIs for the five-fold cross validation conducted for the algorithm development, described below.

To enrich the classification power of the CUIs identified through NLP, a “curated” list of 34 suicide-related predictive factors and 30 protective factors was developed by behavioral health clinicians on the research team, drawing from the literature on risk factors for adolescent suicide [3942] (S1 Table). This curated list was transformed into CUIs by matching the curated terms to text-strings from EHR notes of the total sample.

Machine learning and algorithm development

We used a random forest algorithm, an ensemble classification method with validated applications in mental health research [14, 25, 43], to classify each patient by history of past-year suicide attempt [44]. Random forest sits in the middle of the so-called “axes of machine learning,” relying on human selection and curation of variables that are analyzed using hundreds of decision trees to identify nonlinear variable interactions [45]. The analysis was performed in the statistical software R (version 3.4.4) using the package randomForest [46]. Random forest classification develops decision trees by creating nodes that are related or not to suicidality. For example, the word “depression” (main node) was split between presence or absence of whether the word depression appeared in the note [47].

We then performed five-fold cross-validation of the training set [48] in order to optimize the features (e.g., the set of CUIs) used in the final model. To do so, we first randomly partitioned the data into a training set (80% of patients, n = 58) and an out-of-sample validation set (20% of patients, n = 15). The first step in the cross-validation process was to randomly divide the training set into 5 mutually exclusive datasets, or folds. Two folds contained data from 11 patients while the remaining three folds contained data from 12 patients. During cross-validation, each fold served as the testing dataset for the model that was trained on the remaining four datasets. Thus, each patient in the 80% training set was assigned to a test dataset once (a probability of 20%), and to a training dataset four times.

Second, CUIs were extracted from the clinical notes of the patients from the full dataset. Invenio wrote each CUI contained in each note to a log, which was read by SAS version 9.4 (SAS Institute Inc. Cary, NC, USA), and a dummy variable was generated for each CUI. A matrix of 11806 dummy variables (CUIs) by 9415 notes (records) was generated. To reduce overfitting, we required that a CUI appear at least once for four different patients before including it in the random forests algorithm, which trimmed the matrix to approximately 4000 dummy variables. For the cross-validation and the model performance evaluation (conducted on the 20% out-of-sample dataset), models were trained only on CUIs that originated from notes in each fold’s training dataset.

Third, the 50 CUIs with the largest mean decrease in the Gini impurity index for each fold (S2 Table) were identified using random forest and combined with CUIs from the “curated” list that appeared in the training data for that particular fold. The Gini impurity index is a decision tree split quality measure used in machine learning corresponding to the mean decrease in impurity caused by a node [49]. It is one method of feature importance available in the random forest approach. The average number of CUIs used in this iterative step was 160.

Fourth, a model was built by generating 300 decision trees using the random forest procedure. Fifth, the resulting model was applied to each note of each individual randomized to the test fold to assess performance statistics for that fold. Finally, if the number of notes classifying suicide for an individual person was greater than a specified cutoff (e.g. 0%, 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80%, 90%, or 100%), then the individual was classified as having attempted suicide in the past year. This process was repeated 5 times, with each subset of the training dataset (that is, 16% of the total sample) serving as a holdout dataset once. Statistics across the five folds were then averaged to select the cut-off (the proportion of notes indicating suicide attempt) that determined the best model.

Evaluation of algorithm performance

A final model was built on the training dataset (80% of patients, n = 58), using the features from the best model determined by five-fold cross-validation (in this case, a model that used a 20% cutoff for number of notes classified as suicide attempt). The validity of the model was evaluated by performing validation on the out-of-sample dataset (20% of the total sample, n = 15 patients), determining the sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and accuracy in the comparison of individuals classified as suicide attempt by the algorithm and those who reported a suicide attempt in the gold standard survey measure obtained during the inpatient admission. We also assessed the receiver operating curve (ROC) and estimated the area under the curve (AUC) [14] for classifications that varied by the percentage of notes (0%, 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80%, 90%, 100%) indicating suicide attempt.

Results

A total of 27 patients reported at least one suicide attempt in the year prior to admission. This group was predominantly female, white or Hispanic, and had public insurance (Table 1). Comparisons of demographic and clinical information showed no significant differences between those with and without self-reported suicide attempt (p>0.10).

The random forest procedure identified EHR phrases that, when converted to CUIs in the UMLS, were significantly associated with suicide attempt. Examples of these EHR phrases are found in the S2 Table, which lists the top fifty phrases identified in each of the five training folds. There was a notable preponderance of EHR phrases representing known adolescent suicide risk factors. These included phrases related to suicidal behaviors as previously described in the literature, such as risk factors for suicidality (“suicide attempts,” “thoughts of suicide,” “mood; Depressed”; “pain”), family [5052] (e.g. “fathers,” “brothers,” “grandfather,” “parents,” “siblings”), medication (antidepressants, antipsychotics, anti-inflammatories), and mental health conditions (“severe depression”, “PTSD,” “unspecified psychosis,” disorder, recurrent bipolar,” substance use/abuse,” “attention deficit hyperactivity disorder,”).

Regarding classification statistics for the machine learning model (Table 2), the optimal model we selected in prioritizing sensitivity without overly sacrificing specificity was a model that used a 20% cutoff for number of notes classified as suicide attempt. The performance across all cutoffs and folds of cross-validation can be viewed in S3 Table. The mean model performance at this cutoff included a sensitivity of 0.72 (range 0.25–1.00), a specificity of 0.26 (0.13–0.50), a PPV of 0.33 (0.13–0.75), an NPV of 0.63 (0.25–1.00). At a cutoff of 20%, the system had an accuracy of 42% (range 0.17–0.58), as compared to the most frequent class baseline of 63%.

thumbnail
Table 2. Mean model performance by cutoff (0–100%) Across 5 fold cross-validation.

https://doi.org/10.1371/journal.pone.0211116.t002

Using the model trained off the 80% dataset, and applying the 20% cutoff identified as optimal in the cross validation, we then performed external validation on the 20% out-of-sample dataset, reporting the same performance metrics. The final model had a sensitivity of 0.83, a specificity of 0.22, a PPV of 0.42, an NPV of 0.67, an accuracy of 0.47%, and an AUC of 0.68 (Fig 1). In sensitivity analyses, model performance was measured after decreasing or increasing the percentage of notes indicating suicide attempt. These performance metrics are presented in Table 3, with each row representing a different “cutoff” for the minimum proportion of EHR notes required to indicate suicide attempt.

Discussion

In this proof of concept study, we present a method for developing a classification model for past-year suicide attempt among psychiatrically hospitalized adolescents using natural language processing and machine learning of clinical narratives from electronic health record data preceding admission. From a sample of 73 patients admitted to a psychiatric unit, we developed a classification algorithm with moderate sensitivity and negative predictive value, a modest AUC, and an accuracy below the most frequent class baseline. To the best of our knowledge, this is the first time unstructured data has been analyzed using NLP to develop a machine learning classification algorithm in an adolescent inpatient population. This initial signal of success using a small sample should be validated in larger datasets. The approach demonstrates how EHR notes, enriched with clinically relevant information for adolescent suicide attempt, can be used to identify youth with histories of suicidal behavior, which can aid inpatient treatment planning during a particularly vulnerable time for this high risk population.

Machine learning-based algorithms can lead to methodologically sound but clinically inadequate results due to the lack of clinical context included in the model [53, 54]. This limitation may reveal the benefit of including data derived from natural language processing of clinical notes. EHR notes, especially those from mental health settings, contain a richness of description and context that, when organized for higher level analysis, may identify important relationships with adverse health care outcomes. The model we describe in this paper is therefore notable for its use of NLP-derived phrases for use in the random forest procedure, as opposed to only structured data (e.g. from drop-down and forced selection EHR fields). For example, the algorithm described in this paper yielded a high number of EHR phrases associated with family, which may demonstrate the importance of familial support and/or conflict in understanding suicidal behavior among adolescents with severe mental health difficulties. This relationship between family and suicide attempt is supported by prior machine learning research [50] as well as studies using other research methods [51, 52, 5557]. Future studies can then explore the direction of these relationships to guide clinicians towards the most relevant family-related history to assess, document, and address in treatment planning. Thus the value in using NLP to a health care institution may be that the resulting classification model is attuned to the practice behaviors and patient characteristics of that particular setting.

It is important to note that the performance of the algorithm reported here varied by the threshold of notes that were indicating suicide attempt. Optimizing for sensitivity, we selected a cutoff of 20% of positive notes. To our knowledge, this is the first paper to report these cutoffs and how the classification statistics vary on this parameter. Users of such algorithms can let the clinical need of the tool guide the optimal cut-off in their setting and the related trade-offs between sensitivity and specificity.

Comparing the test metrics of our algorithm to prior research, we note first that there is no common practice of presenting algorithm performance in the literature, although useful guidelines have been published to help standardize the field [58]. Some report performance statistics and probabilities, while others present area under the curve [14]. A study using a self-report measure to determine suicide attempt history achieved sensitivities ranging from 55.8–72.1% and AUC’s ranging from 0.65–0.77 [59]. This study did not use EHR data, but rather measures of sociodemographic and clinical variables (e.g. a depression rating scale). Thus, although the studies are similar, the differences in methodology make it difficult to compare the capabilities of the models.

In the prediction literature, a recent study that also used random forest took care to stratify predictions across a two year period, showing increasingly accurate predictions more proximal to the index attempt (AUCs > 0.83) [25]. Similar to our NLP findings, this paper also found psychotropic medications to be important predictors of suicide attempt.

One potential clinical application of such an algorithm would be to serve as an alert in an EHR system to complement existing risk scales and to aid in clinical risk assessment. Such alerts have been shown to improve recognition of other chronic health conditions, such as pediatric hypertension, although recognition differed by race/ethnicity and gender [60]. A small trial among suicidal adolescents showed a significant increase in safety planning, but with “moderate” satisfaction reported by clinicians [60]. Further research is needed to safely incorporate the signals of automated algorithms into the regular workflows of inpatient treatment planning, in a way that is meaningful to clinicians and families [61].

Limitations

The suicide attempt outcome used in this study allows for the possibility that the algorithm is in fact predicting a suicide attempt. For example, in the event that an individual in the dataset was hospitalized immediately following a suicide attempt, the algorithm using notes in the year prior to that attempt and subsequent hospitalization would, in this case, be predicting an attempt in the future. The specific dates of suicide attempt are unavailable in the survey and so we are unable to determine whether the algorithm is classifying existing events or predicting future events. In future research, the algorithm could be easily re-oriented completely towards prediction by entering only data from EHR notes that come definitively before a suicide outcome with known dates (e.g. an ICD10 code for a suicide attempt documented in an emergency department visit).

Our approach was limited by a small dataset from one community health system, which limits generalizability of findings. The use of natural language processing to define the variables eligible for the random forest procedure limits generalizability further, since they reflect the particular patient, clinician, and treatment settings characteristics within the community health system [62]. Analysis of free text in clinical notes presents both challenges and opportunities due to differences in provider documentation. In comparison, the use of standardized terminology in radiology and cancer staging have resulted in better model performance [63].

The accuracy of classification models using NLP may be improved by including complementary structured variables in the dataset (such as diagnoses, medications, outcome measures, and social determinants) to augment the detection of suicide attempt [25, 64]. For example, suicide outcomes from the electronic record, such as ICD-10 codes for suicide and suicide attempt, provide valid evidence of treatment for self-harm, are less prone to patient response bias, and are therefore suitable for future research in this area. However, such codes would not capture attempts that went untreated, and prior research shows that such codes are under-utilized by clinicians [30], perhaps because they do not indicate a billable diagnosis (e.g. Major Depressive Disorder). By using a self-report variable as the gold-standard outcome, we avoided the under-reporting bias known to affect ICD-9 coding of self-injury [65]. The use of a brief screener for suicide attempt with hospitalized adolescents may also have been vulnerable to a higher false positive rate than longer, structured measures [66]. We look forward to validating this work using such outcomes, in larger samples, and across multiple treatment settings. A validated prediction algorithm can then be studied clinically to determine safe implementation practices from the perspectives of clinicians and the families they serve.

Supporting information

S1 Table. List of “curated” concepts related to adolescent suicide attempt.

https://doi.org/10.1371/journal.pone.0211116.s001

(XLSX)

S2 Table. Top fifty concept unique identifiers from each validation fold.

CUI, Concept Unique Identifiers; UMLS, Unified Medical Language System; Gini impurity index, a measure used in machine learning corresponding to the mean decrease in impurity caused by a node; EHR: electronic health records. UMLS CUI: alphanumerical code designated by UMLS to name each CUI. Concept: Description of CUI, obtained from 2018AA release of the UMLS Metathesaurus Browser. EHR text phrase: text phrase found in electronic health record matching the CUIs.

https://doi.org/10.1371/journal.pone.0211116.s002

(XLSX)

S3 Table. Model performance by cutoff and fold.

https://doi.org/10.1371/journal.pone.0211116.s003

(XLSX)

S1 File. Example data.

Zipped folder containing SAS data files that were created by the research team. These files were not constructed from patient data and are intended for demonstration of our methods. Files include CUI logs, 5 folds of example training data, a file containing data from a sample of “patients” who self-reported past year suicide attempt, a file containing data from a sample of “patients” who did not report a past year suicide attempt, the SAS code used for the analysis, and a modified version of %PROC_R (Wei, X. (2012). %PROC_R: A SAS Macro that Enables Native R Programming in the Base SAS Environment. Journal of Statistical Software, 46(Code Snippet 2), 1–13.), that calls R from SAS.

https://doi.org/10.1371/journal.pone.0211116.s004

(ZIP)

Acknowledgments

The authors would like to thank Wired Informatics for providing access to their Invenio NLP software and technical guidance regarding the software. Statistical support was provided by data science specialist Steven Worthington at the Institute for Quantitative Social Science, Harvard University.

References

  1. 1. Youth Risk Behavior Survey—Data Summary & Trends Report: 2007–2017. Centers for Disease Control, 2018.
  2. 2. QuickStats: Suicide Rates*((,†)) for Teens Aged 15–19 Years, by Sex—United States, 1975–2015. MMWR Morbidity and Mortality Weekly Report: Centers for Disease Control and Prevention; 2017. p. 816.
  3. 3. D'Eramo KS, Prinstein MJ, Freeman J, Grapentine WL, Spirito A. Psychiatric Diagnoses and Comorbidity in Relation to Suicidal Behavior Among Psychiatrically Hospitalized Adolescents. Child Psychiatry and Human Development. 2004;35(1):21–35. pmid:15626323
  4. 4. Becker DF, Grilo CM. Prediction of Suicidality and Violence in Hospitalized Adolescents: Comparisons by Sex. The Canadian Journal of Psychiatry. 2007;52(9):572–80. pmid:17953161.
  5. 5. Wolff JC, Davis S, Liu RT, Cha CB, Cheek SM, Nestor BA, et al. Trajectories of Suicidal Ideation among Adolescents Following Psychiatric Hospitalization. Journal of Abnormal Child Psychology. 2017:1–9. pmid:28349306
  6. 6. Brent DA, Baugher M, Bridge J, Chen T, Chiappetta L. Age-and sex-related risk factors for adolescent suicide. Journal of the American Academy of Child & Adolescent Psychiatry. 1999;38(12):1497–505.
  7. 7. King CA, Kramer A, Preuss L, Kerr DC, Weisse L, Venkataraman S. Youth-Nominated Support Team for suicidal adolescents (Version 1): A randomized controlled trial. Journal of consulting and clinical psychology. 2006;74(1):199. pmid:16551158
  8. 8. Busch KA, Fawcett J, Jacobs DG. Clinical correlates of inpatient suicide. The Journal of clinical psychiatry. 2003;64:14–9.
  9. 9. Qin P, Nordentoft M. Suicide risk in relation to psychiatric hospitalization: evidence based on longitudinal registers. Archives of general psychiatry. 2005;62(4):427–32. pmid:15809410
  10. 10. Wisdom JP, Clarke GN, Green CA. What Teens Want: Barriers to Seeking Care for Depression. Administration and policy in mental health. 2006;33(2):133–45. PMC3551284. pmid:16489480
  11. 11. Hunt IM, Kapur N, Webb R, Robinson J, Burns J, Shaw J, et al. Suicide in recently discharged psychiatric patients: a case-control study. Psychological Medicine. 2008;39(3):443–9. Epub 05/28. pmid:18507877
  12. 12. Yen S, Weinstock LM, Andover MS, Sheets ES, Selby EA, Spirito A. Prospective predictors of adolescent suicidality: 6-month post-hospitalization follow-up. Psychological Medicine. 2012;43(5):983–93. Epub 08/30. pmid:22932393
  13. 13. Czyz EK, King CA. Longitudinal Trajectories of Suicidal Ideation and Subsequent Suicide Attempts Among Adolescent Inpatients. Journal of Clinical Child & Adolescent Psychology. 2015;44(1):181–93. pmid:24079705
  14. 14. Walsh CG, Ribeiro JD, Franklin JC. Predicting risk of suicide attempts over time through machine learning. Clinical Psychological Science. 2017;5(3):457–69.
  15. 15. Huth-Bocks AC, Kerr DC, Ivey AZ, Kramer AC, King CA. Assessment of psychiatrically hospitalized suicidal adolescents: self-report instruments as predictors of suicidal thoughts and behavior. Journal of the American Academy of Child & Adolescent Psychiatry. 2007;46(3):387–95.
  16. 16. King CA, Jiang Q, Czyz EK, Kerr DC. Suicidal ideation of psychiatrically hospitalized adolescents has one-year predictive validity for suicide attempts in girls only. Journal of abnormal child psychology. 2014;42(3):467–77. pmid:23996157
  17. 17. Babeva K, Hughes JL, Asarnow J. Emergency Department Screening for Suicide and Mental Health Risk. Current Psychiatry Reports. 2016;18(11):100. pmid:27671917
  18. 18. Colpe LJ, Pringle BA. Data for Building a National Suicide Prevention Strategy. American Journal of Preventive Medicine. 2014;47(3):S130–S6. pmid:25145730
  19. 19. Longhurst C, Harrington , Robert , Shah , Nigam . A ‘Green Button’ For Using Aggregate Patient Data At The Point Of Care. Health Affairs. 2014;33(7):1229–35. pmid:25006150
  20. 20. Lacson R, Khorasani R. Natural Language Processing: The Basics (Part 1). Journal of the American College of Radiology. 2011;8(6):436–7. pmid:21636059
  21. 21. Monuteaux MC, Stamoulis C. Machine Learning: A Primer for Child Psychiatrists. Journal of the American Academy of Child & Adolescent Psychiatry. 55(10):835–6. pmid:27663936
  22. 22. Poulin C, Shiner B, Thompson P, Vepstas L, Young-Xu Y, Goertzel B, et al. Predicting the Risk of Suicide by Analyzing the Text of Clinical Notes. PLoS ONE. 2014;9(1):e85733. pmid:24489669
  23. 23. Cook BL, Progovac AM, Chen P, Mullin B, Hou S, Baca-Garcia E. Novel Use of Natural Language Processing (NLP) to Predict Suicidal Ideation and Psychiatric Symptoms in a Text-Based Mental Health Intervention in Madrid. Comput Math Methods Med. 2016;2016:8708434. pmid:27752278
  24. 24. Kessler RC, Warner CH, Ivany C, et al. Predicting suicides after psychiatric hospitalization in us army soldiers: The Army Study to Assess Risk and Resilience in Servicemembers (Army STARRS). JAMA Psychiatry. 2015;72(1):49–57. pmid:25390793
  25. 25. Walsh CG, Ribeiro JD, Franklin JC. Predicting suicide attempts in adolescents with longitudinal clinical data and machine learning. Journal of child psychology and psychiatry, and allied disciplines. 2018. Epub 2018/05/01. pmid:29709069.
  26. 26. Goldston DB, Daniel SS, Reboussin DM, Reboussin BA, Frazier PH, Kelley AE. Suicide Attempts Among Formerly Hospitalized Adolescents: A Prospective Naturalistic Study of Risk During the First 5 Years After Discharge. Journal of the American Academy of Child & Adolescent Psychiatry. 1999;38(6):660–71. pmid:10361783
  27. 27. Goldston DB, Daniel SS, Erkanli A, Reboussin BA, Mayfield A, Frazier PH, et al. Psychiatric Diagnoses as Contemporaneous Risk Factors for Suicide Attempts Among Adolescents and Young Adults: Developmental Changes. Journal of consulting and clinical psychology. 2009;77(2):281. PMC2819300. pmid:19309187
  28. 28. Fernandes AC, Dutta R, Velupillai S, Sanyal J, Stewart R, Chandran D. Identifying Suicide Ideation and Suicidal Attempts in a Psychiatric Clinical Research Database using Natural Language Processing. Scientific Reports. 2018;8(1):7426. pmid:29743531
  29. 29. Dulcan's Textbook of Child and Adolescent Psychiatry. Second Edition ed. Arlington, VA: American Psychiatric Association; 2016.
  30. 30. Anderson HD, Pace WD, Brandt E, Nielsen RD, Allen RR, Libby AM, et al. Monitoring Suicidal Patients in Primary Care Using Electronic Health Records. The Journal of the American Board of Family Medicine. 2015;28(1):65–71. pmid:25567824
  31. 31. Downs J, Velupillai S, George G, Holden R, Kikoler M, Dean H, et al. Detection of Suicidality in Adolescents with Autism Spectrum Disorders: Developing a Natural Language Processing Approach for Use in Electronic Health Records. AMIA Annual Symposium Proceedings. 2017;2017:641–9. PMC5977628. pmid:29854129
  32. 32. Safer DJ. Self-Reported Suicide Attempts by Adolescents. Annals of Clinical Psychiatry. 1997;9(4):263–9. pmid:9511952
  33. 33. Ougrin D, Boege I. Brief report: The self harm questionnaire: A new tool designed to improve identification of self harm in adolescents. Journal of Adolescence. 2013;36(1):221–5. https://doi.org/10.1016/j.adolescence.2012.09.006. pmid:23046733
  34. 34. Linehan MM, Comtois KA, Brown MZ, Heard HL, Wagner A. Suicide Attempt Self-Injury Interview (SASII): Development, reliability, and validity of a scale to assess suicide attempts and intentional self-injury. Psychological Assessment. 2006;18(3):303–12. pmid:16953733
  35. 35. Mars B, Cornish R, Heron J, Boyd A, Crane C, Hawton K, et al. Using Data Linkage to Investigate Inconsistent Reporting of Self-Harm and Questionnaire Non-Response. Archives of Suicide Research. 2016;20(2):113–41. PMC4841016. pmid:26789257
  36. 36. Wired Informatics. Invenio. Boston, MA2017.
  37. 37. Savova GK, Masanz JJ, Ogren PV, Zheng J, Sohn S, Kipper-Schuler KC, et al. Mayo clinical Text Analysis and Knowledge Extraction System (cTAKES): architecture, component evaluation and applications. Journal of the American Medical Informatics Association. 2010;17(5):507–13. pmid:20819853
  38. 38. Lindberg C. The Unified Medical Language System (UMLS) of the National Library of Medicine. Journal of the American Medical Record Association. 1990;61(5):40–2. pmid:10104531
  39. 39. Centers for Disease Control. Suicide: Risk and Protective Factors 2017 [cited 2018]. Available from: https://www.cdc.gov/violenceprevention/suicide/riskprotectivefactors.html.
  40. 40. McLean J, Maxwell M, Platt S, Harris F, Jepson R. RISK AND PROTECTIVE FACTORS FOR SUICIDE AND SUICIDAL BEHAVIOUR: A LITERATURE REVIEW In: Health SDCfM, editor.: Scottish Government Social Research 2008.
  41. 41. Association AP. Teen Suicide is Preventable 2018 [cited 2018]. Available from: http://www.apa.org/research/action/suicide.aspx.
  42. 42. Kaslow N. Teen Suicides: What Are the Risk Factors?: Child Mind Institute; 2018 [cited 2018]. Available from: https://childmind.org/article/teen-suicides-risk-factors/.
  43. 43. Jeste DV, Savla GN, Thompson WK, Vahia IV, Glorioso DK, Martin AvS, et al. Association between older age and more successful aging: critical role of resilience and depression. The American journal of psychiatry. 2013;170(2):188–96. pmid:23223917.
  44. 44. Rodriguez JJ, Kuncheva LI, Alonso CJ. Rotation forest: A new classifier ensemble method. IEEE transactions on pattern analysis and machine intelligence. 2006;28(10):1619–30. Epub 2006/09/22. pmid:16986543.
  45. 45. Beam AL, Kohane IS. Big data and machine learning in health care. JAMA. 2018;319(13):1317–8. pmid:29532063
  46. 46. Team RC. R: A Language and Environment for Statistical Computing. 3.3.3 ed. Vienna, Austria R Foundation for Statistical Computing; 2017.
  47. 47. Canty AJ. Resampling methods in R: the boot package. R News. 2002;2(3):2–7.
  48. 48. Dietterich TG. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 1998;10(7):1895–923. pmid:9744903
  49. 49. Khalilia M, Chakraborty S, Popescu M. Predicting disease risks from highly imbalanced data using random forest. BMC Medical Informatics and Decision Making. 2011;11(1):51. pmid:21801360
  50. 50. Bae S, Lee S, Lee S. Predicion by data mining, of suicide attempts in Korean adolescents: a national study. Neuropsychiatric Disease and Treatment. 2015;11:2367–75. https://doi.org/10.2147/NDT.S91111. pmid:26396521
  51. 51. Bridge JA, Goldstein TR, Brent DA. Adolescent suicide and suicidal behavior. Journal of Child Psychology and Psychiatry. 2006;47(3–4):372–94. pmid:16492264
  52. 52. Connor JJ, Rueter MA. Parent-child relationships as systems of support or risk for adolescent suicidality. Journal of Family Psychology. 2006;20(1):143–55. http://dx.doi.org/10.1037/0893-3200.20.1.143. 2006-03561-016. pmid:614454629
  53. 53. Cabitza F, Rasoini R, Gensini G. Unintended consequences of machine learning in medicine. JAMA. 2017;318(6):517–8. pmid:28727867
  54. 54. Caruana R, Lou Y, Gehrke J, Koch P, Strum M, Elhadad N, editors. Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2015; 1721–1730.
  55. 55. Borowsky IW, Ireland M, Resnick MD. Adolescent Suicide Attempts: Risks and Protectors. Pediatrics. 2001;107(3):485–93. pmid:11230587
  56. 56. Lipschitz JM, Yen S, Weinstock LM, Spirito A. Adolescent and caregiver perception of family functioning: Relation to suicide ideation and attempts. Psychiatry Research. 2012;200(2):400–3. http://dx.doi.org/10.1016/j.psychres.2012.07.051.
  57. 57. Lachal J, Orri M, Sibeoni J, Moro MR, Revah-Levy A. Metasynthesis of Youth Suicidal Behaviours: Perspectives of Youth, Parents, and Health Care Professionals. PLoS ONE. 2015;10(5):e0127359. pmid:26001066
  58. 58. Luo W, Phung D, Tran T, Gupta S, Rana S, Karmakar C, et al. Guidelines for Developing and Reporting Machine Learning Predictive Models in Biomedical Research: A Multidisciplinary View. J Med Internet Res. 2016;18(12):e323. Epub 16.12.2016. pmid:27986644.
  59. 59. Passos IC, Mwangi B, Cao B, Hamilton JE, Wu M-J, Zhang XY, et al. Identifying a clinical signature of suicidality among patients with mood disorders: A pilot study using a machine learning approach. Journal of Affective Disorders. 2016;193:109–16. pmid:26773901
  60. 60. Brady TM, Neu AM, Miller ER, Appel LJ, Siberry GK, Solomon BS. Real-Time Electronic Medical Record Alerts Increase High Blood Pressure Recognition in Children. Clinical pediatrics. 2015;54(7):667–75. PMC4455954. pmid:25416626
  61. 61. Harwich E, Layock K. Thinking on its own: AI in the NHS. Reform, 2018 January 2018. Report No.
  62. 62. Barak-Corren Y, Castro VM, Javitt S, Hoffnagle AG, Dai Y, Perlis RH, et al. Predicting Suicidal Behavior From Longitudinal Electronic Health Records. American Journal of Psychiatry. 2017;174(2):154–62. pmid:27609239.
  63. 63. Pons E, Braun LMM, Hunink MGM, Kors JA. Natural Language Processing in Radiology: A Systematic Review. Radiology. 2016;279(2):329–43. pmid:27089187.
  64. 64. Haerian K, Salmasian H, Friedman C. Methods for Identifying Suicide or Suicidal Ideation in EHRs. AMIA Annual Symposium Proceedings. 2012;2012:1244–53. PMC3540459. pmid:23304402
  65. 65. Stewart C, Crawford PM, Simon GE. Changes in coding of suicide attempts or self-harm with transition from ICD-9 to ICD-10. Psychiatric services. 2017;68(3):215–. pmid:27903145
  66. 66. Millner AJ, Lee MD, Nock MK. Single-Item Measurement of Suicidal Behaviors: Validity and Consequences of Misclassification. PLoS ONE. 2015;10(10):e0141606. pmid:26496707