Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Automatic segmentation of multiple cardiovascular structures from cardiac computed tomography angiography images using deep learning

  • Lohendran Baskaran ,

    Roles Conceptualization, Methodology, Project administration, Writing – original draft, Writing – review & editing

    lob2008@med.cornell.edu, lohendran.baskaran@gmail.com

    Affiliations Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America, Department of Radiology, New York-Presbyterian Hospital and Weill Cornell Medicine, New York, New York, United States of America, Department of Cardiovascular Medicine, National Heart Centre, Singapore, Singapore

  • Subhi J. Al’Aref,

    Roles Conceptualization

    Affiliations Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America, Department of Radiology, New York-Presbyterian Hospital and Weill Cornell Medicine, New York, New York, United States of America

  • Gabriel Maliakal,

    Roles Data curation, Formal analysis

    Affiliation Cleerly, Inc, New York, New York, United States of America

  • Benjamin C. Lee,

    Roles Data curation, Formal analysis

    Affiliation Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America

  • Zhuoran Xu,

    Roles Data curation, Formal analysis

    Affiliation Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America

  • Jeong W. Choi,

    Roles Conceptualization

    Affiliation Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America

  • Sang-Eun Lee,

    Roles Resources

    Affiliations Division of Cardiology, Severance Cardiovascular Hospital, Integrative Cardiovascular Imaging Center, Yonsei University College of Medicine, Seoul, South Korea, Division of Cardiology, Department of Internal Medicine, Ewha Womans University Seoul Hospital, Seoul, Korea

  • Ji Min Sung,

    Roles Resources

    Affiliation Division of Cardiology, Severance Cardiovascular Hospital, Integrative Cardiovascular Imaging Center, Yonsei University College of Medicine, Seoul, South Korea

  • Fay Y. Lin,

    Roles Conceptualization, Resources

    Affiliations Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America, Department of Radiology, New York-Presbyterian Hospital and Weill Cornell Medicine, New York, New York, United States of America

  • Simon Dunham,

    Roles Data curation

    Affiliation Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America

  • Bobak Mosadegh,

    Roles Data curation, Resources

    Affiliation Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America

  • Yong-Jin Kim,

    Roles Resources

    Affiliation Department of Internal Medicine, Seoul National University College of Medicine, Cardiovascular Center, Seoul National University Hospital, Seoul, South Korea

  • Ilan Gottlieb,

    Roles Resources

    Affiliation Department of Radiology, Casa de Saude São Jose, Rio de Janeiro, Brazil

  • Byoung Kwon Lee,

    Roles Resources

    Affiliation Division of Cardiology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul, Korea

  • Eun Ju Chun,

    Roles Resources

    Affiliation Department of Radiology, Seoul National University Bundang Hospital, Sungnam, South Korea

  • Filippo Cademartiri,

    Roles Resources

    Affiliation Cardiovascular Imaging Center, SDN IRCCS, Naples, Italy

  • Erica Maffei,

    Roles Resources

    Affiliation Department of Radiology, Area Vasta 1/ASUR Marche, Urbino, Italy

  • Hugo Marques,

    Roles Resources

    Affiliation UNICA, Unit of Cardiovascular Imaging, Hospital da Luz, Lisboa, Portugal

  • Sanghoon Shin,

    Roles Resources

    Affiliation Division of Cardiology, Department of Internal Medicine, Ewha Womans University Seoul Hospital, Seoul, Korea

  • Jung Hyun Choi,

    Roles Resources

    Affiliation Pusan University Hospital, Busan, South Korea

  • Kavitha Chinnaiyan,

    Roles Resources

    Affiliation Department of Cardiology, William Beaumont Hospital, Royal Oak, Michigan, United States of America

  • Martin Hadamitzky,

    Roles Resources

    Affiliation Department of Radiology and Nuclear Medicine, German Heart Center Munich, Munich, Germany

  • Edoardo Conte,

    Roles Resources

    Affiliation Centro Cardiologico Monzino, IRCCS, Milan, Italy

  • Daniele Andreini,

    Roles Resources

    Affiliation Centro Cardiologico Monzino, IRCCS, Milan, Italy

  • Gianluca Pontone,

    Roles Resources

    Affiliation Centro Cardiologico Monzino, IRCCS, Milan, Italy

  • Matthew J. Budoff,

    Roles Resources

    Affiliation Department of Medicine, Los Angeles Biomedical Research Institute, Torrance, California, United States of America

  • Jonathon A. Leipsic,

    Roles Resources

    Affiliation Department of Medicine and Radiology, University of British Columbia, Vancouver, British Columbia, Canada

  • Gilbert L. Raff,

    Roles Resources

    Affiliation Department of Cardiology, William Beaumont Hospital, Royal Oak, Michigan, United States of America

  • Renu Virmani,

    Roles Resources

    Affiliation Department of Pathology, CVPath Institute, Gaithersburg, Maryland, United States of America

  • Habib Samady,

    Roles Resources

    Affiliation Division of Cardiology, Emory University School of Medicine, Atlanta, Georgia, United States of America

  • Peter H. Stone,

    Roles Resources

    Affiliation Cardiovascular Division, Brigham and Women’s Hospital, Harvard Medical School, Boston, Massachusetts, United States of America

  • Daniel S. Berman,

    Roles Resources

    Affiliation Department of Imaging and Medicine, Cedars Sinai Medical Center, Los Angeles, California, United States of America

  • Jagat Narula,

    Roles Resources

    Affiliation Icahn School of Medicine at Mount Sinai, Mount Sinai Heart, Zena and Michael A. Wiener Cardiovascular Institute, and Marie-Josée and Henry R. Kravis Center for Cardiovascular Health, New York, New York, United States of America

  • Jeroen J. Bax,

    Roles Resources

    Affiliation Department of Cardiology, Leiden University Medical Center, Leiden, the Netherlands

  • Hyuk-Jae Chang,

    Roles Resources

    Affiliation Division of Cardiology, Severance Cardiovascular Hospital, Integrative Cardiovascular Imaging Center, Yonsei University College of Medicine, Seoul, South Korea

  • James K. Min,

    Roles Resources

    Affiliations Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America, Department of Radiology, New York-Presbyterian Hospital and Weill Cornell Medicine, New York, New York, United States of America

  •  [ ... ],
  • Leslee J. Shaw

    Roles Resources

    Affiliations Dalio Institute of Cardiovascular Imaging, Weill Cornell Medicine, New York, New York, United States of America, Department of Radiology, New York-Presbyterian Hospital and Weill Cornell Medicine, New York, New York, United States of America

  • [ view all ]
  • [ view less ]

Abstract

Objectives

To develop, demonstrate and evaluate an automated deep learning method for multiple cardiovascular structure segmentation.

Background

Segmentation of cardiovascular images is resource-intensive. We design an automated deep learning method for the segmentation of multiple structures from Coronary Computed Tomography Angiography (CCTA) images.

Methods

Images from a multicenter registry of patients that underwent clinically-indicated CCTA were used. The proximal ascending and descending aorta (PAA, DA), superior and inferior vena cavae (SVC, IVC), pulmonary artery (PA), coronary sinus (CS), right ventricular wall (RVW) and left atrial wall (LAW) were annotated as ground truth. The U-net-derived deep learning model was trained, validated and tested in a 70:20:10 split.

Results

The dataset comprised 206 patients, with 5.130 billion pixels. Mean age was 59.9 ± 9.4 yrs., and was 42.7% female. An overall median Dice score of 0.820 (0.782, 0.843) was achieved. Median Dice scores for PAA, DA, SVC, IVC, PA, CS, RVW and LAW were 0.969 (0.979, 0.988), 0.953 (0.955, 0.983), 0.937 (0.934, 0.965), 0.903 (0.897, 0.948), 0.775 (0.724, 0.925), 0.720 (0.642, 0.809), 0.685 (0.631, 0.761) and 0.625 (0.596, 0.749) respectively. Apart from the CS, there were no significant differences in performance between sexes or age groups.

Conclusions

An automated deep learning model demonstrated segmentation of multiple cardiovascular structures from CCTA images with reasonable overall accuracy when evaluated on a pixel level.

Introduction

In evaluating cardiovascular disease (CVD), the imaging of structures plays a key role in diagnosis, as well as in surveillance of progression. Coronary Computed Tomography Angiography (CCTA) provides isotropic high spatial resolution imaging non-invasively. In both research and clinical workflows, the necessary quantitative and qualitative evaluation of these structures is assisted via available commercial software packages. However, this requires manual input, rendering this process time-consuming and operator-dependent [1].

As it uses large amounts of information to build predictive models via novel algorithmic strategies, machine learning (ML) is well-suited to imaging, and its role in the cardiovascular space is expanding [2]. Deep learning is a subdomain of ML that uses sophisticated frameworks comprising networks with many intermediate layers of “neurons” to perform automated feature extraction. This results in the ability to map inputs to outputs via complex pathways [3]. Here, we apply deep learning to CCTA images to develop, demonstrate and evaluate an automated model for the identification of multiple cardiovascular structures.

Methods

Study population

The population consisted of a convenience sample randomly selected from an international, multicenter, prospective, observational registry that has been described previously [4,5]. Inclusion criteria were patients undergoing clinically indicated CCTA, with images of sufficient quality to be annotated. Known coronary artery disease (CAD), hemodynamic instability, arrhythmia, and uninterpretable CCTA were exclusion criteria. Each site obtained local institutional review or ethics board approval.

Image acquisition and segmentation

Scanners were ≥ 64-detector rows, and acquisition, post-processing and interpretation were performed to current guidelines [6].

Images were obtained and reconstructed at 0.50 mm thickness. Files in Digital Imaging and Communications in Medicine (DICOM) format were transmitted to a core laboratory, and structure annotation was done in a blinded manner by level III–experienced technologists. The luminal segments of eight cardiovascular structures were annotated from the iodinated contrast-tissue border interface using Adobe Photoshop (Adobe Systems, San Jose, California): the proximal ascending and descending aorta (PAA, DA), superior and inferior vena cavae (SVC, IVC), pulmonary artery (PA), coronary sinus (CS), right ventricular wall (RVW) and left atrial wall (LAW). The superior and inferior axial limits of all structures were limited to within the guideline-recommended scan range, from below the tracheal bifurcation or the mid-level of the left pulmonary artery and extending below the cardiac border [6], and any remaining segments of structures outside this range were not analyzed. The PAA was identified as originating at the plane corresponding to the nadirs of all 3 aortic valve cusps, to the plane most proximal to the origin of the brachiocephalic artery. The DA was defined as originating immediately distal to the origin of the left subclavian artery, extending to the most inferior axial slice. The vena cavae were identified as being venous vessels coursing along the right middle mediastinum, adjacent to and to the right of the trachea and PAA and draining into the right atrium [7,8]. The PA included the main, left and right pulmonary arteries. The CS was identified as the cardiac venous structure continuing in the atrioventricular groove from the great cardiac vein, adjacent to left circumflex coronary artery and draining into the right atrium [9]. RVW was defined as the right ventricular myocardial volume derived by the delineation of its endocardial and epicardial borders, excluding papillary muscles and trabeculations, and followed the contours below the atrioventricular valve planes on a three-dimensional isotropic voxel level. LAW identification used the left atrial myocardial volume derived by the delineation of its endocardial and epicardial borders and included the appendage but excluded adjacent veins. These annotations, established and verified by board certified cardiologists, were used as the “ground truth” for the deep learning model.

Splitting of dataset and preprocessing

The entire dataset was split into three parts; training (70%), validation (20%) and testing (10%). No two parts contained images from the same patient. An open source python library known as ‘psd_tools’ was used for the extraction of the ground truth from the annotated images [10]. Annotated slices were extracted from the ground truth Photoshop files and arranged according to each label/structure’s assigned color code. DICOM volumes were converted to have an isotropic voxel spacing of 0.625 mm x 0.625 mm x 0.625 mm, with the same volumetric resolution for the extracted labels. The images were windowed with a Hounsfield unit (HU) window (-300,500) so that all structures of interest were optimally visible. Each input image contained a pair of background and foreground label images. All the images and corresponding labels were then resized to 512x512 pixels and passed to the model.

Deep learning model

As it has previously been used for medical image segmentation in thoracic images, a convolutional neural network (CNN), U-Net was used for the deep learning architecture [11,12]. U-net comprises 4 layers (Fig 1). The image is first down-sampled by a Conv3x3 layer consisting of two runs through a set comprising of a convolution with 3x3 kernel, Rectified Linear Unit (ReLU) and a batch normalization layer. The output feature maps from this layer are further down sampled by half the resolution. After 4 layers of this, the feature maps are now up sampled by transposed convolution (kernel size of 2 and a stride of 2 followed by successive Conv3x3 blocks). The feature maps from the contracting path are concatenated with those of the expanding path. At the final layer the feature maps are reduced from 128 to 2 using a Conv 1x1 block which consists of a 1x1 convolutional kernel, and pixel-wise probabilities for belonging to each class is obtained once this is passed to a Softmax layer. Eight similar but separate networks were trained for each structure. Prior work on other cardiovascular structures using this framework has previously been reported [13].

thumbnail
Fig 1. The deep learning network architecture.

Each red block represents two consecutive sets of 3x3 Convolution layer, ReLU activation and Batch Normalization. The number of feature maps in each layer is on top of each box, while the size of each feature map per layer is indicated on the left. Finally, the feature maps are passed through a softmax layer to obtain pixel wise probabilities of belonging to the given class to obtain the outputs.

https://doi.org/10.1371/journal.pone.0232573.g001

Training strategy and model evaluation

The images were randomly shuffled and passed to the network (batch size 4, resolution 512x512). The network output was binary masks for two classes i.e. the foreground and the background. The Dice loss, obtained by subtracting the mean Dice similarity score from 1, was used to train the network [14]. Training used Adam optimizer with a learning rate of 0.001 to carry out training [15]. Outputs were compared with the ground truth which contains complimentary images of the contour of interest using the Dice loss. The network having the lowest Dice loss on the validation set amongst the epochs (one cycle through the full training dataset) was selected and evaluated on the test set. Data was shuffled every epoch. Each model was trained for 50 epochs, and the model with the best validation loss was chosen among these epochs. Image-based performance metric was based on Dice loss, calculated by subtracting the mean Dice similarity score from 1 [14]. This score quantifies the pixel-wise degree of similarity between the model predicted segmentation mask and the ground truth, and ranges from 0 (no similarity) to 1 (identical) (Fig 2), mathematically expressed as:

thumbnail
Fig 2. Dice score visualization.

The Dice score is used to gauge model performance, ranging from 0 to 1. 1 corresponds to a pixel perfect match between the deep learning model output (red, A and D) and ground truth annotation (green, B and D). The model output with the higher Dice score (A) had greater overlap (C) with the ground truth (B) than the output (D) that had lesser overlap (F), as it did not predict that main pulmonary artery that was annotated in (E).

https://doi.org/10.1371/journal.pone.0232573.g002

Statistical analysis

Statistical analysis was performed using Python 3.7 using the scikit-learn library. Continuous and normally distributed variables were expressed by mean ± standard deviation. Categorical data were expressed by number and percentage. Dice scores were summarized as medians and quartiles. Subgroup analysis of Dice scores by gender and age were compared by Wilcoxon test. A P value of <0.05 was considered significant.

Results

The study comprised 206 patients, with 19,572 images and 5.130 billion pixels. Cohort mean age was 59.9 ± 9.4 yrs., and was 42.7% female (Table 1). Prevalence of diabetes, dyslipidemia, hypertension and smoking were 22.22%, 34.78%, 44.93% and 64.73% respectively. The training set comprised 144 patients (13701 images), validation set 42 patients (3914 images) and testing set 20 patients (1957 images). There were no differences for patient characteristics between the training, validation and test sets (Table 2).

A combined overall median Dice score of 0.820 (interquartile range: 0.782, 0.843) was achieved on the validation set for the structures (PAA, DA, SVC, IVC, PA, CS, RVW, LAW) and demonstrative comparisons between the original image, manual annotation and model prediction are shown in Fig 3. Median Dice scores for PAA, DA, SVC, IVC, PA, CS, RVW and LAW were 0.969 (interquartile range: 0.979, 0.988), 0.953 (interquartile range: 0.955, 0.983), 0.937 (interquartile range: 0.934, 0.965), 0.903 (interquartile range: 0.897, 0.948), 0.775 (interquartile range: 0.724, 0.925), 0.720 (interquartile range: 0.642, 0.809), 0.685 (interquartile range: 0.631, 0.761) and 0.625 (interquartile range: 0.596, 0.749) respectively (Table 3). For PAA, there were no significant differences between sexes (male = 0.954, female = 0.987) or age groups (age<65 = 0.969, age ≥65 = 0.968). For the DA, there were no significant differences in Dice scores between sex- or age-based subgroups (male = 0.949, female = 0.956, age<65 = 0.923, age ≥65 = 0.965), nor were there for the SVC (male = 0.929, female = 0.945, age<65 = 0.946, age ≥65 = 0.929), IVC (male = 0.891, female = 0.918, age<65 = 0.912, age ≥65 = 0.896), PA (male = 0.773, female = 0.778, age<65 = 0.738, age ≥65 = 0.799), RVW (male = 0.669, female = 0.702, age<65 = 0.699, age ≥65 = 0.676), or the LAW (male = 0.586, female = 0.668, age<65 = 0.630, age ≥65 = 0.620) (P > 0.05 for all). For the CS, although there was no significant difference in model prediction performance between scores for age<65 (0.756) and age ≥65 (0.709) (P = 0.365), there was a difference between the scores for male (0.784) and female (0.670) patients (P = 0.048). Automated segmentation for all 8 structures took 19.37 seconds per patient, at 0.198 seconds per slice, whereas manual segmentation took approximately 1 hour per patient.

thumbnail
Fig 3. Demonstrative comparisons between the ground truth manual annotation (green) and prediction overlaid on the image(red) are shown for the: A) PAA, B) DA, C) SVC, D) IVC, E) CS, F) PA, G) RVW and H) LAW.

https://doi.org/10.1371/journal.pone.0232573.g003

Discussion

This study demonstrated the capability of a deep learning model to rapidly identify the majority of the great vessels, the coronary sinus, and the left atrial and right ventricular walls in an automated, pixel-wise manner. This was done within this multicenter, international cohort with reasonable accuracy overall, agreeing well with manual annotation across sex- and age -stratified subgroups.

This model was able to identify the thoracic aorta (PAA and DA) with high accuracy on a pixel level, as reflected by high Dice scores. A prior study using deep learning on 331 abdominal CTs segmented the abdominal aorta with a mean Dice score of 0.796 [16]. However, this performance was on the validation set, rather than on an unseen testing set, as compared to the current study. Another study using dilated CNNs segmented three parts of the thoracic aorta (PAA, aortic arch and DA) and obtained Dice scores of 0.83–0.88 [17]. Although the dataset used only 24 scans with two-fold cross validation, the Dice score was commendable as the CT scans were non-contrast and non-gated. Our study obtained Dice scores of 0.969 and 0.953 for the PAA and DA respectively, and that may be partly attributable to the use of contrast and ECG gating that may have better delineated the border between the vessel wall and lumen. This enhanced edge and visual boundary detection may assist in CNN-based segmentation tasks [18]. Other studies on chest CTs obtained Dice scores of 0.93–0.95 [1921]. However, these were label- or atlas-based rather than deep learning-based, and segmented the thoracic aorta as a whole, rather than distinguishing the PAA from the DA. In a large open-access challenge, Dice scores for the PAA, obtained using deep learning-based techniques, were similar to ours [22]. As the PAA and DA are large tubular structures are already easily identified by the radiologist, the potential added value of an expansion of the current method would be to detect abnormalities along these structures in a rapid manner as well as to quantify measurable parameters, e.g. to identify a larger cross-sectional diameter indicative of an aortic aneurysm. To our knowledge, this is the first demonstration of deep learning-based segmentation on both the PAA and DA in cardiac gated contrast-enhanced scans.

Although the IVC has previously been segmented, it was via the use of masks that required manual input [23]. That study obtained a Dice score of 0.896. Conversely, the current study is automated, and obtained a higher score of 0.903 for the IVC and 0.937 for the SVC, congruent with the overall performance advantage of deep learning methods over atlas-based algorithms in the segmentation of other cardiovascular structures [22]. This performance is despite vena cavae visualization that is below ideal (Fig 4), due to the irregular shape and suboptimal contrast timing, as the contrast for the scans in this study were optimized for coronary artery opacification, rather than the vena cavae. Although the SVC and IVC are easily identified by the radiologist, the current method potentiates measurements of the vessels required for insertion of stents, or for the rapid identification of pathology, e.g. thrombi. To the best of our knowledge, this is the first attempt to use deep learning to segment the vena cavae.

thumbnail
Fig 4. Accurate prediction of the SVC despite suboptimal contrast opacification.

The ground truth annotation is in green and the model prediction in red. Abbreviations: SVC = superior vena cava.

https://doi.org/10.1371/journal.pone.0232573.g004

The indications for imaging and segmentation of the CS have broadened in parallel with developments in left ventricular pacing, arrhythmia ablation and transcatheter mitral valve replacements and repairs. The difficulty in identifying the irregularly-shaped CS, especially with suboptimal contrast opacification, may account for the lower Dice score of 0.720 when compared to the other vessels. When compared to other tubular cardiovascular structures, this disparity in segmentation performance is congruent with a prior study. In a study using model-based frameworks to segment multiple cardiovascular structures from 35 CT scans, segmentation obtained a volume overlap of 0.952 on the aorta, but only 0.704 on the CS [24]. Furthermore, in our study, there was an additional deterioration in performance in images of female subject compared to male. Whilst this may partly due to the generally smaller size of CS in females, further external validation is required to address this. To the best of our knowledge, the current study is the first to use CNN-based deep learning to segment the CS, providing proof of feasibility. This performance is expected to improve with datasets, using specialized CS imaging protocols.

Imaging of the PA may aid in the identification and prognostication of pulmonary hypertension. In the study using model-based frameworks, segmentation of the PA obtained a volume overlap of 0.940 [24]. Although not using deep learning, that performance is superior than the current study’s performance, with a Dice score of 0.775. In another study comprising an open challenge (using both atlas- and deep learning-based models), segmentation of the PA obtained a Dice score of 0.80 [22]. These results suggest that a deep learning approach may not be the best solution to segmentation of the PA, whose variability in terms of shape and appearance is notably greater than structures such as the DA. Additionally, our current model employs a 2D CNN model, that does not incorporate information in serial images along the z-axis. The models previously mentioned incorporate 3D information, that may aid in segmentation. This limitation of our current study is scope for further development.

When compared to the tubular structures, both the RVW and LAW have poorer performance. This may be due to the irregular shape and thin walls for these structures. Although prior studies have segmented the right ventricle, these studies have not distinguished between the blood pool or volume and the wall [22,24]. In contradistinction, to the best of our knowledge, this is the first demonstration of a deep learning-based model to segment the RVW separately. To our knowledge, this is the first demonstration of a deep learning-based LAW segmentation model.

A deep learning model for multiple structures would be consistent, reproducing the same result every time. During this study, certain model outputs obtained lower Dice scores, but on further inspection, this was due to ground truth annotation error by human readers, rather than by the model (Fig 5). This study introduces the feasibility of a deep-learning model as a “second reader”. A second reader reduces interpretative error, and results in changes in decision with meaningful clinical impact. Deep learning could contribute with minimal time and cost issues, improving its feasibility. The rapid throughput could allow future integration into clinical workflows with minimal disruption, helping alleviate the clinical diagnostic burden [25] As such, this study tentatively raises the possibility of clinical integration to enhance diagnostic speed, lower costs and reduce error.

thumbnail
Fig 5. An example of the model as a “second reader”, correcting human error.

The DA was partially missed during that were missed during manual annotation (green). The model correctly identified these (red), but obtained a lower Dice score as it did not match the ground truth set during manual annotation. Abbreviations: DA = descending aorta.

https://doi.org/10.1371/journal.pone.0232573.g005

There are limitations to the current study. Whilst predicting 8 structures, this model omitted the four cardiac chambers and the left ventricular wall. However, these structures have previously been segmented by our group, and this current study is an extension of that work [13]. Although segmenting most of the great vessels, the current study does not segment the pulmonary veins. The pulmonary veins are of therapeutic interest, as potential sites for arrhythmia ablation. The anatomy of the pulmonary veins is variable, as are its boundaries, and mapping of these structures for ablation planning could be possible, with modification of the current model. Further work could include prediction of the pulmonary veins, which has been segmented previously [24]. The scan timing of these images did not optimize contrast opacification for a number of structures, including the vena cavae, RVW and CS. The current study did not include 3D information in the CNN, and future developments to incorporate this will likely improve performance. An additional limitation is the study size. There was a “hold-out” 10% test set that the model never “saw” until final performance evaluation. Whist this comprised only 20 patients, this set comprised 1957 images, a number deemed more than adequate for medical image-based deep learning applications [26]. The high Dice score attests to the model’s robustness between the training, validation and testing cohorts, but cannot exclude overfitting. Overfitting, whilst not excluded, may be less likely as the model did make incorrect predictions (Fig 2). A larger study that serves as external validation will address these limitations and may improve performance.

Concluding, an automated deep learning model demonstrated segmentation of cardiovascular structures from CCTA images with reasonable overall accuracy when evaluated on a pixel level, and to the best of our knowledge, is the first demonstration of deep learning to segment the SVC, IVC, CS, RVW and LAW. This heralds its integration into research and clinical workflows.

References

  1. 1. Fuchs A, Mejdahl MR, Kühl JT, Stisen ZR, Nilsson EJP, Køber LV, et al. Normal values of left ventricular mass and cardiac chamber volumes assessed by 320-detector computed tomography angiography in the Copenhagen General Population Study. Eur Heart J Cardiovasc Imaging. 2016 Sep 1;17(9):1009–17. pmid:26758412
  2. 2. Al’Aref SJ, Anchouche K, Singh G, Slomka PJ, Kolli KK, Kumar A, et al. Clinical applications of machine learning in cardiovascular disease and its relevance to cardiac imaging. Eur Heart J. 2018 Jul 27;
  3. 3. Hinton G. Deep Learning—A Technology With the Potential to Transform Health Care. JAMA. 2018 Sep 18;320(11):1101–2. pmid:30178065
  4. 4. Lee S-E, Chang H-J, Rizvi A, Hadamitzky M, Kim Y-J, Conte E, et al. Rationale and design of the Progression of AtheRosclerotic PlAque DetermIned by Computed TomoGraphic Angiography IMaging (PARADIGM) registry: A comprehensive exploration of plaque progression and its impact on clinical outcomes from a multicenter serial coronary computed tomographic angiography study. Am Heart J. 2016 Dec;182:72–9. pmid:27914502
  5. 5. Lee Sang-Eun, Sung Ji Min, Rizvi Asim, Lin Fay Y., Kumar Amit, Hadamitzky Martin, et al. Quantification of Coronary Atherosclerosis in the Assessment of Coronary Artery Disease. Circulation: Cardiovascular Imaging. 2018 Jul 1;11(7):e007562.
  6. 6. Abbara S, Blanke P, Maroules CD, Cheezum M, Choi AD, Han BK, et al. SCCT guidelines for the performance and acquisition of coronary computed tomographic angiography: A report of the society of Cardiovascular Computed Tomography Guidelines Committee: Endorsed by the North American Society for Cardiovascular Imaging (NASCI). J Cardiovasc Comput Tomogr. 2016 Dec;10(6):435–49. pmid:27780758
  7. 7. Lin FY, Devereux RB, Roman MJ, Meng J, Jow VM, Simprini L, et al. The Right Sided Great Vessels by Cardiac Multidetector Computed Tomography: Normative Reference Values among Healthy Adults Free of Cardiopulmonary Disease, Hypertension, and Obesity. Academic Radiology. 2009 Aug 1;16(8):981–7. pmid:19394871
  8. 8. Sonavane SK, Milner DM, Singh SP, Abdel Aal AK, Shahir KS, Chaturvedi A. Comprehensive Imaging Review of the Superior Vena Cava. RadioGraphics. 2015 Nov;35(7):1873–92. pmid:26452112
  9. 9. Shah SS, Teague SD, Lu JC, Dorfman AL, Kazerooni EA, Agarwal PP. Imaging of the Coronary Sinus: Normal Anatomy and Congenital Abnormalities. RadioGraphics. 2012 Jun 27;32(4):991–1008. pmid:22786990
  10. 10. Python package for reading Adobe Photoshop PSD files: psd-tools/psd-tools [Internet]. psd-tools; 2019 [cited 2019 May 16]. https://github.com/psd-tools/psd-tools
  11. 11. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF, editors. Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015. Springer International Publishing; 2015. p. 234–41. (Lecture Notes in Computer Science).
  12. 12. Javaid U, Dasnoy D, Lee JA. Multi-organ Segmentation of Chest CT Images in Radiation Oncology: Comparison of Standard and Dilated UNet. In: Blanc-Talon J, Helbert D, Philips W, Popescu D, Scheunders P, editors. Advanced Concepts for Intelligent Vision Systems. Springer International Publishing; 2018. p. 188–99. (Lecture Notes in Computer Science).
  13. 13. Baskaran L, Maliakal G, Al’Aref SJ, Singh G, Xu Z, Michalak K, et al. Identification and Quantification of Cardiovascular Structures From CCTA: An End-to-End, Rapid, Pixel-Wise, Deep-Learning Method. JACC: Cardiovascular Imaging [Internet]. 2019 Oct 11 [cited 2019 Oct 21]; http://www.sciencedirect.com/science/article/pii/S1936878X19308733
  14. 14. Sudre CH, Li W, Vercauteren T, Ourselin S, Cardoso MJ. Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations. arXiv:170703237 [cs]. 2017;10553:240–8.
  15. 15. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. arXiv:14126980 [cs] [Internet]. 2014 Dec 22 [cited 2019 May 16]; http://arxiv.org/abs/1412.6980
  16. 16. Roth HR, Oda H, Zhou X, Shimizu N, Yang Y, Hayashi Y, et al. An application of cascaded 3D fully convolutional networks for medical image segmentation. Computerized Medical Imaging and Graphics. 2018 Jun 1;66:90–9. pmid:29573583
  17. 17. Noothout JMH, de Vos BD, Wolterink JM, Isgum I. Automatic Segmentation of Thoracic Aorta Segments in Low-Dose Chest CT. arXiv:181005727 [cs] [Internet]. 2018 Oct 9 [cited 2019 Oct 15]; http://arxiv.org/abs/1810.05727
  18. 18. Kivinen JJ, Williams CKI, Heess N. Visual Boundary Prediction: A Deep Neural Prediction Network and Quality Dissection.:10.
  19. 19. Xie Y, Padgett J, Biancardi AM, Reeves AP. Automated aorta segmentation in low-dose chest CT images. Int J Comput Assist Radiol Surg. 2014 Mar;9(2):211–9. pmid:23877280
  20. 20. Kurugol S, San Jose Estepar R, Ross J, Washko GR. Aorta segmentation with a 3D level set approach and quantification of aortic calcifications in non-contrast chest CT. In: 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society. 2012. p. 2343–6.
  21. 21. Sedghi Gamechi Z, Bons LR, Giordano M, Bos D, Budde RPJ, Kofoed KF, et al. Automated 3D segmentation and diameter measurement of the thoracic aorta on non-contrast enhanced CT. Eur Radiol. 2019 Sep;29(9):4613–23. pmid:30673817
  22. 22. Zhuang X, Bai W, Song J, Zhan S, Qian X, Shi W, et al. Multiatlas whole heart segmentation of CT data using conditional entropy for atlas ranking and selection. Medical Physics. 2015;42(7):3822–33. pmid:26133584
  23. 23. Yan Z, Chen F, Wu F, Kong D. Inferior vena cava segmentation with parameter propagation and graph cut. Int J Comput Assist Radiol Surg. 2017 Sep;12(9):1481–99. pmid:28421319
  24. 24. Ecabert O, Peters J, Walker MJ, Ivanc T, Lorenz C, von Berg J, et al. Segmentation of the heart and great vessels in CT images using a model-based adaptation framework. Med Image Anal. 2011 Dec;15(6):863–76. pmid:21737337
  25. 25. Waite S, Scott J, Gale B, Fuchs T, Kolla S, Reede D. Interpretive Error in Radiology. American Journal of Roentgenology. 2016 Dec 27;208(4):739–49. pmid:28026210
  26. 26. Cho J, Lee K, Shin E, Choy G, Do S. How much data is needed to train a medical image deep learning system to achieve necessary high accuracy? arXiv:151106348 [cs] [Internet]. 2015 Nov 19 [cited 2019 Mar 28]; http://arxiv.org/abs/1511.06348