Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Automated detection scheme for acute myocardial infarction using convolutional neural network and long short-term memory

  • Ryosuke Muraki,

    Roles Data curation, Formal analysis, Investigation, Methodology, Software, Writing – original draft

    Affiliation Graduate School of Health Sciences, Fujita Health University, Toyoake, Japan

  • Atsushi Teramoto ,

    Roles Conceptualization, Investigation, Methodology, Project administration, Writing – review & editing

    teramoto@fujita-hu.ac.jp

    Affiliation Faculty of Radiological Technology, School of Medical Sciences, Fujita Health University, Toyoake, Japan

  • Keiko Sugimoto,

    Roles Data curation, Investigation, Writing – review & editing

    Affiliation Faculty of Medical Technology, School of Medical Sciences, Toyoake, Japan

  • Kunihiko Sugimoto,

    Roles Data curation, Investigation, Writing – review & editing

    Affiliation Fujita Health University Hospital, Toyoake, Japan

  • Akira Yamada,

    Roles Investigation, Validation, Writing – review & editing

    Affiliation School of Medicine, Fujita Health University, Toyoake, Japan

  • Eiichi Watanabe

    Roles Conceptualization, Data curation, Funding acquisition, Investigation, Project administration, Writing – review & editing

    Affiliation School of Medicine, Fujita Health University, Toyoake, Japan

Abstract

The early detection of acute myocardial infarction, which is caused by lifestyle-related risk factors, is essential because it can lead to chronic heart failure or sudden death. Echocardiography, among the most common methods used to detect acute myocardial infarction, is a noninvasive modality for the early diagnosis and assessment of abnormal wall motion. However, depending on disease range and severity, abnormal wall motion may be difficult to distinguish from normal myocardium. As abnormal wall motion can lead to fatal complications, high accuracy is required in its detection over time on echocardiography. This study aimed to develop an automatic detection method for acute myocardial infarction using convolutional neural networks (CNNs) and long short-term memory (LSTM) in echocardiography. The short-axis view (papillary muscle level) of one cardiac cycle and left ventricular long-axis view were input into VGG16, a CNN model, for feature extraction. Thereafter, LSTM was used to classify the cases as normal myocardium or acute myocardial infarction. The overall classification accuracy reached 85.1% for the left ventricular long-axis view and 83.2% for the short-axis view (papillary muscle level). These results suggest the usefulness of the proposed method for the detection of myocardial infarction using echocardiography.

Introduction

Acute myocardial infarction (AMI) is a disease in which myocardial cells become necrotic due to thrombus formation or blood vessel occlusion. AMI causes severe chest pain and requires immediate treatment, such as percutaneous transluminal coronary recanalization or coronary artery bypass grafting. It is important to diagnose AMI as early as possible because it can lead to heart failure, arrhythmia, or sudden death.

Echocardiography, a noninvasive imaging modality used to diagnose AMI, enables the real-time assessment of cardiac function and complications and evaluation of regional abnormal wall motion in patients with AMI. Therefore, it is widely used in cardiology. However, depending on disease range and severity, regional abnormal wall motion can be difficult to recognize. Moreover, the accuracy of its recognition depends on sonographer experience.

Deep learning, an artificial intelligence technique, was recently confirmed to have excellent processing power. Convolutional neural networks (CNNs), which are deep learning models, have been widely applied in various fields such as medical image analysis [112]. In the domain of echocardiography, Kusunose et al. proposed a method for detecting regional abnormal wall motion on echocardiography images and obtained a high detection accuracy with an area under the curve of approximately 0.9 [7]. Huang et al. proposed a technique for the visualization of AMI on echocardiography, and their results showed a Dice index of approximately 0.8 [9]. Thus, CNN is a highly accurate technique for AMI detection and classification.

Deep learning has also been widely applied for the multi-classification of video datasets. Recurrent neural networks (RNNs), which are also deep learning models, are particularly effective at predicting and classifying sequential data such as wave signals, natural language, and video [1316]. RNNs have a recursive structure, meaning that they analyze and produce outputs based on previous time-series and sequential data. Long short-term memory (LSTM), which was recently developed, can be used to analyze even larger and longer-term datasets than RNNs can process [17, 18]. Methods based on LSTM have been widely applied, and Ullah et al. proposed a multi-classification method for video images in large video datasets that obtained a classification accuracy of > 90% [19]. Zhou et al. proposed a natural language classification method based on LSTM and obtained a classification accuracy of >80% [20]. In addition, there has been considerable research done on combining CNNs and RNNs [21].

Kusunose et al. conducted a study on a novel detection method using deep learning techniques for abnormal regional wall motion noted on echocardiography. Multiple deep learning models were utilized to classify echocardiography into routine normal cases and those with abnormal regional wall motion. They adopted the ROC curve as an evaluation method and compared the classification accuracy of the clinical technologists to that of the proposed deep learning model. This comparison resulted in the AUC of the proposed method (0.97) and the AUC of the clinical technologists (0.95) being virtually equivalent, confirming the effectiveness of the proposed method. However, the study had a few limitations. The first limitation was the use of short-axis echocardiogram views only at the level of the papillary muscle. Second, the deep learning input data were limited to only three images per case: end-diastolic, mid-systolic, and end-systolic. Therefore, we attempted to detect AMI by analyzing the wall motion temporal changes and inputting the echocardiograph of one cardiac cycle into deep learning, which can then analyze time series. In addition to the short-axis papillary muscle (PM) level view, we used another view that is easy to use, the left ventricular long axis view, to detect AMI. We believe this method could be integral to the development of detection support technology so that non-specialist physicians in their clinics can accurately detect an AMI. In this study, we focused on CNN and LSTM, which have been widely applied to medical image analysis and can be used for video processing and analysis. Specifically, we aimed to develop an automated detection scheme for AMI using CNN and LSTM in echocardiography.

Materials and methods

Overview of the proposed method

An overview of the proposed method is presented in Fig 1. Echocardiography images were loaded into VGG16 [22], a CNN model, to extract the features. The obtained features were then analyzed using LSTM for the classification of AMI and normal myocardium.

Echocardiography

For this study, we collected short-axis PM level and left ventricular LX view images taken with ultrasound equipment (Vivid E9 and Vivid E95, GE Healthcare) at Fujita health university hospital. A total of 202 cines were collected as inputs: 99 diagnoses of acute anteroseptal infarction of the proximal left anterior descending artery in the American Heart Association Committee Report and 103 normal cases. Cardiologists and experienced sonographers usually estimate the culprit coronary artery in patients with myocardial infarction using left ventricular long-axis, left ventricular short-axis, apical four-chamber, apical two-chamber, and apical long-axis views. Here we employed the short-axis PM level and left ventricular long-axis views because it is relatively easy to detect anteroseptal infarction of the proximal left anterior descending artery, which can be used to evaluate abnormal wall motion in cases of anteroseptal infarction. Moreover, all cases of anteroseptal infarction used in this study underwent coronary angiography, and occlusion of #6 was observed by examination. In addition, patients who underwent percutaneous coronary intervention after coronary angiography and had abnormal wall motion on echocardiography were included. Fig 2 shows the views used. Table 1 and Fig 3 show the baseline clinical characteristics of this cohort and the selection of the study population, respectively. P values indicate differences between patients who have normal myocardium and AMI. P < 0.05 was considered statistically significant.

thumbnail
Fig 2.

Left ventricular long-axis view and short-axis papillary muscle level view (left: views of anatomy; right: normal views).

https://doi.org/10.1371/journal.pone.0264002.g002

thumbnail
Table 1. Baseline clinical characteristics of the study cohort.

https://doi.org/10.1371/journal.pone.0264002.t001

The image preprocessing involved electrocardiogram (ECG) removal, cropping, and frame interpolation of the views. On echocardiography, a two-lead ECG is drawn to identify indicators during the cardiac cycle such as end-diastole and end-systole. Fig 4 shows the preprocessing of the input images. In this study, to recognize the cardiac wall motion on each view, the ECG was removed and trimmed to form a bounding rectangle. To consider differences in heart rate between patients, one cardiac cycle was extracted from each image and the number of frames was interpolated to 30. Interpolation means that if the number of frames in the video image for one cardiac cycle was 50–60, they were interpolated at equal intervals so that the number of frames was 30, whereas if the number of frames was 10–20, they were interpolated so that the total number of echocardiography images for all patients was 30. Linear interpolation was used as the complementation method, and each of the images for one cardiac cycle was extracted based on the length between the peaks (R-R interval) of the simultaneously recorded ECG. This study was approved by an institutional review board of Fujita Health University and informed consents were obtained from patients subject to the condition of data anonymization (No. HM19-345).

thumbnail
Fig 4. Preprocessing of the input data.

All frames of one cardiac cycle taken from the original echocardiography image were interpolated to 30 frames, and the ECG in the image was removed and trimmed.

https://doi.org/10.1371/journal.pone.0264002.g004

Feature extraction

The features of the interpolated images were extracted for input into the classification model [23]. CNNs can extract features simultaneously as the final outputs or extract parameters as intermediate outputs from individual layers. Varshni et al. used CNNs to extract features from chest radiographs [24]. Hyeon et al. used CNNs to extract features from cytology images and used conventional machine learning methods to differentiate between benign and malignant cells [24, 25]. By using CNNs as feature extractors and other models as output layers, new inputs can be added and the accuracy further improved compared to CNN use alone. Therefore, we focused on this method and adopted VGG16 and global pooling [26], another CNN model, as feature extractors. Using these feature extraction methods, we extracted features from all frames of the interpolated echocardiography images.

VGG16 model.

This study used VGG16, a CNN model developed by the Visual Geometry Group at the University of Oxford in 2014, for the feature extraction. The structure of VGG16 is shown in Fig 5. It consists of 13 convolutional layers, 5 pooling layers, and 3 fully connected layers [27]. We introduced the VGG16 pretraining network using the large natural ImageNet image dataset. From the second fully connected layer of VGG16, 4096 features were extracted and input into the classification model.

Global pooling. Global average pooling (GAP) and global max pooling (GMP) are 2 compression methods for feature maps extracted from images by CNNs. These compression methods select only the maximum or average values from the last feature map of the CNN and pool all of the feature maps. Fig 6 shows a simplified diagram of the GAP and GMP methods. These compression methods were applied to the feature maps after completion of the convolutional and pooling processes until just before the fully connected layer of the CNN. GAP, shown in the upper part of Fig 6, extracts the average value from each feature map and outputs only the extracted value as an intermediate output. GMP, shown in the lower part of Fig 6, extracts the maximum value from each feature map and outputs it as an intermediate output. This process significantly reduces the number of dimensions from the original feature map parameters and prevents overfitting. In this study, we performed global pooling of the 7 × 7 × 512 feature maps extracted by VGG16 and output the average or maximum value from each feature map, which resulted in a total of 512 parameters.

thumbnail
Fig 6. Simplified diagram of the global average pooling and global max pooling compression methods.

Pooling is performed to extract the maximum or average values from the last feature map of the CNN.

https://doi.org/10.1371/journal.pone.0264002.g006

LSTM networks

Because the detection of AMI on echocardiography requires the evaluation and analysis of wall motion over time, two-dimensional images with different time phases were input into the CNN. We also focused on RNNs, which are excellent tools for processing sequential data and effective for time-series information such as cine images and wave signals as well as text data and natural language. Therefore, this model is characterized by its ability to control sequential information. Fig 7 demonstrates the principle of RNN, where x, y, and h are the input, output, and weight of the hidden layer, respectively.

thumbnail
Fig 7. Schematic diagram of a recurrent neural network.

In the diagram, x is the input, h is the hidden layer, and y is the output. The RNN learns by passing the weights of the hidden layer to the next hidden layer in the forward direction.

https://doi.org/10.1371/journal.pone.0264002.g007

The RNN connects the layer at time (t) with the previous layer (t−1) and calculates the parameters in the hidden layer (h(t)) and the output according to the following equations: (1) (2) U, W, and V denote the weights calculated during training, and f(z) and g(zm) denote the sigmoid and softmax functions, respectively. The equations for the respective activation functions are as follows: (3) (4)

However, because RNNs theoretically store all past data during training, the vanishing gradient problem arises due to the divergence and disappearance of weights. Therefore, we focused on LSTM, an improved RNN model. The principle of the LSTM is shown in Fig 8. The difference between LSTM and RNN is that LSTM features a mechanism of information selection called “gate” and “cell.” There are 3 types of gates: “input,” “output,” and “forgetting.” Eq (i) shows the formula for the forgetting gate (ft): (i) From the input information, the output of the LSTM layer at (t−1) and the cell, the information that is unnecessary in the learning at (t), is selected and “forgotten.” Eq (ii) is used to determine the input gate (It): (ii) The output of the last LSTM layer and the value of the cell are used to determine the new value to be updated. The value of the updated cell (Ct) is then determined by Eq (iii): (iii) The value of the cell determined by the above equation is propagated to the next LSTM layer, and the output (Ot) of the LSTM layer at (t) is determined by Eqs (iv) and (v) in the output gate section: (iv) (v) Using this gating mechanism, LSTM can analyze long-term series data and solve the vanishing gradient problem of conventional RNNs. We introduced these mechanisms to analyze the wall motion over time using echocardiography.

thumbnail
Fig 8. Diagram of the long short-term memory principle.

Input, forgetting, and output gates are used to determine the information to be input into, retained, and output from the cells, enabling the learning of sequential data over a long period of time.

https://doi.org/10.1371/journal.pone.0264002.g008

Finally, as shown in Fig 9, the features extracted by the VGG16 method were input into the LSTM to classify the normal and AMI cases. For the hyperparameters, we set the learning rate to 1 × 10−5, the number of epochs to 50, the batch size to 30, and the input data size to 4096 × 30.

thumbnail
Fig 9. Classification of myocardial infarction and normal myocardium cases. LSTM, long short-term memory.

Features from 1 to 30 frames were input to each LSTM, and the classification was performed using the softmax function based on the LSTM output.

https://doi.org/10.1371/journal.pone.0264002.g009

Evaluation

Cross-validation method.

The cross-validation method was used to evaluate the classification accuracy of the constructed model. Fig 10 shows a simplified diagram of the cross-validation method. All datasets were initially divided into several groups, one of which was used as the test group for the evaluation, while the remaining data were used for training. Thereafter, the accuracy was comprehensively calculated by repeating the process such that all data are test data. In this study, we used a five-fold cross-validation method in which the 202 echocardiography images were divided into 142 cases for training, 20 cases for evaluation during training, and the remaining 40 cases for testing, with random sampling so that all cases were used as test data.

thumbnail
Fig 10. Simplified diagram of the cross-validation method (number of folds = 5).

https://doi.org/10.1371/journal.pone.0264002.g010

Comparison with conventional artificial neural network.

To demonstrate the effectiveness of our method, the classification accuracy was also evaluated using five-fold cross-validation on a normal artificial neural network (ANN), which does not feature a mechanism to handle time-series relations separately from LSTM [28, 29]. A schematic of the ANN is shown in Fig 11, in which all features extracted by the VGG16 method for each frame were combined and used as input to the neural network to classify the normal and AMI cases. For the hyperparameters, we set the learning rate to 1 × 10−5, the number of epochs to 50, the batch size to 30, and the input data size to 4096 × 30.

thumbnail
Fig 11. Schematic representation of the artificial neural network classification of myocardial infarction and normal cases.

Features from 1 to 30 frames were transformed into one-dimensional data and input into an ANN.

https://doi.org/10.1371/journal.pone.0264002.g011

Results

Tables 2 and 3 show the confusion matrices and overall classification accuracies of LSTM and ANN for the LX images, while Tables 4 and 5 show the results for the PM images. Table 6 shows the classification accuracy for the LX and PM images for the given parameters and classifiers. Tables 7 and 8 show the sensitivities, specificities, and area under the curves (AUC) for the LX and PM images for the given parameters and classifiers. The accuracy of LSTM was the best when GAP was used for both the LX and PM images. The accuracies for the LX and PM images were 0.896 and 0.867, respectively.

thumbnail
Table 2. Overall classification accuracy of long short-term memory for long-axis view images.

https://doi.org/10.1371/journal.pone.0264002.t002

thumbnail
Table 3. Overall classification accuracy of the artificial neural network for the long-axis view images.

https://doi.org/10.1371/journal.pone.0264002.t003

thumbnail
Table 4. Overall classification accuracy of the long short-term memory for the short-axis papillary muscle level images.

https://doi.org/10.1371/journal.pone.0264002.t004

thumbnail
Table 5. Overall classification accuracy of the artificial neural network for the short-axis papillary muscle level images.

https://doi.org/10.1371/journal.pone.0264002.t005

thumbnail
Table 6. Overall classification accuracy with changing parameters.

https://doi.org/10.1371/journal.pone.0264002.t006

thumbnail
Table 8. Results of short-axis papillary muscle view images.

https://doi.org/10.1371/journal.pone.0264002.t008

Figs 1215 compare the ROC curve. Figs 1623 show representative correctly and incorrectly classified cases on echocardiography.

thumbnail
Fig 12. ROC curves of the LSTM versus ANN in long-axis view.

https://doi.org/10.1371/journal.pone.0264002.g012

thumbnail
Fig 13. ROC curves of the GAP versus GMP in long-axis view.

https://doi.org/10.1371/journal.pone.0264002.g013

thumbnail
Fig 14. ROC curves of the LSTM versus ANN in short-axis papillary muscle view.

https://doi.org/10.1371/journal.pone.0264002.g014

thumbnail
Fig 15. ROC curves of the GAP versus GMP in short-axis papillary muscle view.

https://doi.org/10.1371/journal.pone.0264002.g015

thumbnail
Fig 18. False-negative cases on long-axis view images with anteroseptal infarction with regional abnormal wall motion circled in red.

https://doi.org/10.1371/journal.pone.0264002.g018

thumbnail
Fig 20. False-positive cases on short-axis view papillary muscle level images.

https://doi.org/10.1371/journal.pone.0264002.g020

thumbnail
Fig 21. True-negative cases on short-axis view papillary muscle level images.

https://doi.org/10.1371/journal.pone.0264002.g021

thumbnail
Fig 22. False-negative cases on short-axis view papillary muscle level images.

https://doi.org/10.1371/journal.pone.0264002.g022

thumbnail
Fig 23. True-positive cases on short-axis view papillary muscle level images.

https://doi.org/10.1371/journal.pone.0264002.g023

Discussion

In this study, we proposed an automated classification scheme for AMI and normal cases on echocardiography images using deep learning. The VGG16 method was used to extract features from the echocardiography images, while LSTM was used for the classification. The comparison of the classification models (Tables 2 and 3) shows that the results obtained with LSTM were better than those obtained using the ANN. The overall classification accuracy using LSTM was 0.852 for the LX images and 0.832 for the PM images. These results suggest that LSTM can classify AMI and normal cases with higher accuracy than ANN and analyze and classify useful features over time. In addition, the classification accuracy of LSTM suggests that the image information of one cardiac cycle (consisting of 30 frames) is useful for analyzing myocardial motion, thereby distinguishing AMI from normal myocardium. Unlike ordinary ANNs, LSTMs have a mechanism known as “gates,” optimizing them for time-series information analysis. The results showed that the LSTM can detect AMI on echocardiography with better accuracy than ANN by analyzing time series information; moreover, the results confirmed the superiority of the proposed LSTM method. Further, an AI-based study on laryngitis pathology classification showed that the LSTM classification accuracy was 15% higher than regular ANNs [30]. Other studies using AI-based solar radiation prediction have also shown that LSTM has superior accuracy in predicting solar radiation [31]. In this study by comparison, the classification accuracy was improved using LSTM, resulting in more effective use of time series data, once again confirming the effectiveness of this method.

Table 4 shows that the overall classification accuracy of LSTM was best on the LX and PM images when GAP was used. The classification accuracy on the LX images was 0.896, while that on the PM images was 0.867. These results suggest that GAP extracts more useful features for classification than GMP. In addition, the results of the comparison between the features extracted from the fully connected layer and the 512 features extracted by GAP showed that the classification accuracy of LSTM increased when GAP was used. This finding suggests that GAP reduces the number of unnecessary features for classification and achieves more efficient learning by reducing the number of parameters. The reason for the lack of change in classification accuracy between GAP and GMP when LSTM was used in the LX images may be that similar parameters were extracted from the feature maps by GAP and GMP during the pooling process.

Visual comparison of the incorrectly and correctly classified cases showed that those with low video contrast, high noise, or high brightness of the myocardium were misclassified. These results suggest that image quality, such as noise and contrast in the video image, is among the most important factors in the classification of AMI and normal cases on echocardiography. In addition, incorrect cases tended not to be shown adequately in the image: the left ventricle was blurred and a different short-axis level view was shown. In future work, accuracy should be improved by the analysis of echocardiography images and patient data from more facilities to create a robust network model.

Similar studies are listed in Table 9 for comparison with our study. Since there are very few studies with the same images and objectives, a simple comparison with this study may be difficult. However, our method was able to classify MI with an accuracy of more than 80% using 202 cases, confirming its validity.

We then calculated the accuracy of the classification using left ventricular LX and short-axis PM level views, which were subsequently used to detect anteroseptal infarction. Acute anteroseptal infarction was evaluated by cardiologists and experienced sonographers using left ventricular LX and short-axis PM level views as well as apical four-chamber and apical LX views, allowing for observation of the apex. However, inexperienced clinicians, non-cardiologists, residents, and those otherwise unfamiliar with echocardiography may find it difficult to obtain apical four-chamber and apical LX view images with adequate quality. In addition, the detection of AMI is an emergent matter and requires accurate and rapid detection using left ventricular LX and short-axis PM level views, which are relatively easy to obtain. These results indicate that this method correctly identifies acute anteroseptal infarction with superior accuracy clearly distinguishing it from normal myocardium. Therefore, this method can greatly assist non-cardiologists and inexperienced clinicians alike to diagnose acute anteroseptal infarction during initial treatment.

Limitations and future works

This study has a few limitations. First, the echocardiograms were performed at the same institution. Second, this classification method is performed offline; therefore, it is necessary to apply it to real-time processing so that classification during echocardiography can be utilized. Third, we did not evaluate each segment of the heart individually; rather, we examined only the acute anterior wall septal infarction with occlusion of #6 in the American Heart Association Committee Report, which occurs the most frequently. Since this study focused only on acute anteroseptal infarction, the classification and evaluation of infarcts in each segment and in other coronary dominant regions should be performed in the future.

Conclusion

In this study, we developed an automatic detection scheme for AMI on echocardiography images using CNN and LSTM. The accuracy of the classification showed that our proposed method was able to classify AMI and normal cases with high accuracy, confirming its effectiveness as a supplemental tool for the detection of AMI on echocardiography. Here specialists and skilled doctors can easily detect an anteroseptal infarction. However, it may be difficult for residents and physicians who are unfamiliar with echocardiography at the time of the initial visit or physicians in non-cardiology clinics to detect it. Anteroseptal infarctions occur frequently and require accurate detection and diagnosis, regardless of the technician or physician’s experience, field, or situation. This method can contribute to the detection of AMI and is expected to lead to its appropriate treatment of and the prognosis of affected patients. Another technical novelty of this study is the use of LSTM, which enables a time-series analysis of wall motion on echocardiography. The results showed that LSTM can detect AMI more accurately than ANNs without a time-series analysis function, confirming its superiority using LSTM. In addition, here we used left ventricular long-axis and short-axis views (papillary muscle level), which are minimal and easy to depict for diagnosis, as input. Since the classifications were performed using highly accurate views, we found the possibility of applying this method using LSTM to other views, which are easy to take. In addition, although this study focused only on acute anteroseptal infarction, its methodology is expected to be extended to the detection of infarcts in other coronary artery dominant regions.

Acknowledgments

We thank Mr. Kimura and Ms. Yamazaki for the useful discussions as well as Prof. Saito and Prof. Fujita for assisting and teaching.

References

  1. 1. Krizhevsky A, Sutskever I, Hinton GE. ImageNet classification with deep convolutional neural networks. Proceedings of the 25th International Conference on Neural Information Processing Systems. 2012;1: 1097–1105. https://doi.org/10.1145/3065386
  2. 2. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. Proceedings of the 27th International Conference on Neural Information Processing Systems. 2014;2: 2672–2680. https://doi.org/10.1145/3422622
  3. 3. Kooi T, Litjens G, Ginneken BV, Gubern-Mérida A, Sánchez CI, Mann R, et al. Large scale deep learning for computer aided detection of mammographic lesions. Med Image Anal. 2017;35: 303–312. pmid:27497072
  4. 4. Sakai Y, Takemoto S, Hori K, Nishimura M, Ikematsu H, Yano T, et al. Automatic detection of early gastric cancer in endoscopic images using a transferring convolutional neural network. Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society. 2018; 4138–4141. https://doi.org/10.1109/EMBC.2018.8513274
  5. 5. Zhang J, Gajjala S, Agrawal P, Tison GH, Hallock LA, Beussink-Nelson L, et al. Fully automated echocardiogram interpretation in clinical practice. Circulation. 2018;138: 1623–1635. pmid:30354459
  6. 6. Wolterink JM, van Hamersvelt RW, Viergever MA, Leiner T, Išguma I. Coronary artery centerline extraction in cardiac CT angiography using a CNN-based orientation classifier. Med Image Anal. 2018;51: 46–60. pmid:30388501
  7. 7. Kusunose K, Abe T, Haga A, Fukuda D, Yamada H, Harada M, et al. A deep learning approach for assessment of regional wall motion abnormality from echocardiographic images. JACC Cardiovasc Imaging. 2020;2: 374–381. pmid:31103590
  8. 8. Teramoto A, Tsukamoto T, Yamada A, Kiriyama Y, Imaizumi K, Saito K, et al. Deep learning approach to classification of lung cytological images: two-step training using actual and synthesized images by progressive growing of generative adversarial networks. PLoS One. 2018;15: 3. pmid:32134949
  9. 9. Huang M, Wang C, Chiang J, Liu P, Tsai W. Automated recognition of regional wall motion abnormalities through deep neural network interpretation of transthoracic echocardiography. Circulation. 2020;142: 1510–1520. pmid:32964749
  10. 10. Sumitomo M, Teramoto A, Toda R, Fukami N, Fukaya K, Zennami K, et al. Deep learning using preoperative magnetic resonance imaging information to predict early recovery of urinary continence after robot-assisted radical prostatectomy. Int J Urol. 2020;10: 922–928. pmid:32729184
  11. 11. Gadekallu TR, Alazab M, Kaluri R, Maddikunta RKP, Bhattacharya S, Lakshmannna K. Hand gesture classification using a novel CNN-crow search algorithm. Complex Intell Syst. 2021.
  12. 12. Vasan D, Alazab M, Wassan S, Safaei B, Zheng Q. Image-Based malware classification using ensemble of CNN architectures (IMCEC). Comput Secur. 2020;92.
  13. 13. Sutskever I, Hinton GE, Taylor G. The recurrent temporal restricted Boltzmann machine. Proceedings of the 21st International Conference on Neural Information Processing Systems. 2020: 1601–1608.
  14. 14. Mikolov T, Kombrink S, Burget L, Černocký J, Khudanpur S. Extensions of recurrent neural network language model. 2011 IEEE International Conference on Acoustics, Speech and Signal Processing. 2011: 5528–5531. https://doi.org/10.1109/ICASSP.2011.5947611
  15. 15. Sutskever I, Martens J, Hinton GE. Generating text with recurrent neural networks. Proceedings of the Twenty-eight International Conference on Machine Learning. 2011: 1017–1024.
  16. 16. Boulanger-Lewandowski N, Bengio Y, Vincent P. Modeling temporal dependencies in high-dimensional sequences: application to polyphonic music generation and transcription. Proceedings of the Twenty-nine International Conference on Machine Learning. 2011: 1881–1888.
  17. 17. Alazab M, Khan S, Krishnan RSS, Pham Q, Reddy KPM, Gadekallu RT. A Multidirectional LSTM Model for Predicting the Stability of a Smart Grid. IEEE Access. 2020;8: 85454–85463.
  18. 18. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Computation. 1997;9: 1735–1780. pmid:9377276
  19. 19. Ullah A, Ahmad J, Muhammad K, Sajjad M, Baik SW. Action recognition in video sequences using deep bi-directional LSTM with CNN features. IEEE Access. 2018;6: 1155–1166.
  20. 20. Zhou P, Qi Z, Zheng S, Xu J, Bao H, Xu B. Text classification improved by integrating bidirectional LSTM with two-dimensional max pooling. Proceedings of COLING 2016, 26th International Conference on Computational Linguistics: Technical Papers. 2018: 3486–3495.
  21. 21. Rehman A, Rehman US, Khan M, Alazab M, Reddy T. CANintelliIDS: Detecting In-Vehicle Intrusion Attacks on a Controller Area Network using CNN and Attention-based GRU. IEEE Trans Netw Sci Eng. 2021:1–11.
  22. 22. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. International Conference on Learning Representations. 2015: 1409.1556.
  23. 23. Abbas S, Jalil Z, Javed AR, Batool I, Khan MZ, Noorwali A, et al. BCD-WERT: a novel approach for breast cancer detection using whale optimization based efficient features and extremely randomized tree algorithm. PeerJ Comput Sci. 2021. pmid:33817036
  24. 24. Varshni D, Thakral K, Agarwal L, Nijhawan R, Mittal A. Pneumonia detection using CNN based feature extraction. 2019 IEEE International Conference on Electronics, Communication and Computing Technologies. 2019: 1–7. https://doi.org/10.1109/ICECCT.2019.8869364
  25. 25. Hyeon J, Choi H, Lee BD, Lee KN. Diagnosing cervical cell images using pre-trained convolutional neural network as feature extractor. In: 2017 IEEE International Conference on Big Data and Smart Computing. 2017: 390–393. https://doi.org/10.1109/BIGCOMP.2017.7881741
  26. 26. Lin M, Chen Q, Yan S. Network in network. 2013;arXiv: 1312.4400.
  27. 27. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, et al. ImageNet large scale visual recognition challenge. Int J Comput Vision. 2015;115: 211–252.
  28. 28. Jain AK, Mao J, Mohiuddin KM. Artificial neural networks: a tutorial. Computer. 1996;29: 31–44.
  29. 29. Arlot A, Celisse A. A survey of cross-validation procedures for model selection. Statist Surveys. 2010;4: 40–79.
  30. 30. Guedas V, Junior A, Fernandes J, Teixeira F, Teixeira P J. Long Short Term Memory on Chronic Laryngitis Classification. Procedia Computer Science. 2018;138: 250–257. https://doi.org/10.1016/j.procs.2018.10.036.
  31. 31. Zang H, Liu L, Sun L, Cheng L, Wei Z, Sun G. Short-term global horizontal irradiance forecasting based on a hybrid CNN-LSTM model with spatiotemporal correlations. Renewable Energy. 2020;160: 26–41. https://doi.org/10.1016/j.renene.2020.05.150.
  32. 32. Zhang N, Yang G, Gao Z, Xu C, Zhang Y, Shi R, et al. Deep Learning for Diagnosis of Chronic Myocardial Infarction on Nonenhanced Cardiac Cine MRI. Radiology. 2019;291: 606–617. pmid:31038407
  33. 33. Baloglu BU, Talo M, Yildirim O, Tan SR, Acharya RU. Classification of Myocardial Infarction with multi-lead ECG signals and deep CNN. Pattern Recognit Lett. 2019;122: 23–30. https://doi.org/10.1016/j.patrec.2019.02.01610.1016/j.patrec.2019.02.016.
  34. 34. Vece DD, Laumer M, Schwyzer M, Burkholz R, Corinzia L, Cammann LV, et al. Artificial intelligence in echocardiography diagnostics–detection of takotsubo syndrome. Eur Heart J. 2020;41.
  35. 35. Shimizu M, Cho S, Misu Y, Ohmori M, Tateishi R, Kaneda T. Diagnostic Performance of Deep Learning on 12-lead Electrocardiography to Distinguish Takotsubo Syndrome and Acute Anterior Myocardial Infarction. Circulation. 2020;142. https://doi.org/10.1161/circ.142.suppl_3.15414
  36. 36. Shahin IA, Almotairi S. An Accurate and Fast Cardio-Views Classification System Based on Fused Deep Features and LSTM. IEEE Access. 2020;8: 135184–135194.