Clinical research

DOI: 10.4244/EIJ-D-20-00570

Training and validation of a deep learning architecture for the automatic analysis of coronary angiography

Tianming Du1, PhD; Lihua Xie2, MSc; Honggang Zhang1, PhD; Xuqing Liu1, PhD; Xiaofei Wang3, MSE; Donghao Chen3, MSE; Yang Xu3, BSE; Zhongwei Sun2, MSc; Wenhui Zhou3, PhD; Lei Song2, MD; Changdong Guan2, MSc; Alexandra J. Lansky4, MD; Bo Xu2, MBBS

Abstract

Background: In recent years, the use of deep learning has become more commonplace in the biomedical field and its development will greatly assist clinical and imaging data interpretation. Most existing machine learning methods for coronary angiography analysis are limited to a single aspect.

Aims: We aimed to achieve an automatic and multimodal analysis to recognise and quantify coronary angiography, integrating multiple aspects, including the identification of coronary artery segments and the recognition of lesion morphology.

Methods: A data set of 20,612 angiograms was retrospectively collected, among which 13,373 angiograms were labelled with coronary artery segments, and 7,239 were labelled with special lesion morphology. Trained and optimised by these labelled data, one network recognised 20 different segments of coronary arteries, while the other detected lesion morphology, including measures of lesion diameter stenosis as well as calcification, thrombosis, total occlusion, and dissection detections in an input angiogram.

Results: For segment prediction, the recognition accuracy was 98.4%, and the recognition sensitivity was 85.2%. For detecting lesion morphologies including stenotic lesion, total occlusion, calcification, thrombosis, and dissection, the F1 scores were 0.829, 0.810, 0.802, 0.823, and 0.854, respectively. Only two seconds were needed for the automatic recognition.

Conclusions: Our deep learning architecture automatically provides a coronary diagnostic map by integrating multiple aspects. This helps cardiologists to flag and diagnose lesion severity and morphology during the intervention.

Introduction

Coronary artery disease (CAD) is the most common cardiovascular disease1, and the leading cause of death globally during the past two decades2. Therefore, the diagnosis and prevention of CAD is crucial for modern society. Coronary angiography (CAG), which provides assessments of luminal stenosis, plaque characteristics, and disease activity, is an important tool for CAD diagnosis and treatment guidance3,4. In recent years, the use of deep learning has become more commonplace in the biomedical field and its development will greatly assist clinical and imaging data interpretation5. Deep learning can simplify the procedure by directly learning predictive features, thereby strongly supporting the translation from artificial algorithms into clinical application6,7,8. However, much of the previous work to apply deep learning algorithms in the field of CAD has focused on single aspects of the analysis of the coronary artery, such as vessel segmentation9,10, coronary artery centreline extraction11, noise reduction12, coronary artery geometry synthesis13, coronary plaque characterisation14, and calcification detection. Thus, there remains a large gap between the results produced by the aforementioned algorithms and the actual diagnosis of CAD.

High diagnostic accuracy from a coronary angiogram requires correct recognition of lesion morphology and location. Herein, to tackle the above recognition tasks, two unique functional deep neural networks (DNN) were proposed, on which we created, trained, validated, and then tested a coronary angiography recognition system called DeepDiscern. DeepDiscern was evaluated on a test data set of consecutive angiograms collected from clinical cases.

Methods

STUDY POPULATION AND IMAGE ACQUISITION

To develop the DeepDiscern system, 20,612 angiograms from 10,073 patients were consecutively collected using the image acquisition data from a large single centre (Fu Wai Hospital, National Center for Cardiovascular Diseases, Beijing, China). For coronary segmentation DNN training, 13,373 angiograms were consecutively collected from 2,834 patients who underwent CAGs in July 2018. The remaining 7,239 angiograms, with at least one identifiable lesion morphology, such as stenotic lesion, total occlusion (TO), calcification, thrombus, and dissection, collected from 7,239 patients, were used for lesion morphology recognition. The collected angiogram information is listed in Table 1.

Raw angiographic data for our work were acquired during interventional procedures of patients and saved in 512×512-pixel digital imaging and communications in medicine (DICOM) format with angiographic views and video information, without patient identifiers. Each patient’s angiographic DICOM included several angiographic sequences encompassing different angiographic views. The choice of angiographic views was left to the operator’s discretion, to delineate best the lesion severity and morphology. In general, the angiographic views for the left coronary artery included CRA (cranial view), CAU (caudal view), LAO_CRA (left anterior oblique-cranial view), LAO_CAU (left anterior oblique-caudal view), RAO_CRA (right anterior oblique-cranial view), and RAO_CAU (right anterior oblique-caudal view). For the right coronary artery the views included LAO, LAO_CAU, LAO_CRA and RAO. The data flow is presented in Figure 1.

Figure 1. Data flow for the lesion morphology detection task and the coronary segment recognition task. Coronary segment recognition. In total, 13,373 angiograms were used, and divided into seven parts to train and test DeepDiscern DNN. Lesion morphology detection. In total, 7,239 angiograms with 1 to 3 lesion morphology were labelled for model training and testing. There were 12,184 lesion samples of five kinds of lesion morphology. CAU: caudal view; CRA: cranial view; LAO: left anterior oblique view; LAO_CAU: left anterior oblique-caudal view; LAO_CRA: left anterior oblique-cranial view: RAO: right anterior oblique view; RAO_CAU: right anterior oblique-caudal view; RAO_CRA: right anterior oblique-cranial view

REFERENCE STANDARD AND ANNOTATION PROCEDURES

DeepDiscern learns rules from the labelled images in the training phase. To this end, all angiograms collected over a period of 11 months for the training and testing data sets were reviewed by ten qualified analysts in the angiographic core lab at Fu Wai Hospital. Coronary segments were annotated based on pre-established diagnostic criteria and lesion morphology characterised.

For the coronary segment recognition data sets, each angiogram was labelled at a pixel-by-pixel level for coronary segmentation recognition. First, analysts annotated sketch labels of all coronary artery segments on the original angiograms, with different colours representing different arterial segments. Then a group of trained and certified technicians labelled fine ground-truth images pixel by pixel according to the sketch labels. Supplementary Figure 1 illustrates this process.

A total of 20 coronary artery segments were annotated (Supplementary Figure 2), including proximal right coronary artery (RCA prox), RCA mid, RCA distal, right posterior descending (PDA), right posterolateral, left main (LM), proximal left anterior descending (LAD prox), LAD mid, LAD distal, 1st diagonal, add. 1st diagonal, 2nd diagonal, add. 2nd diagonal, proximal circumflex (LCX prox), LCX distal, 1st obtuse marginal (OM), 2nd OM, left posterolateral, left posterior descending (L-PDA) and intermediate.

Although accurate coronary diagnosis requires coronary injections in multiple views to ensure that all coronary segments are seen clearly without foreshortening or overlap, it is not necessary to include all potential views in a given coronary segment. In clinical practice, several dominant projections are typically used to visualise a coronary segment and its morphology. Therefore, during the training and testing process of each coronary segment, DeepDiscern focused mainly on dominant projections provided. The corresponding relationship between the observed coronary segment and the angiographic views is shown in Supplementary Table 1.

For the lesion morphology detection data sets, expert analysts marked all lesion morphologies identified on the angiogram, including stenotic lesion, TO, calcification, thrombosis, and dissection. Stenotic lesion was defined as ≥50% diameter stenosis. TO was defined as angiographic evidence of TOs with Thrombolysis In Myocardial Infarction (TIMI) flow grade 0. Calcification was defined as readily apparent radiopacities noted within the apparent vascular wall (moderate: densities noted only with cardiac motion before contrast injection; severe: radiopacities noted without cardiac motion before contrast injection). Thrombus was defined as a discrete, intraluminal filling defect with defined borders and largely separated from the adjacent wall with or without contrast staining15. Dissection grade was diagnosed based on the National Heart, Lung and Blood Institute (NHLBI) coronary dissection criteria (Supplementary Table 2). The lesion type, location, and extent were labelled using a rectangular box. Supplementary Figure 1 illustrates this process. In total, 7,239 angiograms with one to three lesion morphologies were labelled for model training and testing. There were 12,184 positive samples in these angiograms. The lesion morphology classification data are shown in Table 1.

DEEP LEARNING MODEL

DeepDiscern was designed to use these two DNN to recognise coronary segments and detect lesion morphology (Figure 2). Features of the input angiogram carrying different semantic information were extracted including low-level features, such as vessel edges and background texture, and high-level features, such as the overall shape of the arteries.

Figure 2. The workflow of DeepDiscern. In total, 20,612 angiographic images were collected from DICOM videos of 10,073 patients. Under the supervision of the labelled images, we trained the lesion morphology detection model and coronary artery recognition model of DeepDiscern. After training, the detection and recognition models generate the result images. These two results were combined to generate a high-level diagnosis.

For the coronary artery recognition task, we modified a special DNN - conditional generative adversarial network (cGAN) (Supplementary Figure 3) for image segmentation. For the lesion morphology detection task, we developed a convolutional DNN (Supplementary Figure 4), which outputs the location of all the lesion morphologies that appeared in the input angiogram. The network structure, implementation details, training process and testing process of segment recognition DNN are detailed in Supplementary Appendix 1, and the lesion detection DNN in Supplementary Appendix 2.

For each input angiogram, DeepDiscern combines the two output results from coronary artery recognition DNN and lesion morphology detection DNN to generate high-level diagnostic information, including identification of every coronary artery lesion and the coronary artery segment in which it is located.

MODEL EVALUATION AND STATISTICAL ANALYSIS

For arterial segment recognition, given an input angiogram, the DeepDiscern segment recognition DNN produced an output image with several identified areas that represent the different coronary segments (Supplementary Table 3). For each coronary segment, we calculated the predicted pixel number for true positive (TP), true negative (TN), false positive (FP) and false negative (FN). Based on these results, we evaluated the segment recognition model by several metrics including accuracy ([TP+TN]/[TP+TN+FP+FN]), sensitivity (TP/[TP+FN]), specificity (TN/[TN+FP]), positive predictive value (TP/[TP+FP]), and negative predictive value (TN/[TN+FN]). The recognition model was evaluated using 1,050 images including all the coronary segments.

In terms of lesion morphology detection, the DeepDiscern lesion detection DNN predicts several rectangular areas containing the lesions to describe their location and type. For the algorithmic analysis, lesion morphology is detected correctly if the overlap rate of a predicted rectangle and the ground-truth rectangle (labelled by cardiologists) exceeds a threshold λ_d= 0.5. We measured the performance of the lesion detection model using precision rate P, recall rate R, and F1 score. The precision rate is defined as the percentage of correctly detected lesion cases from all lesion cases detected by the models. The recall rate is defined as the percentage of correctly detected lesions from all ground-truth lesions labelled by cardiologists. The F1 score [F_1= 2×P×R⁄(P+R)], which combines the accuracy rate and recall rate, is a better measure of the overall performance of the detection DNN model. The evaluation process of the recognition model and of the detection model is illustrated in Supplementary Figure 5.

Results

Coronary segment recognition DNN was evaluated using 1,050 images that included all the coronary segments. For segment prediction, the average accuracy, sensitivity, specificity, positive predictive value, and negative predictive value of all coronary artery segments was 98.4%, 85.2%, 99.1%, 76.2%, and 99.5%, respectively. Higher accuracy and sensitivity rates were observed in the proximal segments of major epicardial vessels (99.9% and 91.8% for LM, 99.8% and 92.6% for LAD proximal, 99.8% and 87.9% for LCX proximal, and 99.8% and 87.9% for RCA proximal). Arterial segments that were identified incorrectly were mostly in the distal segments and side branches of the major epicardial vessels. The performance of coronary segment recognition DNN was improved as the amount of data increased (Supplementary Figure 6). Because the majority of pixels in the image were negative, DNN performance cannot be assessed only from specificity and negative predictive value. Table 2 provides detailed results including more metrics, and Figure 3A illustrates result images of the artery recognition task. The results under different angiographic views are shown in Supplementary Table 4.

Figure 3. Result imaging of the segment recognition model and the lesion morphology detection model. A) Segment recognition. First row: input angiograms. Second row: resulting images generated by DeepDiscern segment recognition DNN. Third row: ground-truth labelled images. Different identified areas represent the different coronary segments. B) Lesion morphology detection. First row: input angiograms and ground-truth bounding boxes. There is a TO morphology in the first and second angiograms, and a thrombus morphology in the third angiogram in this row. Second row: bounding boxes and lesion types generated by the DeepDiscern lesion morphology detection model.

One thousand angiograms were used to test the lesion morphology detection DNN model. The test data set included 1,200 (248 stenotic, 228 calcification, 402 TO, 193 thrombus, and 129 dissection) lesion samples. The F1 score, which represents the overall performance of the DNN model, for stenotic lesion, TO, calcification, thrombus, and dissection was 0.829, 0.810, 0.802, 0.823 and 0.854, respectively. For all lesion morphologies, recall rates were higher than precision rates. Results are shown in Table 3 and examples of result images of the lesion morphology task are shown in Figure 3B. The receiver operating characteristic (ROC) curves for different lesion morphologies are shown in Figure 4. The area under the curve (AUC) of the lesion morphology detection DNN for stenotic lesion, TO, calcification, thrombus, and dissection was 0.801, 0.759, 0.799, 0.778, and 0.863, respectively.

Figure 4. ROC curves and AUC values of all lesions. DeepDiscern lesion detection DNN predict several bounding boxes, which may contain lesion morphologies. The bounding box with a correct location and a correct type is a positive sample. The bounding box with a wrong location or a wrong type is a negative sample.

The DeepDiscern system provides an automatic and multimodal diagnosis in a two-step process. DeepDiscern first recognises all the arterial segments in the angiogram, and then it detects the lesions in the angiogram. Processing these two steps took less than two seconds on average for every angiogram (1.280 seconds for the segment recognition task, and 0.648 seconds for the lesion detection task). Combining these two results, DeepDiscern can analyse all lesions appearing in an angiogram (Figure 5).

Figure 5. Combined results of DeepDiscern. In the first column are original angiograms. In the second column are resulting images of artery recognition DNN, where black areas represent background, white areas represent catheter, and other different colours represent different coronary artery segments. In the third column are resulting images of detection DNN. The location of lesion morphology on the angiogram is marked by several boundary boxes. The type of morphology is also predicted. In the fourth column are the combined results of recognition DNN and detection DNN.

Discussion

Many deep learning techniques that focus on a single aspect of the coronary angiogram and coronary computed tomography analysis, and therefore do not provide a high-level analysis, have been described previously9,10,11,12,13,14. Although a single neural network has been applied to medical imaging and other medical signals to generate high-level diagnosis16,17,18, these approaches are different from the DeepDiscern system, which provides a coronary diagnostic map by integrating multiple aspects, including the identification of different coronary artery segments, and the recognition of lesion location and lesion type. Because of the challenges of solving the complexities of coronary angiography using a single end-to-end DNN, DeepDiscern uses multiple DNN to solve multiple sub-problems and combines the results of multiple DNN to produce a high-level diagnosis.

With the current global population expansion, the number of patients with cardiovascular disease is increasing, resulting in a growing workload for cardiologists. DeepDiscern is capable of analysing a coronary angiogram in just a few seconds by learning and understanding medical knowledge from massive medical data. DeepDiscern can be used as an assistant to analyse lesion information quickly and help cardiologists to flag and diagnose lesion severity and morphology during the intervention before making a treatment decision. In addition, the amount of medical documentation routinely recorded has grown exponentially and now takes up to a quarter to a half of doctors’ time19. DeepDiscern can generate detailed angiographic reports automatically, saving cardiologists significant time for patient care. Thus, this approach has the potential to reduce workload and improve efficiency in coronary angiography diagnostics.

In clinical practice, visual interpretations of coronary angiograms by individuals are highly variable20. Inevitable subjective bias can have a great impact on diagnosis and treatment decisions21. Unlike the interpretation of cardiologists, the evaluation criteria of DeepDiscern are consistent for the same data set. In addition, deep learning has the ability to extract features automatically in digital angiographic images at a pixel scale, thereby impacting on angiographic interpretation by allowing analysis of angiographic images and identification of lesion features that are hard to discern by the human eye. Thus, the diagnosis of DeepDiscern is intended to be objective, accurate and reproducible.

DeepDiscern could also alleviate the growing problem of unequal distribution of medical resources and access to advanced health care. In 2017, at least half of the world’s population was unable to access essential health services22. In China, the difference between the highest value of healthcare access and quality (HAQ) and the lowest is 43.5 (the highest in Beijing is 91.5 and the lowest in Tibet is 48.0)23. The number of cardiovascular disease patients in China has reached 290 million. There are more than 2,000 primary hospitals in China providing coronary intervention treatment, with levels of diagnosis and treatment differing from place to place. The DeepDiscern technology can be extended easily and rapidly to major hospitals in the country and even the world. Implementing the DeepDiscern technology in primary hospitals countrywide could relieve the high demand for trained cardiologists, who are scarce, and provide consistency for improving angiographic diagnostic accuracy and treatment decisions, thereby achieving homogenisation of medical standards.

In the future, we will develop coronary artery lesion diagnostic systems that analyse more types of lesion morphology such as trifurcation, bifurcation, and severe tortuosity, among others. Thereafter, many decision-making tools based on recognition of lesion morphology and coronary artery segments can be automated without manual discrimination. For example, the SYNTAX score is a decision-making tool in interventional cardiology, which is determined simply by anatomical features in an angiogram24. The automation of SYNTAX score calculation is of great significance for the diagnosis of coronary angiography as it is an important tool for treatment selection (bypass surgery or percutaneous coronary intervention) in patients with more extensive CAD. The expected automatic SYNTAX score calculation system can generate a result in half a minute, detailing all the information about lesions that appear on a patient’s coronary arteries (Supplementary Figure 7).

Limitations

This study has several limitations. In this initial iteration, the input of DeepDiscern was a single frame angiogram obtained from an angiographic DICOM file, which provides limited information compared to a DICOM video. In actual use, after the procedure starts, the video stream of the contrast image is transmitted to our device. We used an automatic algorithm to extract a single frame with optimal contrast opacification and visualisation of the coronary artery tree and then used this single frame image as the input for DeepDiscern. The diagnosis of coronary lesions is based on the dynamic evaluation of lesion characteristics assessed in multiple coronary angiographic views. Additionally, DeepDiscern has a large requirement for training data volume. The lack of training data can seriously affect the recognition accuracy. In general, models trained with more data have better performance. Moreover, all the angiograms used for training and testing were collected from a large single centre; therefore, external validation by using data from other centres is warranted.

Conclusions

Deep learning technology can be used in the interpretation of diagnostic coronary angiography. In the future it may serve as a more powerful tool to standardise screening and risk stratification of patients with CAD.

Impact on daily practice

In clinical practice, inevitable subjective bias can have a great impact on diagnosis and treatment decisions. Unlike interpretation by cardiologists, the evaluation criteria of DeepDiscern are consistent for the same data set. In addition, DeepDiscern can analyse angiographic images and identify lesion features that are hard to discern by the human eye, which can also impact on angiographic interpretation. Thus, the diagnosis of DeepDiscern is intended to be objective, accurate and reproducible.

Funding

The study was supported by Beijing Municipal Science & Technology Commission - Pharmaceutical Collaborative Technology Innovation Research (Z18110700190000) and Chinese Academy of Medical Sciences - Medical and Health Science and Technology Innovation Project (2018-I2M-AI-007).

Conflict of interest statement

B. Xu reports grants from Beijing Municipal Science & Technology, and grants from the Chinese Academy of Medical Sciences during the conduct of the study. In addition, B. Xu has a patent for a method of coronary artery segmentation and recognition based on deep learning pending, and a patent for an automatic detection method system, and equipment of coronary artery disease based on deep learning pending. H. Zhang reports grants from Beijing Municipal Science & Technology during the conduct of the study. The other authors have no conflicts of interest to declare.

Supplementary data

To read the full content of this article, please download the PDF.


References

Volume 17 Number 1
May 16, 2021
Volume 17 Number 1
View full issue


Key metrics

On the same subject

State-of-the-Art Review

10.4244/EIJ-D-21-00145 Sep 20, 2021
Robotics, imaging, and artificial intelligence in the catheterisation laboratory
Beyar R et al
free

Experimental research

10.4244/EIJ-D-20-00648 May 16, 2021
Deep learning for prediction of fractional flow reserve from resting coronary pressure curves
Zimmermann F et al
free

State-of-the-Art

10.4244/EIJ-D-22-00776 Apr 3, 2023
Computed tomographic angiography in coronary artery disease
Serruys PW et al
free

EXPERT REVIEW

10.4244/EIJ-D-17-00465 Aug 25, 2017
State of the art: coronary angiography
Collet C et al
free

Insight

10.4244/EIJ-D-19-00543A May 20, 2020
The SYNTAX score on its way out or … towards artificial intelligence: part I
Serruys PW et al
free
Trending articles
337.88

State-of-the-Art Review

10.4244/EIJ-D-21-00904 Apr 1, 2022
Antiplatelet therapy after percutaneous coronary intervention
Angiolillo D et al
free
283.98

State-of-the-Art Review

10.4244/EIJ-D-21-00695 Nov 19, 2021
Transcatheter treatment for tricuspid valve disease
Praz F et al
free
226.03

State-of-the-Art Review

10.4244/EIJ-D-21-00426 Dec 3, 2021
Myocardial infarction with non-obstructive coronary artery disease
Lindahl B et al
free
209.5

State-of-the-Art Review

10.4244/EIJ-D-21-01034 Jun 3, 2022
Management of in-stent restenosis
Alfonso F et al
free
168.4

Expert review

10.4244/EIJ-D-21-00690 May 15, 2022
Crush techniques for percutaneous coronary intervention of bifurcation lesions
Moroni F et al
free
150.28

State-of-the-Art

10.4244/EIJ-D-22-00776 Apr 3, 2023
Computed tomographic angiography in coronary artery disease
Serruys PW et al
free
103.48

Expert consensus

10.4244/EIJ-E-22-00018 Dec 4, 2023
Definitions and Standardized Endpoints for Treatment of Coronary Bifurcations
Lunardi M et al
free
X

The Official Journal of EuroPCR and the European Association of Percutaneous Cardiovascular Interventions (EAPCI)

EuroPCR EAPCI
PCR ESC
Impact factor: 6.2
2022 Journal Citation Reports®
Science Edition (Clarivate Analytics, 2023)
Online ISSN 1969-6213 - Print ISSN 1774-024X
© 2005-2024 Europa Group - All rights reserved