Abstract
For intravascular OCT (IVOCT) images, we developed an automated atherosclerotic plaque characterization method that used a hybrid learning approach, which combined deep-learning convolutional and hand-crafted, lumen morphological features. Processing was done on innate A-line units with labels fibrolipidic (fibrous tissue followed by lipidous tissue), fibrocalcific (fibrous tissue followed by calcification), or other. We trained/tested on an expansive data set (6,556 images), and performed an active learning, relabeling step to improve noisy ground truth labels. Conditional random field was an important post-processing step to reduce classification errors. Sensitivities/specificities were 84.8%/97.8% and 91.4%/95.7% for fibrolipidic and fibrocalcific plaques, respectively. Over lesions, en face classification maps showed automated results that agreed favorably to manually labeled counterparts. Adding lumen morphological features gave statistically significant improvement (pā<ā0.05), as compared to classification with convolutional features alone. Automated assessments of clinically relevant plaque attributes (arc angle and length), compared favorably to those from manual labels. Our hybrid approach gave statistically improved results as compared to previous A-line classification methods using deep learning or hand-crafted features alone. This plaque characterization approach is fully automated, robust, and promising for live-time treatment planning and research applications.
Similar content being viewed by others
Introduction
Intravascular optical coherence tomography (IVOCT) is an important technology for planning and assessment of interventional, percutaneous treatments of coronary artery disease. IVOCT is a high contrast, high-resolution imaging modality that uses near-infrared light1. Compared to intravascular ultrasound (IVUS), this modality provides better image resolution with axial resolution ranging from 12 to 18 Ī¼m (as compared to 150ā250 Ī¼m from IVUS) and lateral resolution ranging from 20 to 90 Ī¼m (as compared to 150ā300 Ī¼m from IVUS)1. IVOCT allows to determine different plaque components such as fibrous, lipidous, and calcified tissues, and is the only modality that can identify thin cap fibroatheroma2. IVOCT is used for clinical, live-time intervention planning, and stent deployment assessment. IVOCT-guided percutaneous coronary intervention (PCI) brings valuable benefit for patient treatment as compared to PCI guided by X-ray angiography alone3. In addition, IVOCT is used for clinical research studies such as the calcium scoring analysis4 and calcium crack formation5.
Although IVOCT is clearly an excellent method for intravascular imaging of plaque, it has limitations. One is the cost of transducers. Another is tissue penetration depth, especially in the presence of lipidous plaque. Another is the need for a physician trained in visual interpretation of IVOCT images who is willing to take the time to examine images during a stressful procedure. A single IVOCT pullback typically generates 300ā500 image frames resulting in data overload. Even when IVOCT is used for research purposes, manual labeling of plaque is labor-intensive, requiring up to 1 day for each pullback by a trained cardiologist. Manual labeling is also prone to inter- and intra-observer variability. In prior work6, our group found evidence of up to 5% intra- and 6% inter-observer variability among cardiologists detecting stent struts in IVOCT images. To address the limitation of visual interpretation, researchers are creating automated software.
There are reports of fully automated plaque characterization approaches based on machine and deep learning methods. Ughi et al.7 suggested a fully-automated plaque characterization method using the statistical features and random forest classifier in IVOCT images. They applied the pre-processing step to remove the guidewire and to segment the lumen in (r, Īø) IVOCT images. Classification accuracies were 89.5%, 79.5%, and 72.1% for fibrotic, lipid-rich, and calcified tissues, respectively. Athanasiou et al.8 automatically detected four plaque types using a set of intensity and texture-based features with the random forest classifier. Classification sensitivities were 81%, 71%, 87%, and 81% for calcium, lipid tissue, fibrous tissue, and mixed tissue, respectively. Rico-Jimenez et al.9 extracted the 11 morphological features including the A-line peaks and signal decay, and automatically classified each A-line as one of either intimal thickening, fibrotic, superficial lipid, and fibrotic-lipid based on a linear discriminant analysis algorithm. Some groups have reported that the optical attenuation coefficient7,10,11,12 of the tissue is useful for characterizing coronary plaques. Recently, some researchers have attempted to implement deep learning techniques for plaque characterization. Yong et al.13 proposed a linear regression convolutional neural network (CNN) model to automatically segment the vessel lumen. Their network consisted of four convolutional layers and three fully connected layers, including the output layer. They evaluated their method on 64 pullbacks acquired from 28 patients and achieved mean locational accuracy of 22 microns with very fast computation time (40.6āms per image). To automatically determine vessel borders and coronary plaques (e.g., fibrosis and calcification), Abdolmanafi et al.14,15 used the pre-trained CNN models (e.g., AlexNet16, VGG-1917, and Inception-v318) for extracting deep learning features and combined those features to train the machine learning model. He et al.19 extracted the plaque tissue area and employed a simple CNN model to automatically classify each pixel into one of the five categories: lipid tissue, fibrous tissue, mixed tissue, calcified tissue, and background. Athanasiou et al.20 classified the whole arterial wall into six tissue types: calcium, lipid tissue, fibrous tissue, mixed tissue, non-pathological tissue, and no visible tissue using a CNN model having 45 layers. An overall accuracy was 96.1%. More recently, Gessert et al.21 used two state-of-the-art deep learning models, ResNet50-V222 and DenseNet12123 for automated plaque detection. The ResNet was used for a better gradient flow and improved optimization, and the DenseNet was focused on efficiency of classification. Overall accuracy and sensitivity were 91.7% and 90.9% with high F1 score (0.913). Zhang et al.24 proposed an automated plaque segmentation method based on the simple CNN and an improved random walk algorithm. They obtained the same Jaccard coefficient of 0.864 for lipid and calcified plaque segmentation. Abdolmanafi et al.25 used the VGG-19 model17 as the encoder of the fully convolutional network to identify the calcification, fibrosis, macrophage, and neovascularization. An overall sensitivity of 89.3% was reported. Most recently, our group developed a fully automated plaque characterization method for lipidous and calcified plaques based on semantic segmentation using the SegNet deep learning model26. We obtained the sensitivities/specificities of 87.4%/89.5% and 85.1%/94.2% for lipidous and calcified plaques, respectively. We also provided A-line classification maps based on which plaque was in the majority in each A-line. In a previous report, we obtained good results using A-line classification with a CNN27. In another report, we used a conventional machine learning approach and evaluated over 1,000 features, including a set of lumen morphology features, for classification of A-lines28. This variety of publications suggests that plaque classification is an important problem and that there are a variety of different approaches that should be evaluated. Some of these details are summarized in Table S1.
In this paper, we build on our previous studies and develop an automated, A-line-based plaque characterization method by combining A-line deep learning and hand-crafted, lumen morphology features. Such hybrid learning approaches, which combine deep learning and hand crafted features, have been successfully applied to multiple problems29,30,31,32. Each A-line will be classified as fibrolipidic (fibrous tissue at the lumen border followed by lipidous tissue), fibrocalcific (fibrous tissue followed by calcification), or other. We perform image processing procedures in the polar (r, Īø) domain to avoid interpolation effects seen in the Cartesian (x, y) domain. In addition to conventional quantitative classification metrics, we measure clinically meaningful plaque attributes (arc angle and length). We limit our classification targets to lipidous (fibrolipidic) and calcified (fibrocalcific) plaques, since these types of plaques influence clinical decision making33,34,35.
This study has several important aspects and contributions. First, A-line classification avoids the issue of the indeterminate back border of lipidous plaque found in pixel classification methods. That is, because there is rapid absorption of light in lipidous tissues, one cannot determine the location of the back boarder. Second, we create a powerful, hybrid classification which combines deep learning A-line features and hand-crafted lumen morphological features. The latter exploits the observation that non-circular lumens are often observed in the presence of calcifications28. Essentially, we are combining two methods previously reported by us: A-line CNN classification27 and lumen morphological features which were identified as important among 1000ās of hand-crafted features previously reported28. Third, we use a large number of IVOCT images and corresponding labels, including a wide variety of lesions (i.e., over 6,500 images, 49 patients, and 111 volumes of interest (VOIs)). Having a large variety of lesions is important to ensure generalization. Fourth, in addition to conventional segmentation metrics (e.g., F1 score), we measure clinically meaningful plaque attributes (arc angle and length) for performance evaluation. Fifth, we apply an active learning experiment to possibly reduce inconsistent labeling, which could arise due to analyst fatigue or criteria drift. In this, a highly trained, independent analyst is presented with a number of images containing discrepancies and asked to label them again, blind to previous labels and computer predictions.
Image Analysis Methods
Our A-line plaque classification method involved several steps. First, we preprocessed (r, Īø) data to identify the tissue region of interest and to reduce noise. Second, deep learning CNN features, optimized for A-line classification, were extracted for each A-line. Third, lumen morphological features were extracted from both (r, Īø) and (x, y) data. Fourth, a random forest classifier, which used a hybrid of deep learning and hand-crafted features, was used to classify A-lines as fibrocalcific, fibrolipidic, or other. Fifth, CRF was used to reduce classification errors. Learning was done using standard approaches.
Pre-processing
Pre-processing of (r, Īø) image data consisted of several steps, principally to identify the appropriate tissue region of interest for processing. First, to remove the guidewire and its shadow regions, we created an accumulated intensity map by adding all the pixel values along r in each A-line, as the guidewire usually generates a long dark shadow along the radial direction. Then, we detected the shadow regions using dynamic programming36 and removed it. Second, we identified the lumen boundary using dynamic programming36. Briefly, we computed the image gradient and regarded the contour with the highest cumulative edge strength as the lumen boundary. Third, each A-line was pixel-shifted to the left (smaller r) allowing all A-lines to have the same starting pixel. Fourth, we designated the first 200 pixels (~1āmm) in the r direction as the region of interest for processing, due to the limited penetration depth of IVOCT. Finally, we applied a Gaussian filter to reduce noise, using a kernel with standard deviation 1 and footprint of (7, 7).
Feature extraction
Extraction of convolutional features
We created a CNN to classify A-lines, trained it, and then used the network to compute features from IVOCT images for machine learning, random forest processing. The CNN was trained to classify A-lines as fibrocalcific, fibrolipidic, or other. The CNN model consisted of three convolutional, two maximum pooling, and three fully connected layers (Fig. 1). Each convolutional layer consisted of convolutional, batch normalization, and rectified linear unit (ReLU) layers. Before processing, we padded the left and right edges of the input A-line by 5 pixels (Layer 1). The left padding was set to zero because those regions belong to the luminal area (an anatomical feature of interest), while the five pixels after region of interest (201ā205 pixels in r axis) were used for the right padding. In our model, convolutional layers had a varying number of filters (32, 64, and 96) and kernel size (11, 9, and 7) and generated the relevant feature maps with a stride of 2 pixels. After convolutional operation, batch normalization and ReLU layers were added to accelerate training process and reduce the sensitivity to network initialization16. The maximum pooling layer with a pool size of 2 pixels was subsequently implemented for dimensionality reduction. This layer enabled to prevent overfitting, and kept the network invariant to small transformations, distortions, and translations16. The three fully connected layers were followed by pairs of convolutional and pooling layers. The first two layers included 100 output units with ReLU activation functions and dropout layers, while the Softmax activation was used along with three output units for the last layer. After training, we removed the last fully connected layer and obtained 100 convolutional features from the activations of the previous fully connected layer. These features were later combined with lumen morphology features.
Lumen morphology features
We extracted lumen morphological features using methods described previously28, which are now only briefly described. We estimated the change in lumen area by analyzing the lumen area of current frame to each of its adjacentāĀ±ā3 frames. As shown in Fig. 2A, we computed the lumen eccentricity of current frame as well as the change in lumen eccentricity across adjacentāĀ±ā3 frames. Individual A-line lumen eccentricity was also obtained to localize A-lines containing plaque (Fig. 2B). To determine this, we created a concentric circle having the same center of mass and area as the segmented lumen. Then, we measured the signed distance between the boundary pixel of each A-line and the corresponding circle boundary pixel along the radius from the center of mass. In addition, we found that large calcifications often tend to cause a flat lumen in our dataset, especially when they are close to the vessel wall. Therefore, we generated a best-fit line for lumen boundary (Fig. 2C) and extracted relevant features: sum of squared residuals, goodness of fit, and line magnitude. For the previously described lumen features, we also observed the change in these features across adjacentāĀ±ā3 frames. Furthermore, because of the physics of light reflection, a signal-poor region with diffuse outer border can appear, when the IVOCT signal is obliquely incident on the tissue surface37. To help identify such regions, we computed the R-Īø lumen curvature (Fig. 2D) by assessing the lumen slope, where a slope of 0 indicates perpendicular incidence of the beam, and a large positive or negative value indicates an oblique intersection. In total, we extracted 371 morphological features in IVOCT images. The details of these features are referred in28.
Random forest classification
We used the random forest classifier as it has several advantages for the analysis of IVOCT images. First, this method works efficiently on a large number of input attributes (features) with a low risk of overfitting. Second, it is fairly robust for noisy data. Third, it can provide the relative importance of features in determining classification, thus requiring no separate feature selection step. Fourth, learning and computation time is fast, even for very large data sets. The numbers of trees and variables are the two most important parameters for determining classification performance of random forest classifier38. The optimum numbers of classification trees and variables were determined by considering the out-of-bag error (OOB) data which are not used in bootstrap sampling. Implementation details are described in Section 3.2.
Classification noise reduction using fully connected conditional random field (CRF)
Since A-line classification does not fully consider spatial similarities between nearby A-lines in Īø and z, we employed classification noise cleaning. In particular, we implemented a fully connected CRF that defines the pairwise edge potentials of each probabilistic classification (Īø, z) by a linear combination of Gaussian kernels39. CRF is not only able to model arbitrarily large pixel-neighborhoods, but also computationally efficient, making it ideal for IVOCT plaque characterization. This method plays a very important role in performance improvement as it enables us to consider the spatial similarities.
In a fully connected CRF, there are two Gaussian kernel potentials, which are smoothness and appearance kernels:
where p and I are the spatial position and the intensity vectors, respectively. The configurable parameter ĪøĪ± expresses the size and shape of neighborhoods. The parameters ĪøĪ² and ĪøĪ³ control the degrees of nearness and similarity. The configurable weights Ļ1 and Ļ2 define the relative strength of the two factors. The smoothness kernel (first term) can be expressed by a diagonal covariance matrix and removes small isolated regions in the predicted results40. The appearance kernel (second term) performs optimization based on the observation that adjacent pixels with similar intensity values are likely to be in the same class. Since pixels in the en face image have no specific intensity value, the appearance kernel was dropped by setting Ļ2 to zero. Therefore, our implementation of CRF only included the smoothness kernel. Detailed descriptions and equations are referred in39.
Experimental Methods
Images and their annotation
IVOCT images were acquired with a frequency-domain OCT system (ILUMIEN OPTIS; St. Jude Medical Inc.). During image acquisition, the optical probe was automatically pulled back from distal to proximal at a speed of 36āmm/s, frame rate of 180 frames/sec, and axial resolution of 20 Ī¼m. All IVOCT image frames were initially reviewed, and frames having poor quality due to luminal blood, unclear lumen, artifact, or reverberation were excluded. We used a total of 6,556 image frames across 49 patients with 111 VOIs having calcification (34 VOIs), lipidous regions (37 VOIs), or both (13 VOIs), and 27 VOIs without any identified calcification or lipidous region. The sizes of raw (r,Īø) data before processing were either 968āĆā496 or 968āĆā448 pixels in 16-bit gray scale. We used every IVOCT pullback that was considered analyzable by cardiologists. The images with stents and poor quality were excluded to avoid misleading results.
Each A-line from each VOI was manually annotated as now described. Log compressed, Cartesian (x,y) images were analyzed by two experienced readers from Cardiovascular Imaging Core Laboratory of Harrington Heart and Vascular Institute (University Hospitals Cleveland Medical Center, Cleveland, OH, USA), a laboratory that has done numerous studies requiring expert reading of IVOCT images. Using definitions in the āconsensus documentā2, they manually labeled regions containing lipidous and calcified plaques. Tissue pixels without annotation were considered others. In case of disagreement, readers revaluated the frames and reached a consensus decision. From a simple rule, we next created A-line labels, consisting of fibrolipidic, fibrocalcific, and other, as reported previously27,28. If an A-line included ā„3 pixels of either lipidous or calcified plaques, we determined which of these two classes was in the majority, and then labeled the A-line accordingly as fibrolipidic or fibrocalcific. All other A-lines were identified as being in the āotherā class. Figure 3 shows an example annotation containing lipidous (fibrolipidic), calcified (fibrocalcific), and other classes. Examples of lipidous and calcified plaques are shown in IVOCT images and corresponding cryo-images in Fig. S1 in the Supplementary Material.
Implementation and optimization of the CNN and random forest
The CNN and random forest were trained on the exact same VOIs from the five-fold cross validation. For the CNN classifier, we minimized a cross-entropy loss function over the softmax outputs. Although the CNN model was not used for final classification, overfitting could still degrade the reliability of convolutional features. To limit overfitting, we set the training to stop when the validation loss did not improve over 5 consecutive epochs, or when the number of epochs reached a pre-defined value (50). The former stopping criterion was typically exercised. Because data were imbalanced, we modified the loss function to include relative class weights of each class as determined from the inversed median frequency of class proportions. The CNN model was trained using adaptive moment estimation (ADAM) optimizer41. ADAM optimization is an extension to stochastic gradient descent optimizer that only requires first-order gradients with little memory requirement. This method computes individual adaptive learning rates for different parameters and is well-suited to a non-convex optimization problems41. All parameters of the CNN model were initialized with random Gaussian distributions. The model was trained with a batch size of 30 for 50 epochs, momentum with decay of 0.9, base learning rate of 0.001, learning rate drop factor of 0.2, and learning rate drop period of 5. The size of the receptive field of CNN model was 77āĆā1 pixels. The training progress curve on the one-fold cross validation data set is shown in Fig. S2 in the Supplementary Material. The random forest classifier was optimized. We varied the number of trees from 100 to 1,000 in intervals of 50 and compared OOB error rates. The error rate decreased to a nearly constant value at about 250 trees (Fig. S3), and we set the number of trees to 250 in all other analyses. Similarly, the number of variables was empirically identified to be 16. Image processing and network training were performed using MATLAB software package (R2018b, MathWorks Inc.) on a NVIDIA GeForce TITAN RTX GPU with 120GB of RAM installed in a Dell Precision T7610.
Performance evaluation
To evaluate classification performance and its variation across samples, we used five-fold cross validation. Among the entire data sets, 89 VOIs were divided into five independent subsets, consisting of 15ā22 VOIs each, and during each of the 5 folded evaluations, 3 were assigned for training, 1 for validation, and 1 for testing. This process was repeated five times such that all pullbacks were used in a test set. We determined subsets based on VOIs rather than images, as the latter can leave very similar images in training and testing sets, giving unrealistic performance. The rest (22 VOIs) were used as held-out data for further evaluations. These images were neither used for training nor testing during cross validation. When we compared processing methods, the exact same folds were utilized. Classification performance was quantitatively evaluated using traditional classification metrics as given below:
Above, TP is the number of true positives, TN is the number of true negatives, FP is the number of false positives, and FN is the number of false negatives. In addition, we also measured clinical plaque attributes, such as arc angle and length, to provide a more clinically meaningful analysis (Fig. 4). The arc angle was measured from the center of the catheter to the boundaries of plaque A-lines. The length was the total number of frames containing a contiguous plaque multiplied by the frame interval. These attributes were evaluated using Bland-Altman analysis.
Active learning relabeling
The ground truth labels of IVOCT images used in this study were manually created by two expert readers. Although they are consistently trained and had similar experience, there could be inconsistent criteria applied over time. Therefore, we performed an active learning experiment as follows. Following training on all data, we selected the 2 folds having a low mean F1 score (<0.7) and asked an experienced cardiologist to relabel them. During relabeling, the experienced cardiologist was completely blinded to previous manual and automatic labels. After relabeling, we performed the five-fold cross validation again to determine any improvements.
Results
There was processing prior to classification. Tissue regions were identified for further processing (Fig. 5). The lumen boundary (red) was accurately segmented, and guidewire and corresponding shadow regions were removed (Fig. 5A). The result was pixel shifted to the left (Fig. 5B). This normalized lesion appearance across training data. The region of interest was determined and filtered to reduce speckle noise (Fig. 5C). The latter image was used as the input to the CNN. The CNN was trained to classify A-lines as part of the train/test/validate process across the five folds. Once trained, the CNN was stripped down to create features for further hybrid processing.
Hybrid classification followed by CRF gave good A-line classification results. Figure 6 demonstrates a representative classification result from one of the folds. Collecting results across the five folds, most traditional metrics (i.e., sensitivity and F1 score) were significantly improved for all classes after classification noise cleaning (pā<ā0.05) (Table 1). The proposed method gave better classification results for the fibrolipidic class (F1 score ā 0.887) as compared to those in fibrocalcific class (F1 score ā 0.677).
Active learning improved results. To reduce classification errors, a most experienced cardiologist very carefully reexamined the calcified plaque lesions of two folds (33 VOIs, 1,332 images), with low mean F1 scores (<0.7). Please note that relabeling was done in a very conservative manner. The experienced cardiologist was naĆÆve to automated segmentation results and did not see automated or initial manual annotations during this labeling step. In the 33 VOIs and 1,332 images, most changes between initial and secondary labels were at the edges of a lesion. The experienced cardiologist added calcifications to 16 images not previously identified as having calcification. The mean F1 score of fibrocalcific plaque was improved from 0.677 to 0.719 (Table 1), a changed determined to be significant (pā<ā0.05). There was only a small, insignificant difference in the fibrolipidic and other classes. Because we now had initial and secondary labeling on two folds, this enabled additional analyses. Assuming the secondary labeling to be the gold standard, we determined an F1 score for fibrocalcific of 0.965 and 0.937 for the two folds. Comparing the automated test results from these two folds, we found similar scores (i.e., 0.912 and 0.981, respectively).
To further evaluate processing, we trained our hybrid method using the entire 89 VOIs (4,819 images) and tested on the held-out data (22 VOIs, 1,737 images). Performance was similar to that in the fold analysis (Tables 1 and 2), i.e., nearly within the standard deviation of the foldsā analysis. Fibrocalcific plaque showed higher sensitivity compared to fibrolipidic plaque, while the F1 score and specificity were higher in fibrolipidic class. Since plaques are spatially distributed in coronary artery, we created the en face (Īø, z) map of the predicted A-line results (Fig. 7). The predicted classification maps agreed favorably to the manually annotated counterparts. The confusion matrix is shown in Fig. S4 in the Supplementary Material.
On the held-out data set, we determined the effect of adding hand-crafted morphological features by comparing the CNN classifier to the hybrid classifier. In Fig. 8, the convolutional features showed numerous false negatives, especially for fibrocalcific plaque as compared to the ground truth. Lumen morphology features were hardly able to detect fibrocalcific plaque (F1 scoreā=ā0.011). In Table 2, the hybrid classifier significantly outperformed the CNN classifier, indicating the importance of hand-crafted morphological features. Improvements were obtained both for fibrocalcifican and fibrolipic A-lines.
Bland-Altman analysis was performed to evaluate the statistical agreement of measured clinical plaque attributes (i.e., arc angle and length) between manual and automated methods. In Fig. 9, the fibrolipidic plaque had much lower biases of 6.5āĀ±ā19.2Ā° (angle) and ā8.7āĀ±ā11.3āmm (length) as compared to the fibrocalcific plaque having 13.2āĀ±ā16.8Ā° (angle) and 10.8āĀ±ā20.1āmm (length). Although the fibrocalcific class had a relatively large bias, most of points were included in the limit of agreement. These promising results indicate that the proposed method is well-suited for plaque characterization in IVOCT images.
The proposed hybrid method was compared to our two previous A-line-based classification approaches which used deep learning27 and hand crafted features28 alone. We trained/tested the previous approaches on the exact same data sets (five folds for training and held-out for testing) used in this study. As shown in Table 3, the proposed method was statistically different and better than the previous methods for all plaque types.
Discussion
These first results of hybrid classification/segmentation are promising for clinical treatment planning and research applications. With hybrid features, sensitivities/specificities of fibrolipidic and fibrocalcific plaques over the five folds were 84.8%/97.8% and 91.2%/96.2%, respectively, with substantial F1 score improvement as compared to results before noise cleaning (Table 1). Results from our hybrid method were improved as compared to our previous reports of A-line-based classification using deep learning27 and hand crafted features28 alone. Kolluru et al.27 classified A-lines using a deep learning CNN model similar to the one used in this study. Morphological features were not available to the network due to the use of pixel-shifting. In another study, Prabhu et al.28. analyzed lumen morphology features and for A-line plaque characterization. The current hybrid approach produced very promising results sufficiently quickly (as detailed later) to suggest appropriateness for both research and treatment planning application.
There are important aspects of our hybrid A-line classification/segmentation approach. As described in the Introduction, A-line classification uses the innate, underlying data element in an IVOCT acquisition and incorporates the natural ordering of tissues from the lumen out (e.g., fibrous tissue followed by lipidous tissue). Pre-processing (particularly pixel-shifting) plays a very important role for improved plaque characterization, as reported previously26. However, with pixel shifting, the CNN cannot assess lumen border irregularities (as often found in the presence of calcifications) and physics-of-reflection-based attributes. To address this limitation, we extracted hand-crafted, lumen morphology features and combined with the convolutional features. Comparing results between hybrid, lumen morphology only, and deep learning convolutional only classifications (Fig. 8 and Table 2), we found much improved results with hybrid. We found the random forest classifier to be useful in this application. In a preliminary experiment, we compared classification results between decision tree, random forest, and support vector machine methods (not shown). Decision tree had slightly lower metrics compared to the other two. Support vector machine and random forest gave very similar results with the winner depending upon the fold. We chose the latter for a variety of reasons. As compared to support vector machine, random forest was easier to optimize, did not require a separate feature selection step, trained in ā¼1/4 the time, and computed at run time in ā¼1/3 the time. On our computer system with non-optimized code, our current run-time is only about 1-sec per image (0.05-sec for pre-processing, 0.9-sec for feature extraction, 0.03-sec for classification, and 0.02-sec for post-processing). The proposed method can analyze a high resolution IVOCT pullback in a few minutes. By optimizing coding for speed, the method could be suitable for live-time treatment planning.
Post-processing with CRF remarkably improved classification performance. This is consistent with our previous study using a simpler CNN architecture to classify A-lines27. In this case, the overall sensitivity increased by 10ā15% after CRF noise cleaning. We observed a similar improvement of traditional metrics (e.g., sensitivity and F1 score) for both fibrolipidic and fibrocalcific plaques. This could be expected as both approaches utilized a single A-line as an input. The improvement with CRF is likely due to the consideration of spatial similarity.
The active learning, relabeling step provided improved classification results, particularly for fibrocalcific plaque where the F1 score was improved from 0.677 to 0.719, a significant difference (pā<ā0.05). As argued in Results, relabeling was done in a very conservative manner, where the experienced cardiologist had no knowledge of automated or previous manual results. F1 scores between the initial and secondary manual labeling of the two folds (0.965 and 0.937) indicate that overall changes were relatively small. Nevertheless, the relabeling process improved the consistency of labeling, providing the enhanced performance of the automatic scheme for fibrocalcific plaques. This initial experimental result suggests that improvement of noisy labels could lead to additional benefits.
Unfortunately, it is difficult to compare our results to those from previous studies described in the Introduction7,8,9,10,11,12,13,14,15,19,20,21,24,25,26 because performance will greatly depend on the case mix and criteria for annotation. Nevertheless, it appears that our method compares favorably to previous studies (Table S1). For example, our method showed higher sensitivity in fibrocalcific plaque than those in other reports on the held-out data set. We have used the largest dataset reported. Our method had the very short processing time (~1-sec per image).
This study has two limitations. First, manually labeled images were used for training. Although we improved some ground truth labels using active learning, this is still an imperfect gold standard. In the future, this limitation could be reduced with extensive active learning or with large data sets using labeling based upon registered cryo-images42. Second, we used a relatively simple CNN architecture for convolutional feature extraction. Classification results might be improved with more advanced deep learning models.
In summary, we developed a fully automated, A-line-based plaque characterization method based upon hybrid processing (deep learning and hand-crafted feature machine learning). Hand crafted lumen morphology features provide significant improvement to A-line deep learning, suggesting the importance of these features to determination of IVOCT plaque characterization. The method gives very good plaque classification results and computes in a reasonable time, suggesting that this could be a promising solution for both clinical and research applications.
Data availability
The datasets analyzed during the current study are available from the corresponding author on reasonable request.
References
Bezerra, H. G., Costa, M. A., Guagliumi, G., Rollins, A. M. & Simon, D. I. Intracoronary optical coherence tomography: a comprehensive review clinical and research applications. JACC Cardiovasc. Interv. 2, 1035ā1046 (2009).
Tearney, G. J. et al. Consensus standards for acquisition, measurement, and reporting of intravascular optical coherence tomography studies: a report from the International Working Group for Intravascular Optical Coherence Tomography Standardization and Validation. J. Am. Coll. Cardiol. 59, 1058ā1072 (2012).
Prati, F. et al. Angiography alone versus angiography plus optical coherence tomography to guide decision-making during percutaneous coronary intervention: the Centro per la Lotta contro lāInfarto-Optimisation of Percutaneous Coronary Intervention (CLI-OPCI) study. EuroIntervention 8, 823ā829 (2012).
Fujino, A. et al. A new optical coherence tomography-based calcium scoring system to predict stent underexpansion. EuroIntervention 13, e2182āe2189 (2018).
Maejima, N. et al. Relationship Between Thickness of Calcium on Optical Coherence Tomography and Crack Formation After Balloon Dilatation in Calcified Plaque Requiring Rotational Atherectomy. Circ. J. 80, 1413ā1419 (2016).
Lu, H. et al. Automatic stent detection in intravascular OCT images using bagged decision trees. Biomed. Opt. Express 3, 2809ā2824 (2012).
Ughi, G. J., Adriaenssens, T., Sinnaeve, P., Desmet, W. & Dāhooge, J. Automated tissue characterization of in vivo atherosclerotic plaques by intravascular optical coherence tomography images. Biomed. Opt. Express 4, 1014ā1030 (2013).
Athanasiou, L. S. et al. Methodology for fully automated segmentation and plaque characterization in intracoronary optical coherence tomography images. J. Biomed. Opt. 19, 026009 (2014).
Rico-Jimenez, J. J. et al. Automatic classification of atherosclerotic plaques imaged with intravascular OCT. Biomed. Opt. Express 7, 4069ā4085 (2016).
Xu, C., Schmitt, J. M., Carlier, S. G. & Virmani, R. Characterization of atherosclerosis plaques by measuring both backscattering and attenuation coefficients in optical coherence tomography. J. Biomed. Opt. 13, 034003 (2008).
van Soest, G. et al. Atherosclerotic tissue characterization in vivo by optical coherence tomography attenuation imaging. J. Biomed. Opt. 15, 011105 (2010).
Gargesha, M. et al. Parameter estimation of atherosclerotic tissue optical properties from three-dimensional intravascular optical coherence tomography. J. Med. Imaging 2, 016001 (2015).
Yong, Y. L., Tan, L. K., McLaughlin, R. A., Chee, K. H. & Liew, Y. M. Linear-regression convolutional neural network for fully automated coronary lumen segmentation in intravascular optical coherence tomography. J. Biomed. Opt. 22, 1ā9 (2017).
Abdolmanafi, A., Duong, L., Dahdah, N. & Cheriet, F. Deep feature learning for automatic tissue classification of coronary artery using optical coherence tomography. Biomed. Opt. Express 8, 1203ā1220 (2017).
Abdolmanafi, A., Duong, L., Dahdah, N., Adib, I. R. & Cheriet, F. Characterization of coronary artery pathological formations from OCT imaging using deep learning. Biomed. Opt. Express 9, 4936ā4960 (2018).
Krizhevsky, A., Sutskever, I. & Hinton, G. E. ImageNet Classification with Deep Convolutional Neural Networks. in Proceedings of the 25th International Conference on Neural Information Processing Systems - Volume 1 1097ā1105 (Curran Associates Inc., 2012).
Russakovsky, O. et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 115, 211ā252 (2015).
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J. & Wojna, Z. Rethinking the Inception Architecture for Computer Vision. arXiv:1512.00567 [cs] (2015).
He, S. et al. Convolutional neural network based automatic plaque characterization from intracoronary optical coherence tomography images. Medical Imaging 2018: Image Processing 107, https://doi.org/10.1117/12.2293957 (2018).
Athanasiou, L. S., et al A deep learning approach to classify atherosclerosis using intracoronary optical coherence tomography. in Medical Imaging 2019: Computer-Aided Diagnosis vol. 10950 109500N (International Society for Optics and Photonics, 2019).
Gessert, N. et al. Automatic Plaque Detection in IVOCT Pullbacks Using Convolutional Neural Networks. IEEE Trans. Med. Imaging 38, 426ā434 (2019).
He, K., Zhang, X., Ren, S. & Sun, J. Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 770ā778, https://doi.org/10.1109/CVPR.2016.90 (2016).
Huang, G., Liu, Z., Maaten, L. V D & Weinberger, K. Q. Densely Connected Convolutional Networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2261ā2269, https://doi.org/10.1109/CVPR.2017.243 (2017).
Zhang, H. et al. Automatic Plaque Segmentation in Coronary Optical Coherence Tomography Images. International Journal of Pattern Recognition and Artificial Intelligence, https://doi.org/10.1142/S0218001419540351 (2019).
Abdolmanafi, A., Cheriet, F., Duong, L., Ibrahim, R. & Dahdah, N. An automatic diagnostic system of coronary artery lesions in Kawasaki disease using intravascular optical coherence tomography imaging. Journal of Biophotonics 0, e201900112.
Lee, J. et al. Automated plaque characterization using deep learning on coronary intravascular optical coherence tomographic images. Biomed. Opt. Express, BOE 10, 6497ā6515 (2019).
Kolluru, C. et al. Deep neural networks for A-line-based plaque classification in coronary intravascular optical coherence tomography images. J. Med. Imaging 5, 044504 (2018).
Prabhu, D. et al. Automated A-line coronary plaque classification of intravascular optical coherence tomography images using handcrafted features and large datasets. J. Biomed. Opt. 24, 1ā15 (2019).
Agajanian, S., Oluyemi, O. & Verkhivker, G. M. Integration of Random Forest Classifiers and Deep Convolutional Neural Networks for Classification and Biomolecular Modeling of Cancer Driver Mutations. Front Mol Biosci 6, (2019).
Nijhawan, R., Das, J. & Raman, B. A hybrid of deep learning and hand-crafted features based approach for snow cover mapping. Int. J. Remote. Sens. 40, 759ā773 (2019).
Wu, H., Huang, Q., Wang, D. & Gao, L. A CNN-SVM combined model for pattern recognition of knee motion using mechanomyography signals. J. Electromyography Kinesiol. 42, 136ā142 (2018).
Yu, C. et al. Breast cancer classification in pathological images based on hybrid features. Multimed. Tools Appl. 78, 21325ā21345 (2019).
Imola, F. et al. Association Between Proximal Stent Edge Positioning on Atherosclerotic Plaques Containing Lipid Pools and Postprocedural Myocardial Infarction (from the CLI-POOL Study). Am. J. Cardiology 111, 526ā531 (2013).
Gonzalo, N. et al. Relation between plaque type and dissections at the edges after stent implantation: An optical coherence tomography study. Int. J. Cardiology 150, 151ā155 (2011).
Vavuranakis, M. et al. Stent deployment in calcified lesions: can we overcome calcific restraint with high-pressure balloon inflations? Catheter. Cardiovasc. Interv. 52, 164ā172 (2001).
Wang, Z. et al. Semiautomatic segmentation and quantification of calcified plaques in intracoronary optical coherence tomography images. J. Biomed. Opt. 15, 061711 (2010).
Soest, Gvan et al. Pitfalls in Plaque Characterization by OCT: Image Artifacts in Native Coronary Arteries. JACC: Cardiovascular Imaging 4, 810ā813 (2011).
Wright, M. N. & Kƶnig, I. R. Splitting on categorical predictors in random forests. PeerJ 7, (2019).
KrƤhenbĆ¼hl, P. & Koltun, V. Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials. In Advances in Neural Information Processing Systems 24 (eds. Shawe-Taylor, J., Zemel, R. S., Bartlett, P. L., Pereira, F. & Weinberger, K. Q.) 109ā117 (Curran Associates, Inc., 2011).
Shotton, J., Winn, J., Rother, C. & Criminisi, A. TextonBoost for Image Understanding: Multi-Class Object Recognition and Segmentation by Jointly Modeling Texture, Layout, and Context. Int. J. Comput. Vis. 81, 2ā23 (2009).
Kingma, D. & Ba, J. Adam: A Method for Stochastic Optimization. International Conference on Learning Representations (2014).
Prabhu, D. et al. Three-dimensional registration of intravascular optical coherence tomography and cryo-image volumes for microscopic-resolution validation. J Med Imaging (Bellingham) 3, (2016).
Acknowledgements
This project was supported by the National Heart, Lung, and Blood Institute through grants NIH R21HL108263, NIH R01HL114406, and NIH R01HL143484. This research was conducted in space renovated using funds from an NIH construction grant (C06 RR12463) awarded to Case Western Reserve University. The content of this report is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The grants were obtained via collaboration between Case Western Reserve University and University Hospitals of Cleveland. This work made use of the High-Performance Computing Resource in the Core Facility for Advanced Research Computing at Case Western Reserve University. The veracity guarantor, Yazan Gharaibeh, affirms to the best of his knowledge that all aspects of this paper are accurate.
Author information
Authors and Affiliations
Contributions
J.L. was a major contributor in developing the methods and writing the manuscript. D.P. participated in the initial labeling and helped the comparison. C.K. helped to draft the manuscript and to analyze results. Y.G. participated in developing the methods. V.N.Z. helped to perform the active learning and L.A.P.D. helped to perform the statistical analysis. H.G.B. helped to design and validate the methods. D.L.W. supervised the research and helped to draft the manuscript.
Corresponding author
Ethics declarations
Competing interests
Dr. Bezerra has received consulting fees from Abbott Vascular.
Additional information
Publisherās note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the articleās Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the articleās Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lee, J., Prabhu, D., Kolluru, C. et al. Fully automated plaque characterization in intravascular OCT images using hybrid convolutional and lumen morphology features. Sci Rep 10, 2596 (2020). https://doi.org/10.1038/s41598-020-59315-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-020-59315-6
This article is cited by
-
Deep learning segmentation of fibrous cap in intravascular optical coherence tomography images
Scientific Reports (2024)
-
Roadmap on the use of artificial intelligence for imaging of vulnerable atherosclerotic plaque in coronary arteries
Nature Reviews Cardiology (2024)
-
Prediction of stent under-expansion in calcified coronary arteries using machine learning on intravascular optical coherence tomography images
Scientific Reports (2023)
-
Automated analysis of fibrous cap in intravascular optical coherence tomography images of coronary arteries
Scientific Reports (2022)
-
Ovarian cancer detection using optical coherence tomography and convolutional neural networks
Neural Computing and Applications (2022)