Article | Open

Enhancement of Perivascular Spaces in 7 T MR Image using Haar Transform of Non-local Cubes and Block-matching Filtering

  • Scientific Reports 7, Article number: 8569 (2017)
  • doi:10.1038/s41598-017-09336-5
  • Download Citation
Received:
Accepted:
Published online:

Abstract

Perivascular spaces (PVSs) in brain have a close relationship with typical neurological diseases. The quantitative studies of PVSs are meaningful but usually difficult, due to their thin and weak signals and also background noise in the 7 T brain magnetic resonance images (MRI). To clearly distinguish the PVSs in the 7 T MRI, we propose a novel PVS enhancement method based on the Haar transform of non-local cubes. Specifically, we extract a certain number of cubes from a small neighbor to form a cube group, and then perform Haar transform on each cube group. The Haar transform coefficients are processed using a nonlinear function to amplify the weak signals relevant to the PVSs and to suppress the noise. The enhanced image is reconstructed using the inverse Haar transform of the processed coefficients. Finally, we perform a block-matching 4D filtering on the enhanced image to further remove any remaining noise, and thus obtain an enhanced and denoised 7 T MRI for PVS segmentation. We apply two existing methods to complete PVS segmentation, i.e., (1) vesselness-thresholding and (2) random forest classification. The experimental results show that the PVS segmentation performances can be significantly improved by using the enhanced and denoised 7 T MRI.

Introduction

The perivascular space (PVS) is the normal anatomical structure filled with cerebrospinal fluid (CSF) in the nerve system. It was initially identified by a German pathologist, R. Virchow, and a French biologist, C.P. Robin, more than a century ago. The PVS, which is also known as the perivascular lymphatic space, is related to brain physiological and immune functions1,2,3,4. For example, the diameter of PVSs is typically less than 2mm in all age groups of healthy people. The abnormal enlargement of the diameter or the increase of the PVS number is related to aging5,6,7, cognitive degeneration8, and vessel diseases6, 9, 10. Although PVS has drawn more interest in recent years, the traditional 1.5 T and 3 T magnetic resonance (MR) scanners can hardly capture the thin structures of PVSs precisely. The new-generation 7 T MR scanner, on the contrary, owns much better capability of rendering PVSs10. However, it is still difficult to clearly distinguish the thin PVSs from the noisy background in the 7 T MR images. Figure 1(a) shows the examples of PVSs in a 7 T MR image, where those relatively bright voxels within the red dotted contours are the PVSs. We can see that the PVSs are the thin structures with weak signals in the 7 T MR image.

Figure 1
Figure 1

(a) Illustration of PVSs in a 7 T MR image. Those relatively bright voxels within the red dotted contours are the PVSs. (b) Enhanced image by using spatial correlation filtering; and (c) enhanced image by using our proposed method. (df) Show the denoised images of (ac), respectively. For these three cases, the BM4D method uses the same parameters to denoise the images.

There are several reports that focus on the identification or segmentation of PVSs. For example, Wuerfel et al.11 developed a semi-automatic software that can adjust the intensity threshold for segmenting PVSs by optimizing a predefined PVS model through Markov Chain Monte Carlo. Uchiyama et al.12 used the white top-hat transform to enhance the tubular structures of PVSs, then extracted them according to intensity thresholding, and then finally identified PVSs using the geometric properties. Ramirez et al.13, 14 proposed to segment PVSs in a semi-automatic way by determining the intensity thresholds from T2-weighted and proton density (PD) images adaptively. Wang et al.15 also proposed a semi-automatic method that can adaptively adjust the intensity threshold by using gamma correction and linear mapping. Recently, Park et al.16 proposed a learning-based automatic segmentation method by training a sequence of random forest classifiers with the orientation-normalized 3D Haar features. Then, the segmentation is attained through the sequential classifiers, followed by vesselness thresholding. Although these existing methods can effectively use various characteristics of PVSs, the segmentation performance is still limited due to image quality. In particular, it is difficult to extract informative features of the thin and weak PVS structures from the noisy background.

Accordingly, it is necessary to enhance image quality to effectively distinguish the PVS structures. A common image enhancement is through spatial correlation filtering, which performs convolution operation on every voxel. However, the convolution operation using a single kernel cannot adaptively amplify the PVS signals and also, at the same time, suppress the noises (see Fig. 1(b)). To effectively enhance the specific PVS-like structures, Uchiyama et al.12 incorporated a white top-hat transform method, which is relatively less sensitive to both image noise and inhomogeneous intensity condition. However, since the white top-hat transform is also based on a simple morphological filtering, it cannot fully utilize the nearby spatial and appearance information to enhance the PVSs.

To address these limitations, it is necessary to apply PVS enhancement and image denoising prior to PVS segmentation. Based on the observation that the PVSs are usually presented as thin tubular structures, we can effectively describe the PVSs with the line singularity representation17, 18. Traditional orthogonal wavelet transform19, 20, as widely used in image processing and singularity detection, only represents the point singularity. Accordingly, a series of redundant wavelets (including Ridgelet21, 22, Curvelet23, Contourlet17, NSCT18, etc.) have been developed to capture the line singularity. However, either the orthogonal wavelets or the redundant wavelets just directly convolute the local image neighborhood by using a certain convolution kernel. Thus, image artifacts are inevitably introduced during the local convolution-based transform. These artifacts may seriously decrease the accuracy of the PVS segmentation. For example, NSCT uses the long and narrow filter banks to achieve the linear singularity representation, thus the inverse transform usually introduces halo artifacts due to enhanced transform coefficients, as shown in Fig. 2. From Fig. 2(c), we can see that those near the enhanced coefficients are usually affected by the enhanced coefficients. Thus, they are likely to also be enhanced during the inverse transform, which introduces the halo artifacts. To this end, the local transform may not be optimal to enhance the PVSs for the subsequent PVS segmentation.

Figure 2
Figure 2

Illustration of halo artifacts by NSCT image enhancement. (a) Original image and NSCT forward transform; (b) enhanced transform coefficients; (c) NSCT inverse transform; and (d) enhanced image with halo artifacts.

To suppress the artifacts for better image enhancement and denoising, the nonlocal strategy, such as nonlocal-means (NL_means)24 and block-matching & 3D filtering (BM3D)25, 26, have been developed. These nonlocal methods try to find a number of nonlocal image blocks, which are (1) visually similar to the reference block and (2) located within the nonlocal search range of the reference block. The weighted mean of the nonlocal image blocks is used to remove the noise24. In BM3D method, a separable 3D transform on the stacked 3D array of the matched similar blocks (by implementing 2D transform on each 2D image block and then performing 1D transform on the third dimension) is used to obtain the enhanced sparse representation of the image, and finally, by hard-thresholding the transformed coefficients, the reasonable image denoising result can then be achieved25, 26. On the other hand, the nonlocal filtering methods have been further extended to address the higher-dimensional problems (BM4D)27, 28 in video and 3D medical image data. Although both types of these nonlocal methods and BM3D methods are effective on image denoising, they fail to enhance the PVSs in 7 T MR images since PVSs are tiny structure in the 7 T MR images and thus the matched nonlocal image blocks are highly similar to each other. In this way, PVSs cannot be separated from the background.

In this paper, inspired by the nonlocal image denoising method, we propose a novel PVS enhancement method using the nonlocal line singularity representation. We first extract K neighboring cubes from a small neighborhood of the reference cube, according to their top-left corner coordinates. All K extracted cubes are formed into a group, upon which we perform 1-D Haar transform. Then, we process the Haar transform coefficients by using a nonlinear mapping function, in order to amplify image details and suppress noise simultaneously. Next, the enhanced K neighboring cubes can be reconstructed from the processed Haar transform coefficients. By repeating these steps for reference cubes in the whole image, we can obtain an enhanced 7 T MR image (see Fig. 1(c)). Finally, we apply the existing BM4D method to further denoise the enhanced image, which is referred to as the enhanced and denoised image (see Fig. 1(d)) in this paper. To verify the effectiveness of our proposed image enhancement and denoising method, we use two existing PVS segmentation methods, i.e., (1) vesselness thresholding and (2) random forest learning16, to segment PVSs from the original 7 T MR images and also our enhanced and denoised 7 T MR images. The experimental results show that our proposed enhancement and denoising method can significantly improve image quality and potentially help with PVS segmentation.

Method

Our proposed framework for PVS enhancement and segmentation is summarized in Fig. 3. Our PVS enhancement method is mainly based on the Haar transform of nonlocal cubes. In particular, first, we extract K neighboring cubes within a small neighborhood (i.e., 3 × 3 × 3) of the center of each reference cube in the 7 T MR image. (Note that, if a cube’s top-left corner location is included in the small neighborhood (i.e., 3 × 3 × 3) of the reference cube, it can be extracted as a neighboring cube.) Then, we perform the Haar transform on the group of all the extracted nonlocal cubes. Note that, after Haar transform, the PVSs can be effectively represented in the transformed subbands. Next, to enhance PVSs, according to the characteristics of PVSs, we further process the transformed coefficients by nonlinear mapping to simultaneously amplify the signals (relevant to PVSs) and also suppress the noise. In this way, the enhanced PVSs can be effectively reconstructed by the inverse Haar transform, given the processed (transformed) coefficients. By reconstructing all reference cubes from the whole 7 T MR image, all reconstructed cubes can be aggregated into a single enhanced image. Finally, we apply the existing BM4D image-denoising method28 to further suppress the noise since the enhanced image may still contain noise that can affect PVS segmentation. Note that, we have released our image enhancement software to the public29. To prove that our proposed enhancement method is effective for the PVS segmentation, we compare the PVS segmentations by adopting two existing segmentation methods, i.e., vesselness thresholding based method and random forest based method, from the original 7 T MR images, the denoised images (by directly applying BM4D on the original 7 T MR images), and our enhanced and denoised images (by our complete method). The evaluation was performed on seventeen 7 T MR images16 from healthy volunteers ranging from ages 25 to 37. Informed consents were obtained for both study participation and publication of their images without identification information. The study was reviewed and approved by the institutional reviews board of University of North Carolina and the methods used in this study were carried out in accordance with the approved guidelines. Further details of the experimental setting are described in Section 3.

Figure 3
Figure 3

Framework of the proposed image enhancement (left panel) and PVS segmentation algorithm (right panel). After enhancing the image, two existing methods are used to segment the PVSs, i.e., (1) PVS segmentation based on hard-thresholding and (2) PVS segmentation based on random forest.

Nonlocal Haar Transform

The goal of this step is to enhance the 7 T MR image by better revealing the details of the PVSs. Again, our enhancement strategy is inspired by the nonlocal strategy in the literature, and is particularly implemented through the Haar transform. Specifically, first, we extract K (i.e., K = 8) neighboring image cubes (with the same size as the reference cube) from a small neighborhood (i.e., 3 × 3 × 3) around the center of the reference cube. Then, all the extracted cubes are formed into a group, denoted as C E =[ C 1 , C i ,, C K ] where C i is a 7 × 7 × 7 cube. Figure 4 shows an example of neighboring cube extraction scheme for the case of K = 8. We can observe that the patterns of the PVSs in the extracted cubes are slightly mis-aligned with each other and with the reference cube. Thus, the high-order linear operation (i.e., subtraction) between the two extracted cubes will amplify subtle differences between the included PVSs structures and contribute to capturing the PVSs in the extracted neighboring cubes. This motivates us to apply the Haar transform30 across all extracted nonlocal cubes, for enhancing the PVSs.

Figure 4
Figure 4

An illustration of the extraction of neighboring cubes (K = 8) from the small neighborhood (i.e., small blue block in the top of figure) around the center of the reference cube.

Forward Transform

We apply the forward transform to the extracted neighboring cubes, in order to allow them to represent the PVSs in the extracted cubes. Each neighboring cube is denoted as C i (i=1,,K). Taking K = 8 in our implementation, we can use the following formulation to perform the forward Haar transform:

C ˆ i = j = 1 K Ψ(i,j) C j ,(i=1,,K),
(1)

where Ψ(i, j) is the Haar measure at the crossing of the i-th row and the j-th column of the Haar transform matrix. Specifically, the 8 × 8 Haar transform matrix can be defined as below:

Ψ= ( 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 8 1 / 4 1 / 4 1 / 4 1 / 4 0 0 0 0 0 0 0 0 1 / 4 1 / 4 1 / 4 1 / 4 1 / 2 1 / 2 0 0 0 0 0 0 0 0 1 / 2 1 / 2 0 0 0 0 0 0 0 0 1 / 2 1 / 2 0 0 0 0 0 0 0 0 1 / 2 1 / 2 ) .
(2)

By applying the above Haar transform matrix, all neighboring cubes can be decomposed into 8 subbands. The first subband C ˆ 1 is the weighted average of all neighboring cubes, while the detailed differences among the cubes are preserved in the rest subbands.

Inverse Transform

The Haar transform matrix defined in Eq. (2) is invertible. Thus, we can reconstruct the enhanced K neighboring cubes from the computed coefficients of subbands by using the following inverse transform:

C ˆ ^ i = j = 1 K Ψ 1 (i,j) C ˆ j ,(i=1,,K),
(3)

where Ψ−1 denotes the inverse matrix of Ψ.

After reconstructing all enhanced K neighboring cubes in the whole 7 T MR image, we can put them back into their original locations to build a final reconstructed 7 T MR image. Note that the extracted K neighboring cubes are overlapping with each other, which indicates that their corresponding reconstructed cubes are also overlapping. In this way, we average the signals of the same location from the neighboring reconstructed cubes to obtain the final reconstructed 7 T MR image.

Image Enhancement

Image enhancement can be achieved by amplifying the transformed coefficients related to the signals (i.e. PVSs structures). Thus, the appropriate representation of signals is critical to enhance the 7 T MR image. Note that, if the signals are not well-represented, signals and background noise can both be amplified, making it difficult to separate them. Our proposed nonlocal Haar transform can effectively address this problem by simultaneously amplifying the signals (related to PVSs) and suppressing the background noise. Moreover, the artifacts can be alleviated in our method by avoiding the use of large-size filter banks, which were often used in some local transforms (i.e., Contourlet17, NSCT18, etc.) to represent line singularity.

The proposed image enhancement algorithm can be summarized below:

  1. 1)

    Extract K neighboring cubes for each reference cube.

  2. 2)

    Implement the forward Haar transform on the stack of all the extracted K neighboring cubes according to Eq. (1).

  3. 3)

    Amplify the details and suppress the background noise by mapping the transform coefficients according to Eq. (4) below:

    c En = { c H , i f | c H | > T 1 γ 1 c H , i f T 2 | c H | T 1 γ 2 c H i f T 3 < | c H | < T 2 0 i f | c H | < T 3 ,
    (4)

    where cH is the computed Haar transform coefficient, cEn is the enhanced Haar transform coefficient, and T1, T2 and T3 are the three thresholds used to determine whether the particular Haar transform coefficient corresponds to the common image edge, weak image edge, or noise. Here, we just use the gain factors γ1 and γ2 to amplify the weak edges. All these five parameters (T1, T2, T3, γ1, γ2) are the empirical values. Specifically, we set T1 and T3 so that the Haar transform coefficients of PVSs are included between T1 and T3 and then investigate T2 within T1 and T3 to distinguish the common image edges and the ambiguous structures. Since the most common edges with the Haar transform coefficients within T1 and T2 are PVSs, the gain factors γ1 is determined by a relatively large value so that the enhanced edges are clearly shown. On the other hand, since it is difficult to tell the difference between coefficients of very weak edges and noise, we use a relatively small gain factor γ2 (i.e., half of γ1) to amplify those small magnitude coefficients and potentially also enhance the very weak edges.

  4. 4)

    Implement the inverse Haar transform to enhance those K extracted neighboring cubes according to Eq. (3).

  5. 5)

    Aggregate all the enhanced K neighboring cubes of each reference cube according to the way described in the end of Section 2.1, and finally obtain an entire enhanced 7 T MR image.

Further Image Denoising

By using the above-proposed processing steps, we can obtain the enhanced 7 T MR image. However, because we use a small threshold γ2 to distinguish between the very weak edges and the noise, the enhanced image may still contain some noise. To address this issue, we further use the BM4D method28 to denoise the enhanced image. As briefly mentioned above, BM4D is a state-of-the-art 3D image denoising method. It distinguishes the cubes that are the most similar by cube-matching within a noisy 3D image, and then it stacks similar cubes into a 4D array, and finally performs the separable 4D transform on the stacked 4D array (i.e., applying the separable 3D transform to each cube, and then 1D transform on the fourth dimension). Using this method, noise coefficients can be effectively separated from the signal coefficients in order to extract a reasonably denoised image. As shown in Fig. 1, if denoising is directly performed on the original 7 T MR image, many PVSs are smoothed, but noise can still be effectively suppressed. On the other hand, the PVSs are effectively enhanced while the noise is removed when applying our enhancement and denoising method. Figure 5 further zooms in the original 7 T MR image, the denoised image from the original 7 T MR image, and the denoised image from our enhanced image (with our complete method).

Figure 5
Figure 5

Zoom-in views of the original image (left), the denoised image obtained directly from the original image (middle), and the denoised image from our enhanced image by our complete method (right). Please compare PVSs within contours or pointed by arrows.

PVS Segmentation on enhanced images

After enhancing and denoising the 7 T MR images by our above-proposed method, we turn to segment the PVSs from our enhanced and denoised images. In particular, we adopt two existing PVS segmentation methods, as briefly introduced below.

Vesselness Based Segmentation

Vesselness thresholding method is one of typical ways to extract thin structures31,32,33. We use this as the first method to extract the PVSs from the original 7 T MR images or our enhanced and denoised images. In this method, 7 T MR image is first smoothed through a Gaussian kernel with a scale s, and then the eigenvalues of the second derivative matrix (Hessian matrix) of the smoothed image is computed. Note that the three eigenvalues (λ1, λ2, λ3) represent the magnitudes of derivatives for their associated eigenvectors V1, V2, V3, which represent three principal directions. If the smallest eigenvalue λ1 is small while the other two eigenvalues λ2 and λ3 are relatively large, a high vesselness can be obtained, and thus the respective voxel can be likely belonging to the tubular structure. More details on this vesselness measurement method can be found in Frangi et al.’s work31. In our application, we compute the vesselness values with two small scales (s = 0.5 and s = 1) for extracting very thin or relatively thick PVSs structures, and then use their maximum vesselness as the final vesselness for the respective voxel. Finally, we extract the voxels with higher vesselness than a certain threshold as PVSs. An example of vesselness map is shown in Fig. 6. We can identify a better vesselness map achieved using our enhanced and denoised image, as compared to the original 7 T MR image, especially for the zoom-up regions in the bottom-left corner of each subfigure in Fig. 6.

Figure 6
Figure 6

Vesselness estimation on the original image and our enhanced and denoised image. (a) Original image, (b) estimated vesselness map from (a,c) our enhanced and denoised image, and (d) estimated vesselness map from (c).

Random Forest Based Segmentation

We also apply a learning-based PVS segmentation method16 to our enhanced and denoised image. In the training stage, PVSs and non-PVSs voxels are sampled from a region of interest (ROI) determined by a certain vesselness threshold. Then, 3D randomized Haar features are extracted from those samples, and the sequential classifiers are learned using the random forest model34,35,36. To capture the consistent patterns of PVSs with different orientations and intensity inhomogeneity, both the principal directions and the intensity distribution of local region are normalized during the extraction procedure of the 3D randomized Haar features. In the testing stage, the Haar features are extracted at each target voxel in the ROI and are then passed through trained decision trees from the root node to a leaf node, with respect to the established decision functions of all selected nodes. All the class labels in the selected leaf nodes of all decision trees are averaged. The class label of the underlying voxel is finally determined by the maximal prediction.

Experimental results

For quantitative comparison, we have created ground-truth labels for the 6 images and also for the right hemispheres of the remaining 11 images. We then verify the effectiveness of our enhancement method by comparing the PVS segmentation results achieved from (1) the original 7 T MR images, (2) the denoised images (by directly applying BM4D on the original 7 T MR images), and (3) our enhanced and denoised images (by our complete method). The segmentation performance is evaluated by the Dice similarity coefficient (DSC), sensitivity (SN), and positive prediction value (PPV) as defined below:

DSC= 2 T P 2 T P + F P + F N ,SN= T P T P + F N ,PPV= T P T P + F P
(5)

where TP, FP, and FN denote the true positive, false positive, and false negative, respectively.

Ground Truth

The ground-truth labels were created by intensity-thresholding on a region of interest (ROI) defined manually. Note that it is very challenging to generate accurate PVSs ground-truth labels from the original images even though the ROI is carefully defined, since ambiguous noise is so close to the weak PVSs. On the other hand, in the enhanced images, there is less noise and the intensities of PVSs voxels are more distinguishable from the background voxels. Thus, we selected to generate the ground-truth labels from the enhanced images. Specifically, we manually marked the ROI that contains PVSs, and then extracted the voxels with higher intensities than a certain threshold as PVSs. Figure 7(a,b) show the manually drawn ROI (i.e., red voxels in Fig. 7(a)) and the ground-truth labels generated by the intensity-thresholding, followed by manual editing via two experts.

Figure 7
Figure 7

(a) The manually drawn ROIs of PVSs (red regions), (b) the ground-truth labels (red thin line structures) generated by intensity hard-thresholding upon the ROIs, followed by manual editing via two experts, (c) the same image with our PVS segmentation results (red thin line structures), (d) the 3D rendering of ground-truth labels, and (e) the 3D rendering of our PVS segmentation results.

Parameters

For the forward Haar transform of the nonlocal cubes, we extracted eight 7 × 7 × 7 cubes on a small neighborhood (3 × 3 × 3) of the reference cube, which was selected by the sliding window with the step of 7. As described in Section 2.2, we chose T1 and T3 as 50 and 150 so that the minimum and maximum Haar transform coefficients of PVSs could be included in this range, and then we chose T2 as 110 since most coefficients of PVS were included in the range between 110 and 150. The two parameters γ1 and γ2 for coefficient enhancement were heuristically set as 24 and 12, respectively, since coefficients of PVSs were sometimes around 50, while those of bright parts near the skull were around 1200. The final image denoising was performed by the BM4D method using the available code (http://www.cs.tut.fi/~foi/GCF-BM3D/) and their recommended settings28. The vesselness map was generated first by segmenting the white matter region with FAST37 and then by computing the eigenvalues of the Hessian matrix31 at each voxel within the white matter region. For the random forest based segmentation method, we conducted a leave-one-out cross-validation, i.e., testing an image by using the classifiers trained from all other 16 remaining images.

PVS Segmentation by Vesselness Hard-thresholding

In the vesselness thresholding method16, 38, the optimal vesselness threshold depends on the images under study. Thus, we tested several vesselness thresholds and then used the best one for each image. Table 1 show the segmentation accuracies for the three types of images. Figure 8 shows the box plots of the segmentation accuracies for the three types of images, i.e., (1) the original 7 T MR images, (2) the denoised images (by directly applying BM4D on the original 7 T MR images), and (3) our enhanced and denoised images (by our complete method). The average DSC score on the original 7 T MR images was only about 58%, because of too much ambiguous noise in the original images. When the original images were denoised by BM4D, the DSC score improved about 4%, and the highest DSC on our enhanced and denoised images could be achieved up to 67%. The PPV and SN scores present a similar situation. In summary, the best segmentation accuracies were achieved for the most cases when our enhanced and denoised images were used.

Table 1: Segmentation accuracy of the vesselness hard-thresholding method on the original 7 T MR images, the denoised images, and our enhanced and denoised images.
Figure 8
Figure 8

Segmentation accuracies of the vesselness hard-thresholding method for the original 7 T MR images (left), denoised images (middle), and our enhanced and denoised images (right).

PVS Segmentation by Random Forest

When we use the random forest for PVS segmentation, we need to design sequential classifiers for iteratively performing the segmentation. That is, the outputs of the first classifier are used as features to train the second classifier, along with the features already computed from the input image. This step continues until no performance can be improved with more classifiers. For the original 7 T MR image, we learned three sequential classifiers to achieve the best segmentation results. On the other hand, we found that, most noise was already removed in the denoised images (by directly applying BM4D on the original 7 T MR images) and our enhanced and denoised images, and thus the best results were achieved by using the first classifier (and using more classifiers did not improve the results) in these cases. Therefore, for each type of images, we used the best number of classifiers for segmentation.

Experimental results are shown in Table 2. Figure 9 also shows the box plots of the segmentation accuracies for the three types of images. By comparing Table 2 with Table 1 and by comparing Fig. 9 with Fig. 8, we can see that, in most cases, the random forest based method can achieve much higher PVS segmentation accuracy than the hard-thresholding based method on every type of image. The highest segmentation accuracies were achieved when both our enhancement and denoising method (our complete method) and the random forest based method were used to segment PVSs. Figure 7(c,e) further shows an example of the predicted result by our complete method. As can be seen, our predicted PVSs are very similar to the ground-truth labels with the best PVS segmentation results, indicating that our proposed image enhancement and denoising method is helpful for PVS segmentation.

Table 2: Segmentation accuracy of the random forest method on the original 7 T MR images, the denoised images, and our enhanced and denoised images.
Figure 9
Figure 9

Segmentation accuracies of the random forest learning based method for the original 7 T MR images (left), denoised images (middle), and our enhanced and denoised images (right).

Conclusion and Discussion

In this paper, we have presented a novel method for enhancing PVSs in the 7 T MR images. To demonstrate the effectiveness of our PVSs enhancement method, we have applied two existing PVS segmentation methods, i.e., (1) vesselness thresholding and (2) random forest learning, to the original 7 T MR images, denoised images, and our enhanced and denoised images, respectively. When the thresholding based method or the random forest method was directly applied to the original 7 T MR images without preprocessing, performance was limited due to very similar appearances of thin PVS and noise in the given images. On the other hand, the proposed preprocessing step effectively enlarged the gap between the appearances of thin PVS and noise, and thus the PVS segmentations from our enhanced and denoised images were much more accurate than those from the original and denoised images for both segmentation methods. The performance gains of the thresholding based method and the random forest method were 9% and 15%, respectively, while the random forest based method could achieve 8% higher average DSC accuracy than the thresholding based method. Specially, when using the vesselness thresholding based method, many errors occurred at the edges between white matter and gray matter, as the thin edges were very similar to the structures of PVSs. This issue can be effectively addressed using the random forest based method by utilizing contextual patterns of local neighbors. A potential limitation of our method is the heuristic setting of various thresholds and amplification factors. The intensity ranges of MR images depend on MR acquisition parameters, while the number of available images we have is only 17. Thus, it is hard to find the optimal parameters, since, even if the parameters are optimized, this can lead to overfitting. Accordingly, we left those parameters without parameter optimization. On the other hand, we believe that our released software package29 with multiple parameters would be able to provide more flexibility, when applied to other MR images acquired by different parameter settings. Regarding the PVS segmentation accuracy, some short and small structures were still observed occasionally in the segmented results, but most of the noise was removed in our enhanced and denoised image. In the future, we will further develop our method to remove those small structures (i.e., by using morphological filtering method16, 38) and also improve the overall performance of our complete method (i.e., by better coordinating the two enhancement and denoising steps in our method).

Additional Information

Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. 1.

    Iliff, J. J. et al. Cerebral arterial pulsation drives paravascular CSF–interstitial fluid exchange in the murine brain. J. Neurosci. 33(46), 18190–18199 (2013).

  2. 2.

    Thrane, V. R. et al. Paravascular microcirculation facilitates rapid lipid transport and astrocyte signaling in the brain. Sci. Rep. 3, 2582 (2013).

  3. 3.

    Yang, L. et al. Evaluating glymphatic pathway function utilizing clinically relevant intrathecal infusion of CSF tracer. J. Transl. Med. 11, 107 (2013).

  4. 4.

    Heier, L. A. et al. Large Virchow–Robin spaces: MR-clinical correlation. Am. J. Neuroradiol. 10(5), 929–936 (1989).

  5. 5.

    Zhu, Y. C. et al. Severity of dilated Virchow–Robin spaces is associated with age, blood pressure, and MRI markers of small vessel disease: a population-based study. Stroke 41(11), 2483–2490 (2010).

  6. 6.

    Chen, W., Song, X. & Zhang, Y. & ADNI. Assessment of the Virchow–Robin spaces in Alzheimer disease, mild cognitive impairment, and normal aging, using high-field MR imaging. Am. J. Neuroradiol. 32(8), 1490–1495 (2011).

  7. 7.

    Maclullich, A. M. et al. Enlarged perivascular spaces are associated with cognitive function in healthy elderly men. J. Neurol. Neurosurg. Psychiatry 75(11), 1519–1523 (2004).

  8. 8.

    Rouhl, R. P., van Oostenbrugge, R. J., Knottnerus, I. L. H., Staals, J. E. A. & Lodder, J. Virchow–Robin spaces relate to cerebral small vessel disease severity. J. Neurol. 255(5), 692–696 (2008).

  9. 9.

    Doubal, F. N., MacLullich, A. M. J., Ferguson, K. J., Dennis, M. S. & Wardlaw, J. M. Enlarged perivascular spaces on MRI are a feature of cerebral small vessel disease. Stroke 41(3), 450–454 (2010).

  10. 10.

    Bouvy, W. H. et al. Visualization of perivascular spaces and perforating arteries with 7T magnetic resonance imaging. Investig. Radiol. 49(5), 307–313 (2014).

  11. 11.

    Wuerfel, J. et al. Perivascular spaces — MRI marker of inflammatory activity in the brain? Brain 131, 2332–2340 (2008).

  12. 12.

    Uchiyama, Y. et al. Computer-aided diagnosis scheme for classification of lacunar infarcts and enlarged Virchow–Robin spaces in brain MR images. Conf. Proc. IEEE Eng. Med. Biol. Soc, 3908–3911 (2008).

  13. 13.

    Ramirez, J. et al. Lesion Explorer: a comprehensive segmentation and parcellation package to obtain regional volumetrics for subcortical hyperintensities and intracranial tissue. NeuroImage 54(2), 963–973 (2011).

  14. 14.

    Ramirez, J. et al. Visible Virchow–Robin spaces on magnetic resonance imaging of Alzheimer’s disease patients and normal elderly from the Sunnybrook Dementia study. J. Alzheimers Dis. 43(2), 415–424 (2015).

  15. 15.

    Wang, X. et al. Development and initial evaluation of a semi-automatic approach to assess perivascular spaces on conventional magnetic resonance images. J. Neurosci. Methods 257, 34–44 (2016).

  16. 16.

    Park, S. H., Zong, X., Gao, Y., Lin, W. & Shen, D. Segmentation of perivascular spaces in 7T MR image using auto-context model with orientation-normalized features. NeuroImage 134(7), 223–235 (2016).

  17. 17.

    Do, M. N. & Vetterli, M. The contourlet transform: an efficient directional multi resolution image representation. IEEE Trans. Image Process. 14(12), 2091–2106 (2005).

  18. 18.

    Cunha, A. L., Zhou, J. & Do, M. N. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 15(10), 3089–3101 (2006).

  19. 19.

    Mallat, S. A Wavelet Tour of Signal Processing. 3rd ed. Ch.4, 89–154 Academic Press, 1999).

  20. 20.

    Mallat, S. & Hwang, W. L. Singularity detection and processing with wavelets. IEEE Trans. Information Theory. 38(2), 617–643 (2002).

  21. 21.

    Candès, E. J., Ridgelets: Theory and Applications. PhD thesis, Stanford University (1998).

  22. 22.

    Do, M. N. & Vetterli, M. The finite ridgelet transform for image representation. IEEE Trans. Image Process. 12(1), 16–28 (2003).

  23. 23.

    Candès, E. J. & Donoho, D. L. Curvelets-a surprisingly effective non adaptive representation for objects with edges. Curve and Surface Fitting, TN, Nashville:Vanderbilt University Press (1999).

  24. 24.

    Buades, A., Coll, B. & Morel, J. M. A review of image denoising algorithms, with a new one. Multiscale Modeling & Simulation. 4(2), 490–530 (2005).

  25. 25.

    Dabov, K., Foi, A., Katkovnik, V. & Egiazarian, K. Image denoising by sparse 3D transform-domain collaborative filtering. IEEE Trans. Image Process. 16(8), 2080–2095 (2007).

  26. 26.

    Hou, Y., Zhao, C., Yang, D. & Cheng, Y. Comments on ‘Image denoising by sparse 3D transform-domain collaborative filtering’. IEEE Trans. Image Process. 20(1), 268–270 (2011).

  27. 27.

    Maggioni, M., Boracchi, G., Foi, A. & Egiazarian, K. Video denoising, deblocking, and enhancement through separable 4-D nonlocal spatiotemporal transforms. IEEE Transactions on Image Processing. 21(9), 3952–3966 (2012).

  28. 28.

    Maggioni, M., Katkovnik, V. & Egiazarian, K. Nonlocal transform-domain filter for volumetric data denoising and reconstruction. IEEE Transactions on Image Processing. 22(1), 119–133 (2013).

  29. 29.

    Park, S.H., Hou, Y. & Shen, D. Neuroimaging informatics tools and resources clearinghouse (NITRC) https://www.nitrc.org/projects/pvs_enhance/.

  30. 30.

    Lee, B. Y. & Tarng, Y. S. Application of the discrete wavelet transform to the monitoring of tool failure in end milling using the spindle motor current. The International Journal of Advanced Manufacturing Technology. 15(4), 238–243 (1999).

  31. 31.

    Frangi, A. F., Niessen, W. J., Vincken, K. L. & Viergever, M. A. Multiscale vessel enhancement filtering. Med. Image Comput. Comput. Assist. Interv. 1496, 130–137 (1998).

  32. 32.

    Shi, F. et al. CENTS: cortical enhanced neonatal tissue segmentation. Hum. Brain Mapp. 32(3), 382–396 (2011).

  33. 33.

    Cheng, J. Z., Chen, C. M., Cole, E. B., Pisano, E. D. & Shen, D. Automated delineation of calcified vessels in mammography by tracking with uncertainty and graphical linking techniques. IEEE Trans. Med. Imaging 31(11), 2143–2155 (2012).

  34. 34.

    Criminisi, A., Shotton, J. & Konukoglu E. Decision Forests for Classification, Regression, Density Estimation, Manifold Learning and Semi-supervised Learning, MSR-TR-2011-114, (2011).

  35. 35.

    Criminisi, A. & Shotton, J. Decision Forests for Computer Vision and Medical Image Analysis Ch. classification forest, 25–45 (Springer, 2013).

  36. 36.

    Criminisi, A. et al. Regression forests for efficient anatomy detection and localization in computed tomography scans. Med. Image Anal. 17(8), 1293–1303 (2013).

  37. 37.

    Smith, S. M. Fast robust automated brain extraction. Hum. Brain Mapp. 17(3), 143–155 (2002).

  38. 38.

    Zong, X., Park, S. H., Shen, D. & Lin, W. Visualization of perivascular spaces in the human brain at 7T: sequence optimization and morphology characterization. NeuroImage 125, 895–902 (2016).

Download references

Acknowledgements

This work was supported by the National Science Foundation of China under Grant 61379015, the Natural Science Foundation of Shandong Province under Grant ZR2011FM004. This work was also supported in part by NIH grants (EB006733, EB008374, MH100217, MH108914, AG041721, AG049371, AG042599, AG053867, MH110274).

Author information

Author notes

  1. Yingkun Hou and Sang Hyun Park contributed equally to this work.

Affiliations

  1. School of Information Science and Technology, Taishan University, Taian, 271000, China

    • Yingkun Hou
  2. Department of Robotics Engineering, Daegu Gyeongbuk Institute of Science and Technology, Daegu, 42988, South Korea

    • Sang Hyun Park
  3. Med-X Research Institute, School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200030, China

    • Qian Wang
  4. Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA

    • Yingkun Hou
    • , Jun Zhang
    • , Xiaopeng Zong
    • , Weili Lin
    •  & Dinggang Shen
  5. Department of Brain and Cognitive Engineering, Korea University, Seoul, 02841, South Korea

    • Dinggang Shen

Authors

  1. Search for Yingkun Hou in:

  2. Search for Sang Hyun Park in:

  3. Search for Qian Wang in:

  4. Search for Jun Zhang in:

  5. Search for Xiaopeng Zong in:

  6. Search for Weili Lin in:

  7. Search for Dinggang Shen in:

Contributions

Y.H., S.P., and Q.W. wrote the main manuscript, conducted the experiments, and prepared the figures. J.Z. helped to revise the manuscript. X.Z. and W.L. collected and processed the 7 T MR data. D.S. managed this project and mainly revised the manuscript. All authors reviewed the manuscript.

Competing Interests

The authors declare that they have no competing interests.

Corresponding author

Correspondence to Dinggang Shen.

Comments

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

Creative Commons BY

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.