Dilated Dense U-Net for Infant Hippocampus Subfield Segmentation

Published on April 24, 2019

Dilated Dense U-Net for Infant Hippocampus Subfield Segmentation

Hancan Zhu1,2, Feng Shi3, Li Wang1, Sheng-Che Hung1, Meng-Hsiang Chen4, Shuai Wang1, Weili Lin1 and Dinggang Shen1,5*
1Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC, United States
2School of Mathematics Physics and Information, Shaoxing University, Shaoxing, China
3Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
4Department of Diagnostic Radiology, Kaohsiung Chang Gung Memorial Hospital, Chang Gung University College of Medicine, Kaohsiung, Taiwan
5Department of Brain and Cognitive Engineering, Korea University, Seoul, South Korea
Accurate and automatic segmentation of infant hippocampal subfields from magnetic resonance (MR) images is an important step for studying memory related infant neurological diseases. However, existing hippocampal subfield segmentation methods were generally designed based on adult subjects, and would compromise performance when applied to infant subjects due to insufficient tissue contrast and fast changing structural patterns of early hippocampal development. In this paper, we propose a new fully convolutional network (FCN) for infant hippocampal subfield segmentation by embedding the dilated dense network in the U-net, namely DUnet. The embedded dilated dense network can generate multi-scale features while keeping high spatial resolution, which is useful in fusing the low-level features in the contracting path with the high-level features in the expanding path. To further improve the performance, we group every pair of convolutional layers with one residual connection in the DUnet, and obtain the Residual DUnet (ResDUnet). Experimental results show that our proposed DUnet and ResDUnet improve the average Dice coefficient by 2.1 and 2.5% for infant hippocampal subfield segmentation, respectively, when compared with the classic 3D U-net. The results also demonstrate that our methods outperform other state-of-the-art methods.

Introduction
Hippocampus plays important roles in memory and spatial navigation, and is closely related to neurological diseases, such as autism, attention deficit hyperactivity disorder, and Alzheimer’s Disease (Shi et al., 2009; Bartsch, 2012; Li et al., 2013). Hippocampus consists of several histologically and functionally specialized subfields (Dalton et al., 2017). It has been shown that different pathological conditions affect subfields differently, suggesting that subfields may provide more precise information for earlier disease diagnosis than simply using the whole hippocampus (Small, 2014).
Accurate segmentation of hippocampal subfields from magnetic resonance (MR) brain images is a critical step for studying memory-related neurological diseases. However, it is a challenging task especially in infant subjects, because of the small size of each hippocampal subfield, the blurred boundaries between subfields, and the large inter-subject variations. Manual segmentation is widely adopted, but it suffers high intra- and inter-operator variability, and is also excruciatingly time-consuming. Therefore, automatic hippocampal subfield segmentation methods are desirable. The existing automatic hippocampal subfield segmentation methods can be mainly categorized into three different types: (1) generative model based method (Van Leemput et al., 2009), (2) multi-atlas based method (Wang et al., 2013; Pipitone et al., 2014; Caldairou et al., 2016; Romero et al., 2017), and (3) multi-modality learning based method (Wu et al., 2018).
In the first category (Van Leemput et al., 2009), a generative model of image around the hippocampal area was produced by using a mesh-based probabilistic atlas learned from a set of ultra-high-resolution training images. The model was used to obtain automated hippocampal subfield segmentations on 10 adult subjects with the age range of 22–89 years.
In the past years, the second category of methods, namely multi-atlas based image segmentation (MAIS) methods, have been widely used in the field of medical image segmentation, including hippocampal subfield segmentation on adult subjects (Wang et al., 2013; Pipitone et al., 2014; Caldairou et al., 2016; Romero et al., 2017). In the MAIS methods, all selected atlas images are first registered to the target image, and the corresponding atlas labels are then warped to the target image space. Afterwards, these warped atlas labels are combined to obtain the final segmentation by label fusion. Note, in the MAIS methods, label fusion plays an important role. For example, a weighed voting label fusion was proposed (called joint label fusion) in a previous work (Wang et al., 2013), in which weights were obtained by minimizing the total expected error between the consensus segmentation and the ground-truth segmentation. This method was later combined with a learning-based error correction method for hippocampal subfield segmentation (Yushkevich et al., 2015). In another work (Romero et al., 2017), a new non-local patch based label fusion method was proposed based on a multi-contrast patch matching process. To further improve the segmentation, authors exploited a neural network-based error correction step for minimizing systematic segmentation errors. MAGeT-Brain (Multiple Automatically Generated Templates) was also proposed for automatic segmentation of the hippocampus and subfields, aiming to minimize the number of atlases needed whilst still achieving similar agreement to the multi-atlas approaches (Pipitone et al., 2014),. Besides, a surface patch-based segmentation method (Caldairou et al., 2016) was proposed for hippocampal subfield segmentation by combining surface-based processing with a patch-based template library and feature matching.
Besides the above two categories of methods, learning-based methods in the third category were also proposed for adult hippocampal subfield segmentation using 3T multi-modality MR images, including structural MRI (T1w, T2w) and resting-state fMRI (rs-fMRI) (Wu et al., 2018). In that paper (Wu et al., 2018), authors extracted both appearance features and relationship features to capture the appearance patterns in structural MR images and the connectivity patterns in rs-fMRI, respectively. These extracted features were then fed into a random forest classifier for voxel-wise classification.
Although several automatic methods have also been proposed for hippocampal subfield segmentation, most of them were evaluated only on the adult subjects, and thus cannot be directly applied to infant subjects due to insufficient tissue contrast and fast changing structural patterns of early hippocampal development.
In the recent years, deep convolutional neural networks (CNN) have been widely applied in the medical image segmentation (de Brébisson and Montana, 2015; Zhang et al., 2015; Moeskops et al., 2016). In CNN based segmentation methods, a patch centered at the target voxel (or pixel for 2D images) is taken as input for networks, and the tissue class of the center voxel is produced as the output of the networks. By learning sets of convolutional kernels, CNNs can capture highly non-linear mappings between inputs and outputs. Compared with MAIS methods and the traditional learning-based methods, CNN based segmentation methods are free of image registration and manual feature extraction.
A drawback of the CNN based segmentation approaches is that the input patches from neighboring voxels have huge overlap and the same convolutions are computed for many times. To address this limitation, fully convolutional networks (FCN) were proposed for voxel-wise dense prediction, by reformatting the fully connected layers as convolutional layers (Long et al., 2015). So far, a number of FCNs have been proposed and successfully used in medical image segmentation, including hippocampal segmentation (Ronneberger et al., 2015; Milletari et al., 2016; Chen Y. et al., 2017; Yu et al., 2017; Cao et al., 2018). For example, in the paper (Ronneberger et al., 2015), a U-net architecture was proposed by comprising a contracting (down-sampling) path, followed by an expanding (up-sampling) path. The features in the contracting path are concatenated to the corresponding features in the expanding path to recover the detailed image information that is lost during the down-sampling process. In the work (Milletari et al., 2016), authors extended U-net to a V-net structure by incorporating residual blocks (He et al., 2016a). In the paper (Yu et al., 2017), authors proposed a new volumetric convolutional neural network with mixed residual connections, where both the short connections between successive layers and the long connections between contracting path and expanding path are implemented with residual connections. In the work (Cao et al., 2018), authors proposed a multi-task CNN for joint hippocampal segmentation and clinical score regression with U-net as a subnet for hippocampal segmentation. In the paper (Chen Y. et al., 2017), authors proposed a multi-view ensemble approach to combine multiple decision maps obtained from several deep neural networks for hippocampal segmentation. Besides these contracting-expanding structures, dilated FCNs were also proposed for image segmentation, which can enlarge the receptive field exponentially without reducing any spatial resolution (Liang-Chieh et al., 2015; Yu and Koltun, 2015;Li et al., 2017; McKinley et al., 2017).
The U-net like structures are particularly successful in the field of medical image segmentation. One of the most important factors in the U-net is the long-skip connections which can concatenate the features in the contracting path to the corresponding features in the expanding path to recover the lost image information. However, the levels of features in the contracting path are much lower than those in the expanding path. Thus, it may not obtain optimal results when directly concatenating these features.
In this paper, we develop an automatic method to address the challenging infant hippocampal subfield segmentation problem with state-of-the-art deep learning techniques (LeCun et al., 2015; Litjens et al., 2017; Shen et al., 2017). To overcome the limitation of U-net structure, we propose a novel network by embedding a dilated dense network in the U-net, namely DUnet. The embedded dilated dense network can generate multi-scale features while keeping high spatial resolution, which is useful in fusing the low-level features in the contracting path with the high-level features in the expanding path. To further improve the performance, we use residual connections to group every pair of convolutional layers in DUnet, and obtain the Residual DUnet (ResDUnet).
The proposed method was applied for segmenting infant hippocampal subfields based on the Baby Connectome Project (BCP) dataset, containing 10 infant subjects. To the best of our knowledge, this is the first work to propose an automatic method for infant hippocampal subfield segmentation. To further illustrate the effectiveness of our proposed method, we also validated our proposed method for segmenting adult hippocampal subfields on a publicly available dataset. Experimental results show that our proposed DUnet and ResDUnet, respectively, improve the average Dice coefficient by 2.1 and 2.5% for infant hippocampal subfield segmentation, and 0.5 and 0.6% for adult hippocampal subfield segmentation, compared to the classic 3D U-net (Çiçek et al., 2016). Our proposed ResDUnet also outperforms both the state-of-the-art ConvNet (Yu et al., 2017) and hippocampal subfield segmentation method (HIPS) (Romero et al., 2017).
Materials
Two image datasets were used for validating our method. The first dataset is from BCP, which was funded by the National Institutes of Health (NIH) as a component of the Lifespan Human Connectome Project. The BCP aims to provide scientists with unprecedented information about how the human brain develops from birth through early childhood and will uncover factors contributing to healthy brain development. For this project, researchers are acquiring MRI scans (including T1- and T2-weighted structural MRI, DTI, and rs-fMRI) of 500 typically developing children, ages 0–5 years, over the course of 4 years. In our experiment, 10 infant subjects (6 females/4 males) were randomly selected, each with T1w and T2w images acquired at 12 months old with 3T Siemens Prisma MRI scanners at the Biomedical Research Imaging Center (BRIC) at the University of North Carolina at Chapel Hill. Table 1 lists the imaging protocol for acquiring the T1w and T2w MR images. Five hippocampal subfields were manually labeled for each subject by the consensus of two neuroradiologists, including cornu ammonis sectors 1 (CA1), CA2/3, subiculum (SUB), CA4/dentate gyrus (DG), and Uncus. All T1w and T2w images underwent intensity inhomogeneity correction using the N3 bias field correction, and T2w images were rigidly aligned with corresponding T1w images. All images were aligned to a selected subject with affine registration.

TABLE 1

Table 1. Imaging protocol for acquiring infant T1w and T2w MR images.

The second dataset is a publicly available dataset (https://www.nitrc.org/projects/mni-hisub25), which contains 25 adult subjects (31 ± 7 years, 12 males). Each subject consists of an isotropic 3D-MPRAGE T1-weighted image (TR = 3,000 ms; TE = 4.32 ms; TI = 1,500 ms; flip angle = 7°; matrix size = 336 × 384; FOV = 201 × 229 mm2; 240 axial slices with 0.6 mm slice thickness resulting in 0.6 × 0.6 × 0.6 mm3 voxels; acquisition time = 16.48 min), an anisotropic 2D T2-weighted TSE image (TR = 10,810 ms; TE = 81 ms; flip angle = 119°; matrix size = 512 × 512; FOV = 203 × 203 mm2, 60 coronal slices angled perpendicular to the hippocampal long axis, slice thickness of 2 mm, resulting in 0.4 × 0.4 × 2.0 mm3 voxels; acquisition time = 5.47 min), and a manually labeled image for hippocampal subfields including CA1-3, SUB, and CA4/DG (Kulaga-Yoskovitz et al., 2015). All T1w and T2w images underwent automated correction for intensity non-uniformity and intensity standardization. All images were linearly registered to the MNI152 space and resampled to a resolution of 0.4 × 0.4 × 0.4 mm3. Following the previous work (Romero et al., 2017), we named this dataset as Kulaga-Yoskovitz dataset. Figure 1 shows an example of T1w image and manual hippocampal subfield segmentation from the BCP dataset and the Kulaga-Yoskovitz dataset, respectively.

FIGURE 1

Figure 1. T1w image and manual segmentation of a representative subject from the BCP dataset (top row) and Kulaga-Yoskovitz dataset (bottom row), respectively.

To facilitate the processing, we identified a bounding box that is big enough to cover the hippocampus of unseen target subject (Hao et al., 2014). In particular, for each subject in the BCP dataset and the Kulaga-Yoskovitz dataset, we went through all the training subjects to find the minimum and maximum x, y, z positions of the hippocampus, and empirically add 32 voxels in each direction as a bounding box to cover the hippocampus and its surrounding tissues. This step was done separately for these two datasets given the large hippocampus size differences in infants and adults. Then, we cropped all images with the box and applied a histogram matching method to the cropped images for obtaining similar intensity levels across all training subjects. To leverage the limited data, we left-right flipped each training image to double the number of training subjects.
Methods
We propose a new FCN for hippocampal subfield segmentation. The FCN based segmentation methods can implement dense prediction by estimating the posterior probabilities for each voxel. Given the posterior probability pk(x|θ) of voxel x belonging to the kth category, where θ is the FCN model parameters, the hippocampal subfield label of voxel x is determined by
L(x)=argmaxk∈ℂ pk(x|θ),
where ℂ = {1, 2, …, K}, and K is the number of categories. In the remaining part of this section, we will introduce the details of our proposed FCN architectures and its loss function.
Dilated Dense Network
Recent 3D neural networks often use small convolutional kernels with size 3 × 3 × 3 to reduce the number of parameters, and enlarge the receptive field by alternating convolutions and pooling operations to capture large image contexts (Çiçek et al., 2016). This successive down-sampling process will significantly reduce spatial resolution, which will lose detailed image information. Recently, dilated convolutions were proposed for semantic image segmentation (Liang-Chieh et al., 2015; Yu and Koltun, 2015). By using the dilated convolutions, the feature maps can be computed with a high spatial resolution, and the size of the receptive field can be enlarged arbitrarily. Figure 2 illustrates the dilated convolutional kernels with different dilation rates. Let F:ℤ3 → ℝ be a 3 dimensional discrete function, and h:Ωr → ℝ be a discrete filter with a dilation rate l, where Ωr=[-r,r]3⋂ℤ3. The dilated convolution *l can be defined as (Yu and Koltun, 2015),
(F*lh)(p)=∑s+lt=pF(s)h(t).    (1)
Note that, when l = 1, the dilated convolution becomes the normal convolution.

FIGURE 2

Figure 2. Illustration of dilated convolutional kernels: 1-dilated convolutional kernel (left); 2-dilated convolutional kernel (middle); 4-dilated convolutional kernel (right).

With the dilated convolutions, we design a dilated dense network using dense connections (Huang et al., 2016), as shown in Figure 3. In the dilated dense network, we use dilated convolutions with different dilation rates to enlarge the receptive field, and use dense connections to concatenate all previous generated features to the current feature maps. To avoid overfitting, dropout operations are used after each 3 × 3 × 3 convolution with dropout rate 0.5 (Srivastava et al., 2014). Thus, the dilated dense network can capture contextual image information while keeping high spatial resolution and generate multi-scale image features. This dilated dense network will be embedded in our proposed DUnet, as introduced in the next subsection.

FIGURE 3

Figure 3. The structure of the dilated dense network. The number in each operation rectangle is the number of kernels. All operations are implemented in a 3D manner, and “c” denotes the concatenation.

Dilated Dense U-Net
U-net (Ronneberger et al., 2015) consists of a contracting path to extract abstract features and an expanding path to recover spatial resolution. The features in the contracting path are concatenated to the corresponding features in the expanding path to provide the detailed image information that is lost during the successive down-sampling steps. However, the level of features in the contracting path is much lower than that in the expanding path. It will not obtain the optimal results when directly concatenating these features. To overcome this limitation, we embed the dilated dense network in the U-net to obtain a new network (DUnet). Figure 4 shows the structure of our proposed DUnet.

FIGURE 4

Figure 4. The structure of our proposed DUnet. The number in each operation rectangle is the number of kernels. All operations are implemented in a 3D manner.

Same to U-net, the proposed DUnet consists of a contracting path and an expanding path. The contracting path is built by alternating two 3 × 3 × 3 convolutions and one 2 × 2 × 2 max pooling operation with stride 2. The contracting path is followed by two 3 × 3 × 3 convolutions. Correspondingly, the expanding path is built by alternating one 4 × 4 × 4 deconvolution with stride 2, and two 3 × 3 × 3 convolutions. The expanding path is then followed by a 1 × 1 × 1 convolution, which outputs K feature maps (K is the number of label categories including the background). Each 3 × 3 × 3 convolution is followed by a batch normalization layer and a rectified linear unit (ReLU). Different from the original U-net, some padded convolution layers are also used to maintain the spatial dimension.
The feature maps before the first pooling layer and the last pooling layer are concatenated to the corresponding feature maps in the expanding path. The feature maps before the second pooling layer are first input into the dilated dense network which is introduced in the last subsection of this paper. Then, the output features of the dilated dense network are concatenated to the corresponding feature maps in the expanding path. The dilated dense network can provide multi-scale features while remaining high spatial resolution. Moreover, two different kinds of features provided by the dilated dense network and the contracting-expanding path are fused, providing more abundant image information for dense prediction.
Residual Dilated Dense U-net
To further improve the performance, we use residual connections in DUnet to promote the information flow within the network (He et al., 2016a). Formally, the residual connection can be expressed as:
xl=Hl(xl-1)+xl-1,
where xl−1 and xl are the input and output of the lth unit, and Hl(·) is a non-linear function which is used to learn the residual xl − xl−1 of the lth unit. We group every pair of convolutional layers with one residual connection along the contracting path and the expanding path of DUnet, and obtain the Residual DUnet (ResDUnet). Figure 5 shows the structure of our proposed ResDUnet. The difference between ResDUnet and DUnet is the use of residual connections in ResDUnet, which connects two adjacent convolutions with an identity mapping (or a 1 × 1 × 1 convolution if the number of feature maps is not matched).

FIGURE 5

Figure 5. The structure of our proposed ResDUnet. The number in each operation rectangle is the number of kernels. “⊕” denotes the element-wise summation, and all operations are implemented in a 3D manner.

Loss Function
We train our models using Softmax loss (Gu et al., 2017):
LSoftmax=-∑i=1N∑k=1K1{yi=k}logezk,i∑j=1Kezj,i,
where zk, i represents the kth output of the last network layer for the ith voxel, yi ∈ {1, 2, …, K} represents the corresponding ground-truth label, K and N are the number of categories and the number of voxels, respectively. The term ezk,i∑j=1Kezj,i represents the prediction probability for the kth class of the ith voxel, which is computed by the Softmax function.
Evaluation Metrics
We evaluated the image segmentation results based on two types of segmentation evaluation measures (Jafari-Khouzani et al., 2011): Dice coefficient (Dice) and Average Symmetric Surface Distance (ASSD). Dice is used to measure the relative volumetric overlap between the automated segmentation and the manual segmentation, and ASSD is used to measure the agreement between segmentation boundaries. By denoting A as the manual segmentation, B as the automated segmentation, and V(X) as the volume of segmentation X, the two evaluation measures are defined as:
  Dice = 2V(A∩B)V(A)+V(B),  ASSD = (meane∈∂A(minf∈∂Bd(e,f))+meane∈∂B(minf∈∂Ad(e,f)))/2,
where ∂A denotes the boundary voxels of A, and d(·, ·) is the Euclidian distance between two points.
Experiments and Results
Experimental Details
Five-fold cross validation was used in the experiment for the BCP dataset. In each fold, we selected 7 subjects for training, 1 subject for validation, and 2 subjects for testing. Experiments were performed using a NVIDIA Titan Xp with 12 GB memory. Because of the restriction of limited training subjects and GPU memory, we randomly extracted patches from each training subject, instead of using the whole images as input for each network. We extracted about 1,300 patches from each subject. These patches were extracted as follows. First, we extracted patches one by one with stride of 2 × 2 × 2. The extracted patches that contain at least one hippocampal voxel were taken, and were numbered as 1, 2,…, n. Then, these numbers were randomly reordered. At last, we took the first half part of the reordered patches as our training patches. The patch size was optimally set to 24 × 24 × 24 by comparing the results obtained by the baseline 3D U-net method with different patch sizes, which is shown in Table 2. Since both T1w and T2w images were available, we concatenated the corresponding T1w and T2w image patches as input for each network. The networks were trained by Adam method with a batch size of 5, which were implemented with Caffe (Jia et al., 2014). The learning rates were initially set to 0.0001 and were decreased by a factor of γ = 0.1 every 10,000 iterations. We used a weight decay of 0.0005 and a momentum of 0.9 in all networks. The training process was stopped after 60,000 iterations. For segmenting a testing image, patches were extracted to feed into the trained models with an overlapped sliding windows strategy. The patch size was set to 24 × 24 × 24 with stride of 8 × 8 × 8. We used a majority voting strategy for the overlap regions to get the whole image prediction. Note that we used the same hyper-parameters during the 5-fold cross-validation.

TABLE 2

Table 2. Mean (STD) values of Dice for each subfield segmentation using different patch sizes (R×R×R) on the BCP dataset by 3D U-net.

As the networks are trained based on image patches extracted around the hippocampus, the global spatial information of brain structures may not be perfectly captured. Thus, the obtained network models can well-recognize the hippocampal subfields around the hippocampus, but cannot recognize those far away from hippocampal region. For example, a patch in the caudate (denoted by the pink circle in the left of Figure 6) may look similar to the patches in the hippocampus, and will be classified to hippocampal subfields in the testing stage. As a result, there are some isolated false positives outside the hippocampal region, as shown in Figure 6. To remove these artifacts automatically, our post-processing steps include searching the voxels of each automated segmentation to find the non-zero neighbors of current voxel, and to obtain several connected regions. Then, we selected two regions with maximum volumes for the final left and right hippocampal subfields.

FIGURE 6

Figure 6. An example of isolated tiny blocks, outside the hippocampal region, appeared in the automated segmentation.

Five-fold cross validation was also used in the Kulaga-Yoskovitz dataset. In each fold, we selected 15 subjects for training, 5 subjects for validation, and 5 subjects for testing. The same experimental settings were used as the BCP dataset, except that the patch size was set to 32 × 32 × 32 as the resolution of images in this dataset is much higher, and the batch size was set to 3 because of the GPU memory limit. The same post-processing was used to remove isolated tiny blocks outside the hippocampal region.
The Efficacy of Multi-Modality
In this subsection, we studied the efficacy of multi-modality by comparing the segmentation results obtained using only single modality images (i.e., T1w or T2w) and multi-modality images (T1w+T2w), respectively. All experiments were carried out on the BCP dataset with the same network architecture (ResDUnet) and the same training strategies. Table 3 lists the Dice coefficients of segmentation results using different image modalities. It shows that training using multi-modality images can obtain better results in the most subfields, compared with those using only either T1w or T2w single-modality images. This demonstrates that the network trained with multi-modality images can generate more discriminative features, which improves the performance of hippocampal subfield segmentation. From the results, we also find that T1w images can provide more useful information than T2w images for hippocampal subfield segmentation on the BCP dataset. In some subfields, training using only T1w images obtains similar or even a little better segmentation results than those using multi-modality images.

TABLE 3

Table 3. Mean (STD) values of Dice for each subfield segmentation using different modalities on the BCP dataset.

Comparison With State-of-the-Art Methods
Our proposed method was also compared with two state-of-the-art networks, namely, 3D U-net (Çiçek et al., 2016) and ConvNet (Yu et al., 2017). The 3D U-net is extended from the previous 2D version (Ronneberger et al., 2015) into a 3D variant for volumetric feature representation. For a fair comparison, the 3D U-net used in our experiments consists of three pooling layers and three deconvolutional layers, which are the same as our proposed DUnet. The only difference is that the dilated dense network is used to fuse the middle level features of the contracting path with those of the expanding path in DUnet, instead of directly concatenating them as in 3D U-net. ConvNet (Yu et al., 2017) is a volumetric convolutional neural network with mixed residual connections, which also consists of three pooling layers and three deconvolutional layers. In ConvNet, residual connections are used between the successive convolution layers to form the residual blocks, and also between the feature maps of contracting path and those of expanding path. Besides, ConvNet (Yu et al., 2017) exploits a deep supervision mechanism to accelerate its convergence speed. All these comparative networks use Softmax loss as loss function, and the same post-processing is used to remove the tiny isolated blocks of segmentation results that appear outside of the hippocampal region.
Table 4 reports the Dice coefficients of the segmentation results obtained by different networks on the BCP dataset. It shows that our proposed DUnet outperforms 3D U-net (Çiçek et al., 2016) in segmenting CA1, SUB, CA4/DG and Uncus, and our proposed ResDUnet outperforms 3D U-net (Çiçek et al., 2016) in segmenting CA1, CA2/3, SUB, and Uncus, according to the Wilcoxon signed rank tests with p < 0.05. As can be seen in the table, our proposed ResDUnet achieves the highest Dice coefficient for the average of subfields. Table 5 reports the ASSD coefficients of the segmentation results, which shows that our proposed ResDUnet achieves the best ASSD coefficient for the average of subfields. Figure 7 shows hippocampal subfield segmentations of a randomly selected subject from the BCP dataset, obtained by manual segmentation and four different networks. It can be seen that our proposed ResDUnet achieves the most accurate results.TABLE 4Table 4. Mean (STD) values of Dice for each subfield segmentation by different networks on the BCP dataset.TABLE 5Table 5. Mean (STD) values of ASSD for each subfield segmentation by different networks on the BCP dataset.FIGURE 7Figure 7. Hippocampal subfield segmentations of a randomly selected subject from the BCP dataset, obtained by manual segmentation, and four different networks.Results on a Public Adult DatasetTables 6, 7 list the Dice and ASSD coefficients of the segmentation results obtained by five different networks on the Kulaga-Yoskovitz dataset. The results show that our proposed DUnet outperforms 3D U-net (Çiçek et al., 2016) and ConvNet (Yu et al., 2017) in segmenting CA1-3 and SUB, and our proposed ResDUnet outperforms 3D U-net (Çiçek et al., 2016) and ConvNet (Yu et al., 2017) in segmenting all subfields, according to the Wilcoxon signed rank tests with p < 0.05. Table 6 also lists the comparison of our proposed method with the state-of-the-art hippocampal subfield segmentation method (HIPS), which obtained the best segmentation results on the Kulaga-Yoskovitz dataset so far (Romero et al., 2017). Note that, for a fair comparison, we use the published results of HIPS as reported in Romero et al. (2017). It shows that our proposed DUnet and ResDUnet also outperform HIPS method, especially for segmenting the CA4/DG subfield which is the most difficult task (Dalton et al., 2017). Figure 8 shows hippocampal subfield segmentations of a randomly selected subject from Kulaga-Yoskovitz dataset, obtained by manual segmentation and four different networks. It can be seen that our proposed DUnet and ResDUnet achieve the most accurate results.TABLE 6Table 6. Mean (STD) values of Dice for each subfield segmentation by five different methods on the KULAGA-YOSKOVITZ dataset.TABLE 7Table 7. Mean (STD) values of ASSD for each subfield segmentation by four different networks on the KULAGA-YOSKOVITZ dataset.FIGURE 8Figure 8. Hippocampal subfield segmentations of a randomly selected subject from the Kulaga-Yoskovitz dataset, obtained by manual segmentation, and four different networks.DiscussionFCNs have achieved great success in the field of medical image segmentation, which usually consist of a contracting path to extract abstract features, and an expanding path to up-sample the feature maps for dense prediction (Ronneberger et al., 2015; Çiçek et al., 2016; Chen H. et al., 2017; Lian et al., 2018; Nie et al., 2018). The detailed image information may be lost during these contracting and expanding processes. The existing U-net-like FCNs concatenate the feature maps in the contracting path to the corresponding feature maps in the expanding path to recover the lost image information. However, the levels of features in the contracting path are much lower than those in the expanding path. It may not obtain the optimal results when directly concatenating these features. To overcome this limitation and fully exploit multi-level image features, we proposed a new FCN by exploiting a dilated dense network to connect the features of the contracting path and the features of the expanding path. The dilated dense network uses the dilated convolutions to extract contextual features without reducing spatial resolution, and it also employs dense connections to aggregate multi-scale features. Thus, multi-scale features can be generated from the dilated dense network, which are fused with the corresponding features in the expanding path. To avoid overfitting, dropout operations are also used in the dilated dense network (Srivastava et al., 2014).By using the dilated dense network to connect the feature maps in the contracting path and expanding path, our proposed method provides a way to fuse the finer-grained low-level features in the contracting path and the coarse high-level features in the expanding path. Moreover, the multi-scale features extracted by the dilated dense network are useful for segmenting multi-structures with different shapes and different scales. To further promote information propagation and accelerate the convergence, we introduce residual connections to group every pair of convolutional layers (He et al., 2016a,b).Different from natural images, many imaging modalities are 3D in the field of medical image analysis. In the past few years, a lot of effort has been dedicated to exploit CNNs to process volumetric data. Some of them applied 2D CNNs to each slice of volumetric images (Prasoon et al., 2013; Setio et al., 2016; Chen Y. et al., 2017). To effectively make full use of the 3D spatial information, recent studies applied 3D CNNs to deal with volumetric images (Çiçek et al., 2016; Chen H. et al., 2017;Nie et al., 2018; Wachinger et al., 2018). Following these methods, our proposed FCNs were also implemented in a 3D manner. As the number of our training subjects is limited, we randomly extracted patches from each training subject, instead of using the whole image as the input for each network. The patch size was set to 24 × 24 × 24 for the BCP dataset and 32 × 32 × 32 for Kulaga-Yoskovitz dataset, considering different image resolutions in these two datasets.As both T1w and T2w images were available for each subject, we concatenated the extracted T1w and T2w image patches as input to the networks. Compared with single modality data, multi-modality MR images can provide complementary contextual information, which contributes to better segmentation performance. From our experiments, we find that training using multi-modality images can obtain better results than using only single-modality images, and we also find that T1w images can provide more discriminative information than T2w images for hippocampal subfield segmentation.Experimental results on the BCP dataset show that our proposed DUnet and ResDUnet improve the average Dice coefficient by 2.1 and 2.5%, respectively, for infant hippocampal subfield segmentation, compared with the 3D U-net (Çiçek et al., 2016). To further validate the effectiveness, we also applied our proposed method for adult hippocampal subfield segmentation based on a publicly available dataset. The results show that our proposed DUnet and ResDUnet improve the average Dice coefficients of 0.5 and 0.6%, respectively, compared with the 3D U-net (Çiçek et al., 2016). The improvement of our proposed ResDUnet method on both infant dataset and adult dataset comes from (1) multi-scale image features aggregation for distinguishing different hippocampal subfields; (2) utilization of the embedded dilated dense network for effectively fusing the low-level features in the contracting path and the high-level features in the expanding path; and (3) utilization of residual connections for promoting information propagation and accelerating the convergence.However, the proposed method was mainly designed for infant hippocampal subfield segmentation on the BCP dataset. First, the embedded dilated dense network can provide multi-scale image features, which are especially useful for segmenting infant hippocampal subfields, since tissue contrast between infant hippocampal subfields are much blurrier than in adults. Second, the task of infant hippocampal subfield segmentation on the BCP dataset is to segment hippocampus into five parts (CA1, CA2/3, SUB, CA4/ DG, and Uncus), while there are only three parts (CA1-3, SUB, and CA4/DG) on the Kulaga-Yoskovitz dataset. Therefore, the segmented hippocampal subfields in the infant subjects are much smaller than those of the adult subjects. In our proposed network, the embedded dilated dense network can capture contextual image information without losing detailed image information, which is extremely useful for segmenting small structures.ConclusionIn this paper, we have proposed a new FCN by integrating U-net and dilated dense network for hippocampal subfield segmentation. Our proposed method can avoid losing the detailed image information in the successive down-sampling steps, effectively fuse the low-level features of the contracting path with the coarse high-level features of the expanding path, and generate multi-scale image features. Experimental results show that our proposed method outperforms the state-of-the-art methods.Ethics StatementThis study was approved by the Institutional Review Board in School of Medicine, the University of North Carolina (UNC) at Chapel Hill.Author ContributionsHZ, FS, LW, SW, and DS contributed to the method development and the manuscript preparation. WL contributed to providing infant imaging data, and S-CH and M-HC did manual labeling of infant hippocampal subfields. HZ undertook the statistical analysis.Conflict of Interest StatementFS was employed by company Shanghai United Imaging Intelligence Co., Ltd., China.The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.AcknowledgmentsThis work utilizes data collected by a NIH grant (1U01MH110274) and the efforts of the UNC/UMN Baby Connectome Project Consortium. HZ was partially supported by National Natural Science Foundation of China (Nos. 61602307, 61877039), and Natural Science Foundation of Zhejiang Province (No. LY19F020013).References Bartsch, T. (2012). The Clinical Neurobiology of the Hippocampus: An Integrative View. Vol. 151. Oxford: Oxford University Press. doi: 10.1093/acprof:oso/9780199592388.001.0001CrossRef Full Text | Google Scholar Caldairou, B., Bernhardt, B. C., Kulaga-Yoskovitz, J., Kim, H., Bernasconi, N., and Bernasconi, A. (2016). “A surface patch-based segmentation method for hippocampal subfields,” in Paper Presented at: International Conference on Medical Image Computing and Computer-Assisted Intervention (Athens).Google Scholar Cao, L., Li, L., Zheng, J., Fan, X., Yin, F., Shen, H., et al. (2018). Multi-task neural networks for joint hippocampus segmentation and clinical score regression. Multimedia Tools Applic. 77, 29669–29686. doi: 10.1007/s11042-017-5581-1CrossRef Full Text | Google Scholar Chen, H., Dou, Q., Yu, L., Qin, J., and Heng, P.-A. (2017). VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images. Neuroimage 170, 446–455. doi: 10.1016/j.neuroimage.2017.04.041PubMed Abstract | CrossRef Full Text | Google Scholar Chen, Y., Shi, B., Wang, Z., Zhang, P., Smith, C. D., and Liu, J. (2017). “Hippocampus segmentation through multi-view ensemble ConvNets,” in 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017) (Melbourne, VIC) 192–196.Google Scholar Çiçek, Ö., Abdulkadir, A., Lienkamp, S. S., Brox, T., and Ronneberger, O. (2016). “3D U-Net: learning dense volumetric segmentation from sparse annotation,” in Paper presented at: International Conference on Medical Image Computing and Computer-Assisted Intervention (Athens).Google Scholar Dalton, M. A., Zeidman, P., Barry, D. N., Williams, E., and Maguire, E. A. (2017). Segmenting subregions of the human hippocampus on structural magnetic resonance image scans: an illustrated tutorial. Brain Neurosci. Adv. 1:2398212817701448. doi: 10.1177/2398212817701448PubMed Abstract | CrossRef Full Text | Google Scholar de Brébisson, A., and Montana, G. (2015). “Deep neural networks for anatomical brain segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (Boston, MA), 20–28.Google Scholar Gu, J., Wang, Z., Kuen, J., Ma, L., Shahroudy, A., Shuai, B., et al. (2017). Recent advances in convolutional neural networks. Pattern Recogn. 77, 354–377. doi: 10.1016/j.patcog.2017.10.013CrossRef Full Text | Google Scholar Hao, Y., Wang, T., Zhang, X., Duan, Y., Yu, C., Jiang, T., et al. (2014). Local label learning (LLL) for subcortical structure segmentation: Application to hippocampus segmentation. Hum. Brain Mapp. 35, 2674–2697. doi: 10.1002/hbm.22359PubMed Abstract | CrossRef Full Text | Google Scholar He, K., Zhang, X., Ren, S., and Sun, J. (2016a). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV), 770–778.Google Scholar He, K., Zhang, X., Ren, S., and Sun, J. (2016b). “Identity mappings in deep residual networks,” in European Conference on Computer Vision (Amsterdam: Springer), 630–645.Google Scholar Huang, G., Liu, Z., van der Maaten, L., and Weinberger, K. Q. (2016). “Densely connected convolutional networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI), 4700–4708.Google Scholar Jafari-Khouzani, K., Elisevich, K. V., Patel, S., and Soltanian-Zadeh, H. (2011). Dataset of magnetic resonance images of nonepileptic subjects and temporal lobe epilepsy patients for validation of hippocampal segmentation techniques. Neuroinformatics 9, 335–346. doi: 10.1007/s12021-010-9096-4PubMed Abstract | CrossRef Full Text | Google Scholar Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, S., et al. (2014). “Caffe: convolutional architecture for fast feature embedding,” Proceedings of the 22nd ACM International Conference on Multimedia (Orlando, FL: ACM), 675–678.Google Scholar Kulaga-Yoskovitz, J., Bernhardt, B. C., Hong, S. J., Mansi, T., Liang, K. E., van der Kouwe, A. J., et al. (2015). Multi-contrast submillimetric 3 Tesla hippocampal subfield segmentation protocol and dataset. Sci. Data 2:150059. doi: 10.1038/sdata.2015.59PubMed Abstract | CrossRef Full Text | Google Scholar LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539PubMed Abstract | CrossRef Full Text | Google Scholar Li, J., Jin, Y., Shi, Y., Dinov, I. D., Wang, D. J., Toga, A. W., et al. (2013). “Voxelwise spectral diffusional connectivity and its applications to Alzheimer's disease and intelligence prediction,” in Paper Presented at: International Conference on Medical Image Computing and Computer-Assisted Intervention (Nagoya).Google Scholar Li, W., Wang, G., Fidon, L., Ourselin, S., Cardoso, M. J., and Vercauteren, T. (2017). “On the compactness, efficiency, and representation of 3D convolutional networks: brain parcellation as a pretext task,” in International Conference on Information Processing in Medical Imaging (Boone, NC), 348–360.Google Scholar Lian, C., Zhang, J., Liu, M., Zong, X., Hung, S. C., Lin, W., et al. (2018). Multi-channel multi-scale fully convolutional network for 3D perivascular spaces segmentation in 7T MR images. Med. Image Anal. 46, 106–117. doi: 10.1016/j.media.2018.02.009PubMed Abstract | CrossRef Full Text | Google Scholar Liang-Chieh, C., Papandreou, G., Kokkinos, I., Murphy, K., and Yuille, A. (2015). “Semantic image segmentation with deep convolutional nets and fully connected crfs,” in International Conference on Learning Representations (San Diego, CA).Google Scholar Litjens, G., Kooi, T., Bejnordi, B. E., Setio, A. A. A., Ciompi, F., Ghafoorian, M., et al. (2017). A survey on deep learning in medical image analysis. Med. Image Anal. 42, 60–88. doi: 10.1016/j.media.2017.07.005PubMed Abstract | CrossRef Full Text | Google Scholar Long, J., Shelhamer, E., and Darrell, T. (2015). “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Boston, MA), 3431–3440.Google Scholar McKinley, R., Jungo, A., Wiest, R., and Reyes, M. (2017). “Pooling-free fully convolutional networks with dense skip connections for semantic segmentation, with application to brain tumor segmentation,” in International MICCAI Brainlesion Workshop (Quebec, QC), 169–177.Google Scholar Milletari, F., Navab, N., and Ahmadi, S-A. (2016). “V-net: fully convolutional neural networks for volumetric medical image segmentation,” in 2016 Fourth International Conference on 3D Vision (3DV) (Stanford, CA), 565–571.Google Scholar Moeskops, P., Viergever, M. A., Mendrik, A. M., de Vries, L. S., Benders, M. J., and Išgum, I. (2016). Automatic segmentation of MR brain images with a convolutional neural network. IEEE Trans. Med. Imaging 35, 1252–1261. doi: 10.1109/TMI.2016.2548501PubMed Abstract | CrossRef Full Text | Google Scholar Nie, D., Wang, L., Adeli, E., Lao, C., Lin, W., and Shen, D. (2018). 3-D fully convolutional networks for multimodal isointense infant brain image segmentation. IEEE Trans. Cybernet. 99, 1–14. doi: 10.1109/TCYB.2018.2797905CrossRef Full Text | Google Scholar Pipitone, J., Park, M. T. M., Winterburn, J., Lett, T. A., Lerch, J. P., Pruessner, J. C., et al. (2014). Multi-atlas segmentation of the whole hippocampus and subfields using multiple automatically generated templates. Neuroimage 101, 494–512. doi: 10.1016/j.neuroimage.2014.04.054PubMed Abstract | CrossRef Full Text | Google Scholar Prasoon, A., Petersen, K., Igel, C., Lauze, F., Dam, E., and Nielsen, M. (2013). “Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (Nagoya), 246–253.Google Scholar Romero, J. E., Coupé, P., and Manjón, J. V. (2017). HIPS: a new hippocampus subfield segmentation method. Neuroimage 163, 286–295. doi: 10.1016/j.neuroimage.2017.09.049PubMed Abstract | CrossRef Full Text | Google Scholar Ronneberger, O., Fischer, P., and Brox, T. (2015). “U-net: convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (Munich), 234–241.Google Scholar Setio, A. A. A., Ciompi, F., Litjens, G., Gerke, P., Jacobs, C., van Riel, S. J., et al. (2016). Pulmonary nodule detection in CT images: false positive reduction using multi-view convolutional networks. IEEE Trans. Med. Imag. 35, 1160–1169. doi: 10.1109/TMI.2016.2536809PubMed Abstract | CrossRef Full Text | Google Scholar Shen, D., Wu, G., and Suk, H-I. (2017). Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 19, 221–248. doi: 10.1146/annurev-bioeng-071516-044442PubMed Abstract | CrossRef Full Text | Google Scholar Shi, F., Liu, B., Zhou, Y., Yu, C., and Jiang, T. (2009). Hippocampal volume and asymmetry in mild cognitive impairment and Alzheimer's disease: meta-analyses of MRI studies. Hippocampus 19, 1055–1064. doi: 10.1002/hipo.20573PubMed Abstract | CrossRef Full Text | Google Scholar Small, S. A. (2014). Isolating pathogenic mechanisms embedded within the hippocampal circuit through regional vulnerability. Neuron. 84, 32–39. doi: 10.1016/j.neuron.2014.08.030PubMed Abstract | CrossRef Full Text | Google Scholar Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2014). Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15, 1929–1958.Google Scholar Van Leemput, K., Bakkour, A., Benner, T., Wiggins, G., Wald, L. L., Augustinack, J., et al. (2009). Automated segmentation of hippocampal subfields from ultra-high resolution in vivo MRI. Hippocampus 19, 549–557. doi: 10.1002/hipo.20615PubMed Abstract | CrossRef Full Text | Google Scholar Wachinger, C., Reuter, M., and Klein, T. (2018). DeepNAT: Deep convolutional neural network for segmenting neuroanatomy. Neuroimage 170, 434–445. doi: 10.1016/j.neuroimage.2017.02.035PubMed Abstract | CrossRef Full Text | Google Scholar Wang, H., Suh, J. W., Das, S. R., Pluta, J. B., Craige, C., and Yushkevich, P. A. (2013). Multi-atlas segmentation with joint label fusion. Pattern Anal. Mach. Intell. IEEE Trans. 35, 611–623. doi: 10.1109/TPAMI.2012.143PubMed Abstract | CrossRef Full Text | Google Scholar Wu, Z., Gao, Y., Shi, F., Ma, G., Jewells, V., and Shen, D. (2018). Segmenting hippocampal subfields from 3T MRI with multi-modality images. Med. Image Anal. 43, 10–22. doi: 10.1016/j.media.2017.09.006PubMed Abstract | CrossRef Full Text | Google Scholar Yu, F., and Koltun, V. (2015). Multi-scale context aggregation by dilated convolutions. arXiv [preprint] arXiv:1511.07122.Google Scholar Yu, L., Yang, X., Chen, H., Qin, J., and Heng, P-A. (2017). “Volumetric ConvNets with mixed residual connections for automated prostate segmentation from 3D MR images,” in AAAI'17 Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (San Francisco, CA), 66–72.Google Scholar Yushkevich, P. A., Pluta, J. B., Wang, H., Xie, L., Ding, S. L., Gertje, E. C., et al. (2015). Automated volumetry and regional thickness analysis of hippocampal subfields and medial temporal cortical structures in mild cognitive impairment. Hum. Brain Mapp. 36, 258–287. doi: 10.1002/hbm.22627PubMed Abstract | CrossRef Full Text | Google Scholar Zhang, W., Li, R., Deng, H., Wang, L., Lin, W., Ji, S., et al. (2015). Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. Neuroimage. 108:214–224. doi: 10.1016/j.neuroimage.2014.12.061PubMed Abstract | CrossRef Full Text | Google ScholarKeywords: fully convolutional network, dilated dense network, deep learning, hippocampal subfield segmentation, infant hippocampusCitation: Zhu H, Shi F, Wang L, Hung S-C, Chen M-H, Wang S, Lin W and Shen D (2019) Dilated Dense U-Net for Infant Hippocampus Subfield Segmentation. Front. Neuroinform. 13:30. doi: 10.3389/fninf.2019.00030Received: 07 February 2019; Accepted: 02 April 2019;Published: 24 April 2019.Edited by: Tianzi Jiang, Institute of Automation (CAS), ChinaReviewed by: Liang Zhan, University of Pittsburgh, United StatesYalin Wang, Arizona State University, United StatesCopyright © 2019 Zhu, Shi, Wang, Hung, Chen, Wang, Lin and Shen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.*Correspondence: Dinggang Shen, dgshen@med.unc.edu

Read Full Article (External Site)