1.IntroductionNeurons are the basic units of the structure and function of the nervous system.1 And the neuronal morphology reflects the function of the brain. A complete segmentation of neuronal structure from optical microscopy (OM) images is essential to reveal the connections of brain circuits and investigate synaptic integration, neuron morphology, neuron connectivity, and brain mechanisms.2 Analyzing the morphology and connection of neurons enables us to have a deeper understanding of the operating mechanism of brains and facilitates diagnosis of brain disease.3 For example, congenital nystagmus affects starburst amacrine cells,4 and amyotrophic lateral sclerosis affects upper and lower motor neurons.5 Alzheimer’s disease (AD) is a neurodegenerative disease associated with synaptic loss and neuronal degeneration.6 A precise segmentation of 3D neuron structures is vital to probe impaired brain functions of AD animal models and to determine early treatment strategies.7 We might gain a better understanding of such disorders if we could find specific neuronal morphology in disease models.8 Through the accurate signals of neuron morphology, we can systematically classify brain cells. In recent years, labeling techniques9,10 and optical imaging methods11,12 have made a series of breakthroughs, which can generate terabyte (TB) level neuronal morphological data at whole brain scale.13,14 Micro-optical sectioning tomography (MOST)15 and its follow-up study fluorescence micro-optical sectioning tomography (fMOST)16 have realized the whole brain imaging of specific neural structures at submicron resolution, including tens of thousands of complete neuronal morphologies. Therefore, it is highly difficult to extract and analyze neuron morphology in neuron tracing caused by the uneven distribution of fluorescence,17 differences of imaging system,18 complex morphology of neurons,19 and low signal-to-noise ratio (SNR).20 Currently, many efforts have been devoted to develop automatic or semi-automatic neuron tracing algorithms. However, most of the neuron tracing methods are not applicable in challenging datasets where the 3D neurites are contaminated by strong background noises or containing weak light signals. These methods degrade in performance of neuron segmentation from more complex image blocks. Therefore, an efficient segmentation method is urgently needed to reduce the impact of noises and enhance the weak neurite signals, which should improve the result of neuron tracing. In recent years, deep learning has achieved remarkable success in processing natural images. Convolutional neural networks (CNNs) have been employed to segment tubular structures due to its capability of feature learning and its nonlinear relationship capturing between inputs and outputs. Deep learning techniques have been adopted in neuron segmentation to improve the segmentation quality. A deep learning toolbox, DeepNeuron, was designed to trace neurons with manually reconstructed neurons as training samples.21 The general supervised deep learning algorithms require a large number of manual labels for reaching an efficient and accurate neuron segmentation, but the segmentation labeling is pretty expensive and labor-consuming. Moreover, extensive a priori knowledge of neuroscience is also required for manual labels. To solve the above problems, we develop a weakly supervised learning framework [Docker-based deep network (DDeep3M+)] for neuron segmentation from 3D OM images without any manual labeling. It can extract accurate neuron signals robustly from noisy backgrounds. We take advantage of both conventional methods and deep learning techniques as follows. An adaptive enhancement filter is first employed to produce pseudo-labels for training a deep segmentation neural network (DDeep3M). The proposed DDeep3M+ method improves prediction of DDeep3M by iteratively optimizing the training labels and retraining model via region growing and image fusion based on neurite structure information to mine undetected weak neurites from the probability map predicted by DDeep3M. The experimental results on the fMOST16 datasets and public BigNeuron22 datasets have demonstrated the high precision and generalization achieved by the proposed neuron segmentation method; see Fig. 1. Our main contributions are summarized as follows:
2.Related WorkMorphology of neurons is essential for the organization and function of the brain. Segmentation or tracing of 3D neuron structures from OM images are the main approaches to characterize and analyze neuronal morphology. There are 2,32,613 digitally reconstructed neurons contributed by 800 laboratories worldwide in NeuroMorpho19 dataset, which took more than 2.5 million hours of manual work in the past decades. In computational neuroscience, there are high requirements for automatic and accurate neuron segmentation methods. With the development of optical imaging and molecular labeling technology, we can obtain large-scale sub-micron resolution neuron images through optical imaging of mammalian brain. These advances promote the generation of various optical images for different applications, and pose new challenges in neuron segmentation. The limit is the significant differences in image quality and attributes between different datasets due to a variety of factors, such as intensity range, image size, neuron structure, etc.22 Another challenge is that microscopy images usually have high background noise. It has difficulty in separating weak neuron voxels from an inhomogeneous background.23 It is tough to distinguish the neurites with low intensity and non-uniformity from the noise background, especially for the large-scale neuron images with discontinued segments of neurites and ultra-low SNR. The variability between different datasets and the low SNR of optical images increase the difficulty of generalization and improvement of neuron segmentation algorithms. Many semi-automatic or automatic methods have been proposed for neuron segmentation or neuron tracing.24 In these algorithms, various computational methods addressing global and local image features have been used to realize neuron segmentation and tracing. These include but are not limited to region growing, tubular model, full path pruning, graph theory method, and support vector machine based on self-learning.25,26 These algorithms usually show good performance on OM images with clear structure. However, the above methods are designed for specific datasets or specific problems, and their performance on different types of datasets may degrade, so complex parameter tuning is needed. In addition, most algorithms perform poorly in tracing neurites from low SNR images, and are unable to recognize weak neurite signals or prone to over-segmentation due to background noise. Deep network27 has been widely used in many fields, including computer vision, natural language processing, speech recognition, and so on, and has achieved excellent results in these tasks.28 Since the full convolution network (FCN),29 various deep networks have been proposed to improve the accuracy of image segmentation, including U-Net,30 3D deeply supervised network (DSN),31 and VoxResNet.32 The excellent performance of deep network encourages researchers to apply it to neuron segmentation.33 In Ref. 34, a deep network is designed to improve the performance of existing tracing algorithms. The main purpose of the deep network is to denoise the image containing a single neuron or multiple sparse neurons. DeepNeuron,21 a deep learning toolbox, is designed to trace neurons with human-reconstructed neurons as training samples. But they have poor performance in the case of weak neurites. Yang et al.35 employed a two-stage 3D neuron segmentation approach to segment neuron images from the BigNeuron dataset, via learning deep features by an FCN model and enhancing weak neuronal structures based on Hessian Eigenvalues in a fibrous structure. Structure-guided segmentation network (SGSNet),36 a two-branch network for 3D neuron segmentation, contains a shared encoding path but utilizes two decoding paths to enhance weak neuronal structures and remove background noises. There are 3D CNNs specifically designed for neuron segmentation and tracing that show better performance on images with high noise.37,38 However, these algorithms require a lot of labor, time-consuming, and expensive manual annotation to segment neurons. Users need to annotate enough new samples for neuron datasets from different brain regions to achieve reliable prediction. The need for manual annotation has greatly limited the promotion of deep learning-based methods for various optical neuron images. Here, we propose a weakly supervised learning method for deep CNN without manual labeling. Integrating the advantages of traditional image enhancement and deep learning, we can extract neuron signals accurately and robustly from noisy background, and realize neuron segmentation in OM image. 3.Materials and MethodsTo segment neuron population from low SNR optical microscopic images, our neuron segmentation framework consists of four key modules: pseudo-labels module, DDeep3M segmentation network, region growing module, and image enhancement module, which are illustrated in Fig. 2. The automatic neuron segmentation framework includes four steps: (1) initial neurite segmentation is automatically obtained by an adaptive enhancement filter and threshold segmentation algorithm from optical images as their pseudo-labels, (2) training of the DDeep3M segmentation network using input images and their pseudo-labels as training sets, and prediction of the neurite as probability map using DDeep3M, (3) refinement of pseudo-labels from probability map based on region growing, (4) combining the original image intensities with the refined probability map. Steps (1) to (4) are iterated to update the probability map. The segmentation network is optimized, and the completeness and accuracy of neuron segmentation are improved. By fusing the prediction probability map of neuron voxel with the original intensity to enhance the original image, we can preserve the global structure from probability map and local signal from original image at the same time. We gradually improve the performance of neuron segmentation by combining the traditional image enhancement method and the DDeep3M network without any manual annotation. Finally, we complete the automatic segmentation of large-scale neuronal population from noisy and low-contrast OM images through whole framework. 3.1.Pseudo-LabelsThere are some problems in the 3D optical images of neurons, such as weak signal, strong noise, and uneven signal distribution. All these problems seriously restrict neuron segmentation, so we enhance the original image to remove the noise in the first place. Usually, a neuron consists of a soma and a great number of fibers. The fibers are formed from an axon and dendrites. Seeing a neuron in mathematical graphical structure, we can understand that a cell body is a dot-like structure and a neurite is a fiber-like or line-like structure.39 We propose an adaptive enhancement for the 3D image of neurons based on distance transform (DT) and Hessian matrix. The purpose of enhancement filters for neuron is to get clearer and higher contrast images, and then the neuron can be segmented after being enhanced. The Hessian matrix is a mixed second derivate matrix, which simply indicates that the gray intensity trends of continuous voxels in a 3D image. A 3D raw OM image volume was presented as . Hessian matrix was defined as the second order derivatives of the image intensity where is a Gaussian kernel. denotes the first order Gaussian derivatives. is convolution. denotes the mixed second derivative along dimensions and .First, the window size of the Hessian matrix is determined by DT: DT can measure the thickness of different neuron fibers effectively, and it can be used to calculate the Hessian matrix. Second, the method constructs a structural response function to enhance the neuron fibers: when the image signal is bright and the background is dark, the Hessian matrix eigenvalue is approximately equal to 0, and are and their amplitudes are close to each other. The method takes advantage of this characteristic to enhance the neuron fibers. Third, the method enhances the soma with different strategy: the signal value of the soma is high and the radius is large. Accordingly, this method sets a higher threshold for DT to localize the soma for hole filling. Through the above three steps, the enhancement of neuron with different structures is finally achieved. Through a Hessian analysis-based neuron enhancement and neurite threshold segmentation, the initial neurite segmentation is automatically obtained, and then they are regarded as initial training labels (pseudo-labels). We devise a way to get an adaptive window size, which can guarantee that the Hessian matrix has a proper window size in our image of uneven intensity. First, we distinguished the foreground regions and background regions with a determined value derived from a priori knowledge, the image was transformed to be binary. Then we used the Euclidean DT to describe the degree of the thickness. Here, the Euclidean DT between and is . Here and are -tuples ( in this study), and are the ’th coordinates of and , in Euclidean distances. The purpose of computing the DT is to determine the window size of the Hessian matrix. As a second-order partial derivative matrix, the Hessian matrix needs to determine how many points to calculate when calculating between voxels. The size of the Hessian matrix window is closely related to the thickness of the nerve fibers. It has a great influence on the subsequent nerve fiber enhancement. First, the center point of a cell body is set as . After the Euclidean DT, the DT value of the cell center point is set to . Next, the window size of the Hessian matrix is calculated by the obtained DT value. Then the window radius is determined by the following formula: . The intensity of neurites in the middle is higher than that at the edge. DT image reveals the structure information of the graph, to a certain extent. We made use of this useful information and defined an optimal window radius for each voxel. The DT value for each location is defined as , and all the DT values for each voxel in the image build a DT matrix, which has the same size as the original image. Then we normalize between 1 and 256. We defined the normalized result for each voxel as DN ). We make a logarithmic operation for DN, the result of it comes to be the window radius. Indeed, in this way, we basically used a window with radius between 1 and 8. The window size, that is, window diameter DI is defined as twice the radius and plus one. The above formulas are as follows: , . It can provide a proper size for each voxel for the uneven gray-scale characteristics of different locations in neurites to calculate the Hessian matrix. Since the Hessian matrix is symmetric, there exist three eigenvalues , , and . For convenience, they are sorted based on the absolute value: . In previous studies, such as Frangi’s40 method in 1998 when the signal is bright and the background is dark, the Eigenvalues of an ideal 3D dot-like, line-like, and plate-like area satisfy the following conditions: line-like: ; ; dot-like: ; plate-like: ; . Initially, taking account of the neurite enhancement, Frangi’s filter proposed three derived parameters based on the above-mentioned Eigenvalues: , , and . The response function was defined as follows: where , , and are thresholds that control the sensitivity of the line filter to the measures , , and . Through the test of multiple groups of data, we found the optimal values of parameters: , , and . When a voxel in the image is a point in a plate-like structure, approaches 0. The calculation result of is 0 without response. When a voxel is a point in a dot-like structure, approaches 0. The result of is 0 without response. When a voxel is background signal, approaches 0. The result is with no response. Therefore, nerve fibers can be specifically enhanced by the Eq. (2).The above work is for line-like structures enhancement, the soma of a neuron will be turned into a circular ring or an irregular shape with a hole,41 as shown in Step 1 of Fig. 2. To solve the problem and make our algorithm complete and useful, we must design a method to detect the holes. In the original fluorescent neuron image, the intensity of the soma and middle part of the fiber is always higher than the other pixels. We take advantage of this feature and combine the characteristic of DT, then design the following approach to detect the holes in soma and thicker fiber. First, we choose a relatively higher determined value , which is bigger than , with a priori knowledge. Then, we recalculate DT matrix with the higher threshold , defined as , the result will show the location of the holes in soma and fiber. We can use the location to fill the hole in the enhanced image by replacing it with high intensity at the same location. If , ; otherwise, it remains . Where is the voxel, is the local DT value for with higher threshold, is the threshold to determine whether a voxel is within the soma area or not. And is the enhanced result before soma detection. If the voxel is assumed as a soma voxel, we replace the intensity value. Otherwise, the previous enhanced result will be retained. The is the enhanced result of original image, and then we obtain the initial training labels (pseudo-labels) through threshold segmentation after enhancement by our Hessian analysis based adaptive enhancement filter. 3.2.Segmentation Network—DDeep3MDocker is an open source lightweight virtualization technology. DDeep3M42 integrates the CDeep3M43 network by Docker, and it can run on the local computer by downloading a separate Docker image file containing all necessary libraries, software and framework. DDeep3M greatly reduces the threshold for biomedical researchers to accept deep learning technology. It achieves excellent segmentation performance on multiple datasets and performs better than most related models, especially in the neuroscience image segmentation tasks. The architecture of DDeep3M has some similarities to the U-Net architecture,30 which consists of a bottom-up path and a resolution recovery path. We aggregate multiscale context information required for classification and minimize the effect of noise via using the -direction information. However, there are inception and residual modules in the coding path, which are some key differences from U-Net. In deep learning, deeper layers can lead to overfitting and vanishing gradient problems. The size of the receptive field is another important factor affecting network performance. The inception module allows information to pass through multiple sized cores in parallel, applying convolution reduces the number of input feature maps before using large kernel convolution to effectively reduce the complexity of the model. Residual learning is able to alleviate the gradient vanishing problem. We apply skip connections similar to those in FCN,29 but add pyramid dilated deconvolution to further enhance the ability to represent multi-scale information in the spatial recovery path. Because neurites only account for a small part of the data block, the categories of foreground (i.e., segmented neurites) and background are usually unbalanced, which may lead to prediction bias. The weighted cross-entropy (WCE) loss function44 and dice loss (DL) function45 reduce the impact of class level imbalance. We use a mixed loss function to combine these two loss functions to prevent class imbalance and maintain regional continuity. The WCE loss is Then, the DL is where is the prediction probability of pixel ; is the corresponding pseudo-labels of foreground or background, and its values are 1 or 0 respectively; is the length, is the width, and is the height of volume data; is the ratio of the number of voxels of foreground pseudo labels to the number of voxels of image volume data ; is the smoothing parameter set to 1. Thus, the combination loss function to train the model is , where is the weight of cross-entropy loss , which is used to balance the ratio of loss function.3.3.Pseudo-Label RefinementSome weak and uneven neuron signals are not detected in the pseudo-labels obtained by an adaptive enhancement filter, which reduce the network prediction accuracy of neuron images with low SNR. Therefore, it is necessary to improve the pseudo-labels for better segmentation. Huang et al.46 proposed a weak supervised learning method for neuron reconstruction. The region growing and skeleton method is used to mine more weak neurites from the probability map predicted by CNN iteratively. Here, we use the region growing method employed by Huang et al.46 to find more weak neurites from the probability map predicted by DDeep3M iteratively. The steps of the regional growth method are as follows:
Compared with the seed region, the grown region contained more voxels from neurites. Some weak neuron signals in the post-processing prediction graph are marked, which can be fused with the original image to obtain new enhanced training data for the next round of network training, to improve the accuracy of network prediction. 3.4.Image EnhancementThe output of DDeep3M network is the 3D probability map , which is calculated by the SoftMax activation function after the 3D data block passes through convolution layers of the network. denotes the probability that voxel becomes the part of neuron. To use the prediction results, a natural method is to segment neurons directly from the prediction graph. However, because the accuracy of pseudo labels is not as good as manual annotation, especially in early iterations, some local details may be lost in the prediction graph. Therefore, we enhance the original image by fusing the prediction image and the original image to maintain the accurate neuron structure and suppress the noise signal effectively. By fusing the prediction result graph with the original image signal, the enhanced image block can be obtained, to preserve the local signal and the global structure at the same time (the image enhancement formula is shown below). When the enhanced image block is delivered to the Hessian module, a more complete neural community can be segmented and a better pseudo-label can be provided for the next iteration of network training. In the process of iterative learning, DDeep3M and Hessian modules complement and promote each other, thus gradually improving the performance of neuron segmentation. Specifically, for an input image , where is a voxel, we identify the foreground voxels through the threshold () selected by experience, to screen the probability map.34 If the foreground probability of voxel is , we set the intensity to zero, otherwise we keep the original intensity value unchanged. We use a step-edge function to describe an intermediate image. if , ; if , . Then, intermediate images and probability maps are fused, and the final image adjusted by our enhancement function is defined as where is the maximum intensity of the image, and is the weight controlling the contribution of voxel in the original intensity, whereas is the weight controlling thecontribution of voxel in the prediction graph, which were both set between 0 and 1. Using more reliable segmentation results as pseudo-labels, the segmentation network can be further trained to learn more distinctive and representative features for generating the probability map, which is conducive to the Hessian matrix based enhancement filter to segment more complete neurons in the next iteration. To explore the influence of parameter for image enhancement, we adopt different values for . In this paper, we empirically select to increase the robustness of the whole framework.During each iteration of the training process, we calculate the value () of the test dataset. If the difference between and the value from the previous iteration () is , and the difference between and the value from the next iteration () is also (i.e., and ), we consider the iteration to have converged to the optimal solution. Based on our experimental results (as shown in Fig. 3) and experience, we have found that the number of iterations required to converge to the best solution is usually between 4 and 6. Once convergence is achieved, the network training is considered complete. 4.Results4.1.Datasets and SettingsWe evaluate the proposed method and state-of-the-art segmentation methods based on deep learning on various 3D optical neuron datasets, including the fMOST datasets and the BigNeuron datasets. For the fMOST dataset, we select an image stack with the voxel size of from the images of mouse neurons at the single fiber level for training and testing. The fMOST dataset used in the experiment is 3D volume data obtained from transgenic mouse Thy1 sample using two-photon fMOST (2P fMOST) system, which combines slicer and two-photon microscopy to realize continuous block imaging. In this study, the prediction set included the fMOST image stack of from the V layer of the mouse cerebral cortex. During the training phase, we input a representative fMOST image stack of size into the DDeep3M framework for iterative network training. The image stack was divided into three datasets for training, validation, and testing, respectively. The training dataset comprised a volume of , whereas the validation and testing datasets were and in size, respectively. For each iteration, we fed the training and validation datasets, along with the corresponding pseudo-labels, separately into the neural network for training. The network’s performance on the test dataset and the ground truth data were used to evaluate its prediction accuracy at each iteration, which was then used to decide whether to continue training the network or terminate the iteration. 4.2.Training DetailsThe fMOST image stack is processed by a histogram equalization technology to strengthen the contrast, and then normalized by merely dividing the value of all pixels by 255. Finally, the data augmentation technique is applied to the original training set. Data augmentation can expand the effective size of training data, thereby decreasing the problem of over fitting. It is normally used to train CNN for image classification. By applying several spatial transformations to the input image, we enhance the training and test data. These transformations are a combination of horizontal and vertical mirrors, rotation , , and 180 deg in the plane, and flipping in the direction (horizontal mirror). A total of 5 datasets of 1000 images are employed in training. After all the converted data is provided to the network, we apply reverse conversion for each probability map. We take the predicted average of each transformation graph and the original graph as the final output of the boundary probability map. We initialize the network by utilizing a Gaussian distribution with a standard deviation of 0.01. The mini batch size of our training model is 4, and the base learning rate is 0.01. The stochastic gradient descent optimization method with a momentum of 0.9 is adopted. Each model trains 30,000 iterations, and the mixed cross-entropy loss function is taken as our training objective. The network segmentation was trained and executed on a computer equipped with an Intel Xeon w-2123 CPU (64 GB RAM) and a Nvidia P5000 GPU (16 GB RAM). The prediction task for the fMOST image stack () was completed by DDeep3M+ in , resulting in an average processing rate of 18.4 s per image. 4.3.Evaluation MetricsTo conduct quantitative comparisons among different segmentation methods, we first perform a binarization operation on the segmentation results obtained by different segmentation methods with a threshold value of 200, and then utilize common metrics to quantitatively evaluate the segmentation result between the ground truth labeled by human experts. They are precision, recall, , and Jaccard, and their definitions are given by here TP, TN, FP, and FN denote the true positive, true negative, false positive, and false positive.To demonstrate the segmentation performance of our weakly supervised learning strategy, four widely-used deep segmentation networks, including U-Net, 3D DSN, VoxResNet, and DDeep3M, are tested to generate the neuron probability map in our framework. Eight iterations are tested on the fMOST dataset, and the value improvement of segmentation results is shown in Fig. 3. It can be seen that our DDeep3M+ algorithm achieves the best segmentation effect, with the value up to 0.973 after fourth iterations. After the fifth iteration, the value decreases, and the network segmentation effect does not improve when the iteration number increases. It indicates that the prediction performance of the weakly supervised learning framework has reached the best after the fourth iteration. In addition, the performance of the DDeep3M network is better than the other three deep networks. Moreover, the neuron segmentation results on a test block at different iterations are shown in Fig. 4, which depict the mining process of pseudo-label refinement. In Fig. 4, we can see that the prediction effect of DDeep3M network is more and more accurate, and many weak nerve fiber signals can be accurately segmented in the later stage, after several iterations of training. 4.4.Comparison with Other CNN ModelsBy comparing the detection performance of the proposed method against that of the common deep learning methods, we tested the effectiveness and accuracy of the proposed weakly supervised learning method for neuron detection and segmentation. The deep learning method used in this experiment is based on the supervised learning of the DDeep3M network. Figure 5 showed the segmentations of the proposed method on the fMOST datasets. Figure 6 displays the 3D rendering of the prediction results for a larger fMOST image stack (), as well as the maximum intensity projection of the original images and segmentations for 50 slices of the image stack. The proposed method detects almost all neurites from OM images. According to Table 1, the score of the proposed method and the supervised deep learning method are 0.973 and 0.980, respectively; the recall and precision scores of the proposed method are 0.952 and 0.994, respectively; and the recall and precision scores of the deep learning method are 0.970 and 0.985, respectively. In Table 1, we also implement segmentation models that contain different parts of our DDeep3M+ modules in the fMOST image trainings, and the results show the performance for the different key structures. Table 1 shows that our DDeep3M+, which contains our enhancement filter to obtain pseudo-labels, has the best performance for fMOST neuron segmentation. We implement four segmentation networks with a threshold method (named Otsu) for pseudo-labels. And the results show that their performances are far worse than our DDeep3M+ method. Table 1Quantitative comparison results of neuron segmentation by different methods on fMOST16 data.
Zhao et al.47 proposed a progressive learning method to reconstruct neurons via 3D DSN network from ultra-large scale optical microscopic images, and it is currently the excellent method for weakly supervised learning in the field of neuron segmentation. Figure 7 shows that neither U-Net segmentation network nor progressive learning method has an ideal segmentation effect on MOST dataset by comparing the segmentation performance of different network. We tested the proposed method and compared the final result with other three networks with a mouse brain dataset as shown in Fig. 8, which shows more details for comparison results between our method and other three segmentation networks. The result of proposed method shows more details of the fibers. As pointed out by the yellow box images in Fig. 8, the VoxResNet network was unable to detect neurites with weak or sudden changes in intensity. Similarly, it is difficult to trace some weak and uneven neurites from high noise images by 3D DSN network. In Fig. 8(f), there are some false positives in the result of U-Net network, which failed to trace some weak neurites precisely from the noisy background. Our method, employing weakly supervised learning, can accurately segment neuron cell body and fiber, and even some weak neurites can be accurately identified, and its segmentation performance is comparable to that of traditional deep learning method with manual annotations.42 4.5.Comparison on Other DatasetsWe also test the performance of the proposed method on other datasets, as shown in Table 2. The proposed method also performs well on the BigNeuron datasets, with the highest value up to 0.880, much higher than the 0.809 of U-Net network. The value of U-Net, 3D DSN, and VoxResNet were 0.809, 0.816, and 0.649, respectively, whereas the performance of DDeep3M in our experiment is 0.855 (recall), 0.907 (precision), and 0.880 ( value). After the comparison of various quantitative metrics, the proposed method is better than other segmentation methods in all aspects. Figure 9 shows the segmented neurons on three test images from the BigNeuron dataset. Table 2Quantitative comparison results of neuron segmentation by different methods on BigNeuron22 data. 4.6.Enhancement ParameterTo explore the influence of parameter in enhancement function on image enhancement of our segmentation framework, we use different values of and show the corresponding segmentation results of our method in Fig. 10. means that the original data block is directly used as the input of neuron segmentation. means that only the probability map is used as the input of neuron segmentation. This shows that the performance is improved by combining the probability map with the original image signals. The main reason is that the probability map reflects the long-range trajectory structure, whereas the original image signal carries more details of the neurites. The experimental results show that the segmentation performance can be improved by combining the probability map with the original image signal, which is mainly because the probability map reflects the long-range trajectory structures, whereas the original image signal has more details of subtle neural processes. Since we used pseudo labels to train the CNN model, its performance was limited when comparing it against the supervised learning method. To reduce the influence of false positive prediction on the probability map and combine with experimental verification, we chose in this paper to improve the robustness of the whole framework. 5.DiscussionDeep learning-based algorithms provide an effective and automatic method for neurite detection in highly noisy backgrounds. However, an accurate and robust estimation relies on large numbers of voxel-wise annotations. In recent years, the method of weakly supervised learning has gradually become a new trend in biomedical image segmentation. Huang et al.,46 Zhao et al.,47 and Chen et al.48 proposed three weakly supervised methods for neuron tracing and reconstruction. They solved the problem of the scarcity of manually annotated training samples and obtained better neurite detection results. Here, we employ a well-established CNN (DDeep3M) with inception and residual modules to achieve accurate, automatic, and universal neuron segmentation of 3D optical images for different types of neuron datasets with low and non-uniform signal strength. The proposed method is superior to several novel weakly supervised segmentation methods. First, we apply the Hessian-based adaptive enhancement filter for 3D neuron images based on DT to obtain more accurate pseud-label. Our enhancement filter can preserve the neuron signals of subtle intensity. The enhanced images have shown better visual effect, higher image quality, more abundant information. In addition, the proposed algorithm has shorter time consumption. Thus, this algorithm will benefit the morphological study of neurons. The adaptive enhancement filter based on Hessian matrix is used to extract the global feature of the neuron voxel, which can effectively provide the approximate location of the neuron. However, some weak, disconnected, and heterogeneous neurites are not detected in the pseudo-labels obtained through the Hessian matrix, which reduces the accuracy of image segmentation for low SNR neuron images. Therefore, finer neuronal signals must be mined for better segmentation. Deep CNN has achieved impressive performance in both natural and medical image segmentation, and its efficient performance in extracting image features gives us the possibility of our work. Further use of the DDeep3M network on the adaptively enhanced image can detect almost all the neurites in the neuron image, including the difficult to identify thin neurites with relatively low intensity. The initial training labels are constructed by the traditional image enhancement algorithm based on the neurite structure features. Then, the DDeep3M network improves prediction by iteratively optimizing the training labels and retraining model, while using regional growth to mine undetected weak neurites from the probability map predicted by DDeep3M. With more iterations of network training and image enhancement, the pseudo-labels are progressively refined and the DDeep3M model can gradually learn discriminative features of the neurites and background. Therefore, DDeep3M based segmentation method and traditional image enhancement technique can complement and promote each other. Finally, completer and more accurate neuron segmentation is obtained by our DDeep3M+ method. In the training phase, deep learning completes the model establishment and parameter optimization, and in the inference phase, it completes the specific implementation and application. After carefully adjusting the weights, the neural network is basically a bulky and huge database. To make full use of the training results and complete the segmentation task in practical application, we used data quantification, layer fusion, and model pruning to optimize the inference process of the network. After the acceleration technology in inference phase, the speed of network prediction can be increased by about 12 times. At the same time, the value of the prediction results did not decrease significantly, indicating that the inference speed was improved and the accuracy was not lost. There are still some limitations of our work. (1) Our method is currently not as good as supervised learning methods in accuracy, although we have an advantage in the cost of manual labeling. In the future work, we need to make better use of the prior knowledge of neuronal signals and improve the architecture of deep networks to improve the prediction accuracy of weakly supervised learning and achieve more accurate segmentation of neuronal protrusions. (2) The selection of the value of the parameter in the enhancement formula in the work of image fusion is empirical, and there is no theoretical support. We choose as the weighting value of the original image according to the experience of the experiment. The enhancement result is the best, which is only obtained from limited discrete data. (3) The selection of the number of iterations in the whole weakly supervised learning scheme is also obtained according to the experience of experiments, which also lacks the explanation of the algorithm principle. (4) We have not carried out neuron reconstruction experiments on the results of neural network segmentation. In the future work, we will use neuronal images enhanced by neural networks for neuron tracing and reconstruction, so as to analyze the advantages of our method in promoting neuron tracing, reconstruction and visualization compared with other deep learning methods. Compared to supervised learning methods, our method adapts to different types of optical neuron datasets without manual labeling, model redesign, or parameter tuning. The high accuracy results on challenging 3D optical images from different types of datasets demonstrate the accuracy and generalization of the proposed neurite segmentation method. The proposed weakly supervised learning framework can be extended to multi-modal biomedical big data, such as soma, vessels, and tumors, which helps life science experts to efficiently use deep learning techniques and promote further brain research. In addition, the source code of our method is open on the GitHub, which helps neuroscience researchers to easily use deep learning tools for their work on neuron segmentation and reconstruction. 6.ConclusionsWe propose a weakly supervised deep learning method (i.e., DDeep3M+) for automatic neuron segmentation in 3D OM images. Our DDeep3M+ framework consists of an adaptive enhancement filter that extracts initial neuron signals as initial neurite segmentations and a deep segmentation network for precise neurite segmentations from low SNR optical images without manual labeling. The accuracy of segmentation is close to that one of the best supervised learning methods. The comparison between the proposed method and several new segmentation methods proves the superiority of the proposed method in the segmentation of neurons from low SNR images. The proposed framework is effective for both public fMOST and BigNeuron datasets, and can be used for automatic segmentation of super-large neuronal population in the entire brain region. The codes and the pretrained model weights of our method are available at GitHub, which will help neuroscience researchers to easily use deep learning tools for their work on neuron segmentation and reconstruction. We believe it will promote further brain studies, including neuron tracing, neuron reconstruction, and so on. Code, Data, and Materials AvailabilityThe codes, datasets, and the pretrained model weights of our method are available at https://github.com/xiaoronghust/DDeep3Mplus/. AcknowledgmentsThis work was supported in part by the National Natural Science Foundation of China (Grant Nos. 61890951 and 61371014), and the Director Fund of WNLO, Open Project Program of Wuhan National Laboratory for Optoelectronics (Grant No. WNLOKF027). ReferencesA. Peters,
“Golgi, Cajal, and the fine structure of the nervous system,”
Brain Res. Rev., 55
(2), 256
–263 https://doi.org/10.1016/j.brainresrev.2006.12.002 BRERD2 0165-0173
(2003).
Google Scholar
S. L. Senft,
“A brief history of neuronal reconstruction,”
Neuroinformatics, 9
(2–3), 119
–128 https://doi.org/10.1007/s12021-011-9107-0 1539-2791
(2011).
Google Scholar
H. Zeng and J. R. Sanes,
“Neuronal cell-type classification: challenges, opportunities, and the path forward,”
Nat. Rev. Neurosci., 18
(9), 530
–546 https://doi.org/10.1038/nrn.2017.85 NRNAAN 1471-003X
(2017).
Google Scholar
F. Crick and E. Jones,
“Backwardness of human neuroanatomy,”
Nature, 361 109
–110 https://doi.org/10.1038/361109a0
(1993).
Google Scholar
J. P. Taylor, R. H. J. Brown and D. W. Cleveland,
“Decoding ALS: from genes to mechanism,”
Nature, 539
(7628), 197
–206 https://doi.org/10.1038/nature20413
(2016).
Google Scholar
J. M. Long and D. M. Holtzman,
“Alzheimer disease: an update on pathobiology and treatment strategies,”
Cell, 179
(2), 312
–339 https://doi.org/10.1016/j.cell.2019.09.001 CELLB5 0092-8674
(2019).
Google Scholar
J. Zhang, B. Long and A. Li,
“Whole-brain three-dimensional profiling reveals brain region specific axon vulnerability in 5xFAD mouse model,”
Front. Neuroanat., 14 608177 https://doi.org/10.3389/fnana.2020.608177
(2020).
Google Scholar
H. Gong, D. Xu and J. Yuan,
“High-throughput dual-colour precision imaging for brain-wide connectome with cytoarchitectonic landmarks at the cellular level,”
Nat. Commun., 7 12142 https://doi.org/10.1038/ncomms12142 NCAOBW 2041-1723
(2016).
Google Scholar
T. Ragan, L. R. Kadiri and K. U. Venkataraju,
“Serial two-photon tomography for automated ex vivo mouse brain imaging,”
Nat. Methods, 9
(3), 255
–258 https://doi.org/10.1038/nmeth.1854 1548-7091
(2012).
Google Scholar
L. Silvestri, A. Bria and L. Sacconi,
“Confocal light sheet microscopy: micron-scale neuroanatomy of the entire mouse brain,”
Opt. Express, 20
(18), 20582
–20598 https://doi.org/10.1364/OE.20.020582 OPEXFF 1094-4087
(2012).
Google Scholar
S. Liu, J. Nie and Y. Li,
“Three-dimensional, isotropic imaging of mouse brain using multi-view deconvolution light sheet microscopy,”
J. Innov. Opt. Heal. Sci., 10
(5), 1743006 https://doi.org/10.1142/S1793545817430064
(2017).
Google Scholar
P. Osten and T. W. Margrie,
“Mapping brain circuitry with a light microscopy,”
Nat. Methods, 10
(6), 515
–523 https://doi.org/10.1038/nmeth.2477 1548-7091
(2013).
Google Scholar
D. Cyranoski,
“China launches brain-imaging factory,”
Nat. News, 548
(7667), 268
–269 https://doi.org/10.1038/548268a
(2017).
Google Scholar
E. Meijering,
“Neuron tracing in perspective,”
Cytometry Part A, 77
(7), 693
–704 https://doi.org/10.1002/cyto.a.20895 1552-4922
(2010).
Google Scholar
A. Li, H. Gong and B. Zhang,
“Micro-optical sectioning tomography to obtain a high-resolution atlas of the mouse brain,”
Science, 330
(6009), 1404
–1408 https://doi.org/10.1126/science.1191776 SCIEAS 0036-8075
(2010).
Google Scholar
H. Gong, S. Zeng and C. Yan,
“Continuously tracing brain-wide longdistance axonal projections in mice at a one-micron voxel resolution,”
NeuroImage, 74 87
–98 https://doi.org/10.1016/j.neuroimage.2013.02.005 NEIMEF 1053-8119
(2013).
Google Scholar
M. N. Economon, N. G. Clack and L. D. Levis,
“A platform for brainwide imaging and reconstruction of individual neurons,”
Elife, 5 e10566 https://doi.org/10.7554/eLife.10566
(2016).
Google Scholar
M. Halavi, K. A. Hamilton and R. Parekh,
“Digital reconstructions of neuronal morphology: three decades of research trends,”
Front. Neurosci., 6 49 https://doi.org/10.3389/fnins.2012.00049
(2012).
Google Scholar
R. Parekh and G. Ascoli,
“Neuronal morphology goes digital: a research hub for cellular and system neuroscience,”
Neuron, 78
(1), 206 https://doi.org/10.1016/j.neuron.2013.03.015 NERNET 0896-6273
(2013).
Google Scholar
S. Li, T. Quan and H. Zhou,
“Identifying weak signals in inhomogeneous neuronal images for large-scale tracing of sparsely distributed neurites,”
Neuroinformatics, 17 497
–514 https://doi.org/10.1007/s12021-018-9414-9 1539-2791
(2019).
Google Scholar
Z. Zhou, H. C. Kuo and H. Peng,
“DeepNeuron: an open deep learning toolbox for neuron tracing,”
Brain Inf., 5
(2), 3 https://doi.org/10.1186/s40708-018-0081-2
(2018).
Google Scholar
H. Peng, M. Hawrylycz and J. Roskams,
“BigNeuron: large-scale 3D neuron reconstruction from optical microscopy images,”
Neuron, 87 252
–256 https://doi.org/10.1016/j.neuron.2015.06.036 NERNET 0896-6273
(2015).
Google Scholar
K. Svoboda,
“The past, present, and future of single neuron reconstruction,”
Neuroinformatics, 9
(2), 97
–98 https://doi.org/10.1007/s12021-011-9097-y 1539-2791
(2011).
Google Scholar
J. Lu,
“Neuronal tracing for connectomic studies,”
Neuroinformatics, 9
(2-3), 159
–166 https://doi.org/10.1007/s12021-011-9101-6 1539-2791
(2011).
Google Scholar
L. Feng, T. Zhao and J. Kim,
“NeuTube 1.0: a new design for efficient neuron reconstruction software based on the SWC format,”
eNeuro, 2
(1), ENEURO.0049-14.2014 https://doi.org/10.1523/ENEURO.0049-14.2014
(2015).
Google Scholar
H. Chen, H. Xiao and T. Liu,
“SmartTracing: self-learning-based neuron reconstruction,”
Brain Inf., 2
(3), 135
–144 https://doi.org/10.1007/s40708-015-0018-y
(2015).
Google Scholar
Y. LeCun, Y. Bengio and G. Hinton,
“Deep learning,”
Nature, 521
(7553), 436
–444 https://doi.org/10.1038/nature14539
(2011).
Google Scholar
A. Krizhevsky, L. Sutskever and G. Hinton,
“ImageNet classification with deep convolutional neural networks,”
in Proc. 25th Int. Conf. Neural Inf. Process. Syst.,
1097
–1105
(2012). Google Scholar
J. Long, E. Shelhamer and T. Darrell,
“Fully convolutional networks for semantic segmentation,”
IEEE Trans. Pattern Anal. Mach. Intell., 39
(4), 640
–651
(2015).
Google Scholar
O. Ronneberger, P. Fischer and T. Brox,
“U-Net: convolutional networks for biomedical image segmentation,”
Lect. Notes Comput. Sci., 9351 234
–241 https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743
(2015).
Google Scholar
Q. Dou et al.,
“3D deeply supervised network for automated segmentation of volumetric medical images,”
Med. Image Anal., 41 40
–54 https://doi.org/10.1016/j.media.2017.05.001
(2017).
Google Scholar
H. Chen et al.,
“VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images,”
NeuroImage, 170 446
–455 https://doi.org/10.1016/j.neuroimage.2017.04.041 NEIMEF 1053-8119
(2018).
Google Scholar
X. Wu, Y. Tao and G. He,
“Boosting multilabel semantic segmentation for somata and vessels in mouse brain,”
Front. Neurosci., 15 610122 https://doi.org/10.3389/fnins.2021.610122 1662-453X
(2021).
Google Scholar
R. Li, T. Zeng and S. Ji,
“Deep learning segmentation of optical microscopy images improves 3D neuron reconstruction,”
IEEE Trans. Med. Imaging, 36
(7), 1533
–1541 https://doi.org/10.1109/TMI.2017.2679713 ITMID4 0278-0062
(2017).
Google Scholar
B. Yang, W. Chen and Y. Wang,
“Neuron image segmentation via learning deep features and enhancing weak neuronal structures,”
IEEE J. Biomed. Health Inf., 25
(5), 1634
–1645 https://doi.org/10.1109/JBHI.2020.3017540
(2021).
Google Scholar
B. Yang, M. Liu and E. Meijering,
“Structure-guided segmentation for 3D neuron reconstruction,”
IEEE Trans. Med. Imaging, 41
(4), 903
–914 https://doi.org/10.1109/TMI.2021.3125777 ITMID4 0278-0062
(2022).
Google Scholar
M. Dong, D. Liu and F. Wu,
“3D CNN-based soma segmentation from brain images at single-neuron resolution,”
in 25th IEEE Int. Conf. Image Process. (ICIP),
126
–130
(2018). https://doi.org/10.1109/ICIP.2018.8451389 Google Scholar
Q. Li and L. Shen,
“3D neuron reconstruction in tangled neuronal image with deep networks,”
IEEE Trans. Med. Imaging, 39
(2), 425
–435 https://doi.org/10.1109/TMI.2019.2926568 ITMID4 0278-0062
(2020).
Google Scholar
T. Jerman and B. Likar,
“Enhancement of vascular structures in 3D and 2D angiographic images,”
IEEE Trans. Med. Imaging, 35
(9), 2107
–2118 https://doi.org/10.1109/TMI.2016.2550102 ITMID4 0278-0062
(2016).
Google Scholar
R. F. Frangi, W. J. Niessen and K. L. Vincken,
“Multiscale vessel enhancement filtering,”
Lect. Notes Comput. Sci., 1496 130
–137 https://doi.org/10.1007/BFb0056195 LNCSD9 0302-9743
(1998).
Google Scholar
Z. Zhou, S. Sorensen and H. Peng,
“Adaptive image enhancement for tracing 3D morphologies of neurons and brain vasculatures,”
Neuroinformatics, 13
(2), 153
–166 https://doi.org/10.1007/s12021-014-9249-y 1539-2791
(2015).
Google Scholar
X. Wu, S. Chen and J. Huang,
“DDeep3M: Docker-powered deep learning for biomedical image segmentation,”
J. Neurosci. Methods, 342 108804 https://doi.org/10.1016/j.jneumeth.2020.108804 JNMEDT 0165-0270
(2020).
Google Scholar
M. G. Haberl, C. Churas and L. Tindall,
“CDeep3M–Plug-and-Play cloud-based deep learning for image segmentation,”
Nat. Methods, 15 677
–680 https://doi.org/10.1038/s41592-018-0106-z 1548-7091
(2018).
Google Scholar
Q. Huang, J. Sun and H. Ding,
“Robust liver vessel extraction using 3D U-Net with variant dice loss function,”
Comput. Biol. Med., 101 153
–162 https://doi.org/10.1016/j.compbiomed.2018.08.018 CBMDAW 0010-4825
(2018).
Google Scholar
T. Falk, D. Mai and R. Bensch,
“U-Net: deep learning for cell counting, detection, and morphometry,”
Nat. Methods, 16 67
–70 https://doi.org/10.1038/s41592-018-0261-2 1548-7091
(2019).
Google Scholar
Q. Huang, Y. Chen and S. Liu,
“Weakly supervised learning of 3D deep network for neuron reconstruction,”
Front. Neuroanat., 14 38 https://doi.org/10.3389/fnana.2020.00038
(2020).
Google Scholar
J. Zhao, X. Chen and Z. Xiong,
“Neuronal population reconstruction from ultra-scale optical microscopy images via progressive learning,”
IEEE Trans. Med. Imaging, 39
(12), 4034
–4046 https://doi.org/10.1109/TMI.2020.3009148 ITMID4 0278-0062
(2020).
Google Scholar
X. Chen, C. Zhang and J. Zhao,
“Weakly supervised neuron reconstruction from optical microscopy images with morphological priors,”
IEEE Trans. Med. Imaging, 40
(11), 3205
–3216 https://doi.org/10.1109/TMI.2021.3080695 ITMID4 0278-0062
(2020).
Google Scholar
|
Image segmentation
Neurons
Education and training
Machine learning
Voxels
Image enhancement
3D image processing