Significance: Confocal endoscopy images often suffer distortions, resulting in image quality degradation and information loss, increasing the difficulty of diagnosis and even leading to misdiagnosis. It is important to assess image quality and filter images with low diagnostic value before diagnosis. Aim: We propose a no-reference image quality assessment (IQA) method for confocal endoscopy images based on Weber’s law and local descriptors. The proposed method can detect the severity of image degradation by capturing the perceptual structure of an image. Approach: We created a new dataset of 642 confocal endoscopy images to validate the performance of the proposed method. We then conducted extensive experiments to compare the accuracy and speed of the proposed method with other state-of-the-art IQA methods. Results: Experimental results demonstrate that the proposed method achieved an SROCC of 0.85 and outperformed other IQA methods. Conclusions: Given its high consistency in subjective quality assessment, the proposed method can screen high-quality images in practical applications and contribute to diagnosis. |
1.IntroductionConfocal endoscopy employs laser scanning confocal imaging technology to achieve real-time observation of mucosal cells and subcellular structures with micron-scale resolution to accurately locate lesions.1,2 Probe-based confocal endoscopes are commonly used. They transmit laser and fluorescence images via a fiber bundle,3 which has flexibility and accessibility for in vivo clinical imaging. The miniature objective system is the core component of confocal endoscopy for high resolution to directly visualize cells and is often assembled from multiple optical elements to correct aberrations.4 It improves biopsy accuracy and contributes to the early diagnosis of cancer in various clinical fields such as human brain tumors5 and gastrointestinal cancer.6 However, distortions, such as blur, noise, and decreased contrast, are common in confocal endoscopy imaging. Blur is the most common type of distortion in confocal endoscopy, which is caused by defocus, probe fiber core cross coupling,7 and motion caused by the difference in movement speed between the investigated anatomical structures and the physician.8 Owing to the small field-of-view of the confocal endoscopy,9 to obtain a comprehensive view, a typical endoscopy examination produces thousands of images, most of which are not useful for diagnostic purposes5 owing to image degradation and information loss caused by the above distortions. Manual removal of nondiagnostic and low-quality images is time-consuming and labor-intensive. Therefore, it is desirable to automatically screen high-quality images accurately and efficiently. An image enhancement method is observed to be beneficial in the automatic diagnosis of confocal endoscopy images,10 and its development and evaluation also require the involvement of image quality assessment (IQA). Furthermore, the imaging performance evaluation of the confocal endoscopy also requires the participation of IQA; for example, Wang et al. analyzed the image histogram distribution to assess image contrast to validate the proposed confocal microendoscope.11 Therefore, it is essential to develop an IQA method because it can benefit clinical applications of confocal endoscopy. IQA is mainly divided into full reference (FR) and no reference (NR) methods. FR-IQA requires an ideal undistorted image when evaluating the quality, which is difficult to obtain in practical applications. NR-IQA is gaining attention because it does not require the use of reference images. The feature extraction and prediction model form the general NR-IQA framework. Common features describe poetry of images, including natural scene statistics (NSS) feature,12,13 gradient feature,14 frequency domain feature,15 curvelet domain feature,16 and discrete cosine transform (DCT) domain feature.17 In addition, there are methods based on analysis of the perceptual process of image reception by the human eye, that is the human vision system (HVS), such as free energy theory18 and phase congruency.19 Owing to the powerful ability of image description, local descriptors in IQA have aroused extensive attention, such as local binary pattern (LBP),20,21 from accelerated segment test (FAST),22 speeded-up robust features (SURF),23 Weber local descriptor (WLD),24 and they have remarkable performance in multiply-distorted images. Currently, NR-IQA has promising performance for images with a single distortion, such as blur, noise, and JPEG compression, while it is unsatisfactory when it comes to authentic and multiply distorted images25 owing to joint distortion interactions. Deep learning techniques for IQA have been studied; however, the size of the dataset limits the network structure.26 Bianco et al.27 proposed the DeepBIQ method that uses a fine-tuned convolutional neural network (CNN) to exact features and feeds it to support vector regression (SVR) to predict the image quality. Liu et al.28 proposed RankIQA, first pretrained the network on a large-scale self-build image pair database for an quality comparison of image pair task and fine-tuned the network to achieve promising performance. Ma et al.29 proposed MEON, which first pretrained the network by an image distortion classification task on synthetic distortion images and then fine-tuned the network to achieve end-to-end image quality prediction. Zhu et al.30 proposed MetaIQA method, first adopted the meta-learning strategy to learn the prior knowledge from different NR-IQA tasks and fine-tuned the network to address the small sample problem. In general, deep learning requires a great number of training images, and the scale of the image quality database severely limits the performance of deep networks. Establishing larger datasets and designing a new method to reduce the requirement for training images is a direction that needs further exploration for deep IQA. Medical images are multiply distorted because of variable imaging conditions; meanwhile, the content of medical images is distinct from that of natural images, mainstreaming IQA may decline in performance.31 It is necessary to develop IQA of medical images. The study of medical IQA has increased because of the development of deep learning techniques, and it mainly focuses on ultrasound imaging, MRI, and OCT images. Zhang et al.32 proposed DCNN-IQA-14 and ResNet-IQA based on traditional networks to predict the quality of ultrasound images. To overcome overfitting, a transfer-learning strategy was employed. Liu et al.33 proposed a nonlocal residual neural network to assess slicewise MRI image quality and applied random forest for the volumewise quality grade. Semisupervised learning and iterative self-training strategies were used for a few quality-annotated images. Wang et al.34 analyzed the performance of four classic deep networks for assessing the quality of retinal OCT images by transfer learning, and the ResNet-50 network achieves highest performance. Medical images are more difficult to acquire and label than natural images, making the deep learning IQA method difficult to develop, transfer learning strategies based on traditional networks are widely adopted, and the potential of deep learning has yet to be fully explored. The analysis of confocal image quality has made some progress. Aubreville et al.8 proposed an improved Inception v3 network to detect motion artifacts in confocal endoscopy images. Kamen et al.35 screened high quality and information-rich images by calculating image entropy before classifying images. Izadyyazdanabadi et al.36 proposed a binary classification network to classify diagnostic and nondiagnostic images and applied fine-tuning and ensemble modeling techniques to improve performance and achieve high accuracy.37 Despite this, none of these methods can quantify image quality, which limits the applications for screening and evaluating image enhancement algorithms. The signal-to-noise ratio is determined by a confocal endoscopy IQA38 that can quantify image quality. However, regions of interest need to be manually selected before calculation, which does not apply to practical applications. To meet the needs of practical applications, in this paper, we propose a new NR confocal laser endoscopy IQA (CEIQA) method, which combines local descriptors and Weber’s law. First, we used a differential excitation (DE) map to describe the local variation information, calculated the LBP map of the image to describe structure information, and then computed the joint distribution histogram of DE and LBP as the first feature set. Second, for better perception of the ability to describe image information, we improved local ternary pattern (LTP) by changing its threshold function by referring to Weber’s law and computed the histogram and entropy of improved LTP, which was used to measure the distribution of local variation patterns. Finally, SVR was applied to map the perceptual features to the quality score. Our main contributions are summarized as follows:
The remainder of this paper proceeds as follows. In Sec. 2, we present the feature-extraction method for the proposed IQA. In Sec. 3, we present the details and results of the performance experiments. In Sec. 4, we discuss the limitations of the study and future work. Finally, in Sec. 5, we conclude the main contributions and provide potential application prospects for the proposed IQA. 2.Materials and Methods2.1.Weber’s Law and Differential ExcitationThe perceptual process is sensitive to relative variations in pixel intensity in images when recording to the HVS. Weber’s law can be used to describe this mechanism,39 which is expressed as follows: where represents the perceptual threshold, represents the initial stimulus intensity, and is a constant. Based on Weber’s law, Chen et al.40 proposed the WLD to extract local variation information in the image. One of the components is DE, which is calculated as follows: where denotes the original image, characterizes the image local variation, is the DE map, is the central pixel, is the neighborhood pixels, is the number of neighborhood pixels, and arctan function is used to prevent computation instability. After the above calculation, the range of becomes . Compared to Weber’s law, DE regards the sum of the difference between the neighbors and the center as the change in the image. In this study, we adopted DE to describe the local variation in confocal endoscopy images. Figures 1(a) and 1(d) show two confocal endoscopy images with different MOSs and DE maps, respectively.Figures 1(b) and 1(e) show that DE highlights the variation region in the image; thus, the distribution of the DE values shown in Figs. 1(c) and 1(f) represents the distribution of levels of local variation in the image. A low variation region means “the expected” according to the free energy theory,18 which carries less information than a high-variation region. Figures 1(c) and 1(f) demonstrate that the DE value of the low-quality image is more often located around the zero point, whereas that of the high-quality image is more evenly distributed. In conclusion, the distributional characteristics of DE can be used as perceptual features that indicate image quality. Nonetheless, Eq. (2) shows that during the accumulation phase, positive and negative variations will counteract each other, resulting in image variation information loss; meanwhile, the pattern and direction of local variation are ignored. Therefore, further analysis based on a DE map is required. 2.2.Improved LBP by Differential ExcitationThe LBP41 is a local descriptor with remarkable performance and has attracted considerable attention in IQA studies20,21 owing to its ability to describe the local structure of an image. By comparing the interpixel relationships between the central pixel and its neighbors, LBP divides pixels into different patterns. The general form of LBP is rotation-invariant uniform ,41 defined as follows: where and are the central and circular neighborhood pixels, respectively, is the radius of circular neighborhood, is the number of , and is the thresholding function defined as follows: where defined in Eq. (5) is a bitwise transition function that recognizes a uniform pattern whose number of bitwise transitions is less than two. Uniform patterns contribute more to the description of the image structure than basic patterns. has uniform patterns and one other pattern, coding from 0 to :The image structure contains information, and quality degradation will affect it, resulting in pattern shifts in the LBP.20 Therefore, LBP can be used in the representation of image quality. The obtained LBP map is defined as follows: DE only calculates the intensity information and ignores local variation information of pattern and direction; meanwhile, LBP contains interpixel relationships without pixel intensity information. Therefore, it is helpful to calculate the joint distribution histogram of DE and LBP to compensate for both the shortage in describing the structure and preserving the local variation intensity and pattern of pixels. The joint distribution histogram was calculated as follows: where is a two-dimension joint histogram, indicates the frequency function. , where is the number of uniform patterns of and , where is the number of bins in the histogram of the . After the joint histogram is obtained, it is an dimension feature to characterize the local variation and structure of the image.2.3.Improved LTP by Weber’s LawThe LTP42 is the generalization of LBP by changing the threshold function to obtain more detailed information of the local interpixel relationship. LTP was calculated as follows: The obtained LTP map contains patterns when because LTP codes can be , 0, and 1. For lower computational complexity, the original LTP map is converted to an up-pattern and low-pattern map. The up-pattern map is obtained by turning the LTP code C from to 0 in Eq. (10), and a low-pattern map is obtained by turning the LTP code C from 1 to 0 and to 1. Therefore, the up-pattern and low-pattern maps both contain patterns with values ranging from 0 to 255 as LBP. Figure 2 shows the LTP map of the two confocal endoscopy images with different MOSs. Note that because of the threshold in the threshold function, LTP emphasizes the high variation region and underestimates the low variation region, which is similar to DE, while the choice of threshold value affects the screening strength of local variations. However, Fig. 2 shows that the LTP map retains an image structure like LBP; thus, LTP can describe image quality degradation. Considering that LTP captures the local variation before extracting the structure, we apply Weber’s law to the threshold function of LTP as follows: where consults the form of Weber’s law regarding as , as , and as threshold . To prevent the instability of the results, the overall image is added by one. After introducing Weber’s law in the threshold function, the judgment of local variation is under the HVS.Threshold in the threshold function of LTP determines the ability to capture an image variation and further affects the image description. Therefore, the threshold plays a key role in the performance of WB-LTP. Referring the form of Weber’s law, the threshold is calculated as follows: where is the DE map in Eq. (2), denotes the average variation intensity of image . Function tan() and factor 1/256 are used to transfer the range of from to the range of for corresponding to Eq. (11). The threshold indicates that the region above or below average variation intensity of the image will be regarded as the high variation or low variation region, respectively.After LTP improvement, we conducted further analysis to enrich the information expression of images. Considering that the up and low channels denote different directions of patterns, we referred to the calculation of the gradient magnitude14 and computed the magnitude channel as follows: Because LTP reveals local structure distribution, it is useful to calculate the entropy of LTP to characterize image information as follows: where denotes the frequency of grayscale value in the image. We calculated , , and . Finally, we obtained the WB-LTP feature , where , , and are the histograms of , , and , respectively.2.4.Feature Extraction and Quality ModelBecause the field-of-view of confocal endoscopy is circular, acquired images appear as circular effective regions surrounded by black regions, and the latter interferes with the image description. Therefore, before feature extraction, the image should be preprocessed by adopting the square inscribed in the valid circle region, as shown in Fig. 3. After preprocessing, the DE-LBP of the image was calculated using in DE, , and in LBP and to obtain 100-dimensional features. Then, computing the WB-LTP of the image with 15 bins in the histogram and acquire 48 dimension features. To obtain multiscale information of the image, the image is downsampled twice beside the origin scale. Features are extracted in three scales; thus, the features have 444 dimensions. In WB-LTP, the structure information differs from the three scales of images, leading to different thresholds. Therefore, the thresholds of different scales was calculated as , where is the scale factor, is the number of image downsampling. After obtaining the image features, SVR43 with a radial basis function kernel was adopted to build a quality prediction model. 2.5.Confocal Endoscopy Image DatabaseTo compare the performance of IQA methods, we established a database of confocal endoscopy images. The imaging experiment was conducted using a confocal endoscope designed by Wang et al.1 The confocal endoscope has a field of view of and a resolution of , and the image was obtained with at a frame rate of 4 to 16 fps. Imaging experiments were conducted with ex vivo imaging of colonic and gastric tissues of female specific pathogen free and Sprague Dawley rats weighing . Imaging experiments obtained 656 images with blur, contrast distortion, and motion artifacts. All imaging experiments were approved by the animal experiment guidelines of the Animal Experimentation Ethics Committee of Huazhong University of Science and Technology (HUST, Wuhan, China). To obtain meaningful MOS, eight researchers with extensive experience in instrument operation and image processing of confocal endoscopy rated the quality of images. Subjective quality assessment experiments apply single-stimulus (SS) methods. Every observer watches one confocal image 10 s on a computer monitor every time and gives a quality index ranging from one to five, where one denotes the lowest quality and five denotes the highest quality. To avoid observer exhaustion, a session lasted for half an hour and the observers watched 180 images. After a session, the observers rested for 8 min. Thus, there were four sessions in subjective quality assessment experiments. After obtaining the eight rates of every image, the standard deviation (STD) of every image quality score was calculated. The image with STD higher than 1.5 was discarded because the result cannot reflect objective image quality. Eight images were discarded in this process, and 642 images remained. Finally, the MOS of the images was computed by averaging the scores of the researchers, and the distribution of the image quality scores is shown in Fig. 4(a). The relationship between STD and MOS is shown in Fig. 4(b), which we can see that the consistency of observers’ opinions is higher when faced with relatively high-quality and low-quality images compared with images of medium quality. 3.Results3.1.Experimental ProtocolWe compared the proposed method with 11 state-of-the-art NR-IQAs with publicly available source codes. The methods used for comparison were the NSS feature-based IQA methods including BRISQUE12 and SINQ44 in the spatial domain, NBIQA13 in the spatial and DCT domains, and CurveletQA16 in the curvelet domain. There are local descriptor-based IQA methods including GWH-GLBP20 using a gradient magnitude-weighted LBP feature, ORACLE45 using the FAST algorithm, and the fast RetinaKeypoint descriptor (FREAK), NOREQI23 using the SURF algorithm, and RATER46 using FAST. In addition, there are image spatial and spectral entropy feature-based method SSEQ,15 free-energy feature-based NFERM,18 and gradient magnitude feature-based GM-LOG.14 All the above methods follow the process of feature extraction, SVR training, and prediction. Before feature extraction, all IQA models apply the same preprocessing method. For IQA employing color space information, grayscale values were used as inputs to the three color channels. In the performance experiment, a confocal endoscopy image dataset was applied. First, 80% of the dataset were randomly chosen to train the SVR model, and the rest were used to test the performance. During the training phase of all IQA methods, SVR parameters were optimized using a grid search to achieve the best performance for fair comparison. In the performance evaluation, Spearman rank order correlation coefficient (SROCC), Pearson’s linear correlation coefficient (PLCC), and root mean square error (RMSE) were used to characterize the monotonicity and accuracy of prediction. SROCC and PLCC values closer to 1 and RMSE closer to 0 indicates better performance. Before the PLCC and RMSE are calculated, the nonlinear logistic regression shown in Eq. (15) is required,47 where is the predicted score, is the fitting score, and is the regression parameter: The random 80% to 20% train-test is repeated 1000 times, and the median of performance criteria are reported. For a fair comparison, all methods use the same training and testing sets in each repeat.3.2.Performance ComparisonTable 1 shows the performance of NR-IQA, and the best method is shown in bold. As shown in Table 1, CEIQA outperforms the other IQAs in all criteria, followed by NOREQI. Table 1Performance comparison of NR-IQA methods. The best IQA methods are highlighted in boldface.
To further verify whether the performance difference is significant, we conducted the corrected resampled paired Student’s -test48 between different methods of SROCC values of 1000 train-test repeats. The results are shown in Fig. 5, where symbols “1,” “,” and “0” mean that the method in the row is statistically (with 95% confidence) better, worse, or similar to the method in the column, respectively. Figure 5 shows that CEIQA is significantly superior to all other NR-IQA methods on the confocal endoscopy image dataset, followed by NOREQI. To further analyze the relationship between the algorithm’s performance and the consistency between the predicted quality scores and MOS, the scatter plots of MOS against the predicted quality scores of the test dataset from the IQA methods in a train-test repeat are shown in Fig. 6. The corresponding SROCC values were labeled in the plots. For a clear view, we only show the results of BRISQUE, SSEQ, CurvletQA, NOREQI, GWHGLBP, ORACLE, GMLOG, and the proposed CEIQA method. As shown in Fig. 6, the CEIQA with the highest SROCC shows the promising performance with most densely and evenly distribution of points around the fitting curve, which indicates the great consistency between MOS and the predicted quality scores. From Fig. 6, we can also see that the higher SROCC indicates better consistency between MOS and the quality scores predicted by the method. The robustness of the algorithm is an important factor in the performance. To evaluate the robustness of IQA, we calculated the STD of the three criteria across 1000 repeats, which are shown in Table 1. The lower the STD, the more stable the algorithm performs for varying images. According to Table 1, CEIQA is the most stable among the IQA methods. The rightmost column of Table 1 shows the time taken by different algorithms to extract the confocal endoscopy image features, which accounts for most of the total algorithm runtime. CEIQA has a promising speed and runs faster than NOREQI with the second-best performance. Experiments were performed on MATLAB R2017a with an Intel i7-8750HQ CPU at 2.20 GHz. The experimental results demonstrated that CEIQA with a combination of perceptual laws and local descriptors has promising performance. Furthermore, to verify the enhancement of the proposed method using perceptual laws and the performance of the IQA components, we compared the performance of LBP and LTP before and after being improved using perceptual laws using the same 1000 train-test procedure. The results are presented in Table 1. According to the results, the performance of LBP and LTP is remarkable, which demonstrates that the local descriptor is suitable for the quality prediction of confocal endoscopy images with multiple distortions. The same conclusion can be drawn from the fact that the NOREQI with the second-best performance also uses the local descriptors SURF, as shown in Table 1. Note that LTP uses the threshold calculation method proposed by Freitas et al.49 Furthermore, DE-LBP, which combines LBP and DE, significantly outperforms origin LBP, while WB-LTP also performs better than origin LTP owing to Weber’s law engagement. Introducing perceptual laws enhances the capability to describe image information and assess image quality. 3.3.Analysis of ParametersThere are two types of parameters in the proposed CEIQA method. The first is the neighborhood radius and the number of neighborhood pixels in DE-LBP. The second is the number of bins in the WB-LTP histogram. To verify if the performance of CEIQA is sensitive to the variations of parameters, we performed two experiments with a variety of parameters. First, we compared the performance of DE-LBP for different values of and . The 1000 train-test process, as in Sec. 3.1, is conducted, and the median of the SROCC of the entire loop is shown in Fig. 7(a). Then, the performance experiment of WB-LTP for different numbers of bins was conducted, and the result is shown in Fig. 7(b). As shown in Fig. 7, DE-LBP’s performance is stable across different settings of LBP and WB-LTP’s performance is stable in different bins number of the histogram in a moderately large range. In conclusion, the CEIQA, that consists of DE-LBP and WB-LTP, is robust to parameter variations and has good generalization capability. 4.DiscussionIt is meaningful to analyze the relationship between the algorithm’s performance in predicting MOS and screening images. In clinical practice, image screening was employed by setting the quality threshold first, and then the images with a quality score lower than the quality threshold are labeled as “the low-quality image” which should be screened out and the images with a quality score higher than or equal to the quality threshold are labeled as “the high-quality image” that need to be kept. Therefore, the high consistency of the predicted quality scores and MOS indicates the high accuracy of image filtering. As shown in Table 1 and Fig 6, the proposed CEIQA method with great consistency between the predicted quality scores and MOS has the potential for practical application. This study has some limitations. First, the confocal endoscopy images were obtained from a single confocal endoscopy instrument and two tissue types. The type of distortion is limited. This also limits the ability of the proposed CEIQA method to effectively characterize image quality. The lack of image morphology can cause the overfitting of the IQA. For example, IQA will provide higher quality scores to images with the tissue of high local variability and ignore distortion during imaging. The algorithm also must be improved to ignore image content and focus on image distortion. Second, the objectivity of subjective IQA of image MOS must be further improved, which in turn will affect the performance evaluation and application potential of the algorithm. Future work will focus on obtaining confocal endoscopy images of more tissues, imaging conditions, and more detailed and clear types of distortion, as well as conducting more comprehensive subjective IQA experiments to obtain more meaningful MOS. With additional data, CEIQA is expected to show better performance and can be universally applied. Other directions of research could involve using the IQA method to evaluate and improve confocal endoscopy image enhancement and deconvolution algorithms or using the IQA method to select high-quality images in clinical practice and analyzing the effectiveness of the method. 5.ConclusionIn this study, we proposed a new NR-IQA method named CEIQA based on Weber’s law and a local descriptor. The image structural information is measured using the local descriptor, which is then improved by Weber’s law to extract perceptual features. The method is compared with 11 state-of-the-art NR-IQA methods on the introduced dataset of confocal endoscopy images. The dataset contains 642 images with authentic distortion and the corresponding MOS assessed by eight experimenters. As shown in the experimental results, CEIQA is significantly superior to other NR methods in terms of accuracy and robustness, which demonstrates that CEIQA has great potential for practical application and contributes to clinical diagnosis. AcknowledgmentsThis work was supported by the National Natural Science Foundation of China (Grant No. 81971692). Code, Data, and Materials AvailabilityThe confocal endoscopy image dataset can be obtained from the authors upon reasonable request. The CEIQA source code and demo code is available in the Figshare repository via the link https://figshare.com/articles/software/CEIQAcode/16987909. ReferencesJ. Wang et al.,
“A confocal endoscope for cellular imaging,”
Engineering, 1
(3), 351
–360
(2015). https://doi.org/10.15302/J-ENG-2015081 ENGNA2 0013-7782 Google Scholar
H. Li et al.,
“Advanced endoscopic methods in gastrointestinal diseases: a systematic review,”
Quantum. Imaging Med. Surg., 9
(5), 905
–920
(2019). https://doi.org/10.21037/qims.2019.05.16 Google Scholar
Y. Wang et al.,
“Four-plate piezoelectric actuator driving a large-diameter special optical fiber for nonlinear optical microendoscopy,”
Opt. Express, 24
(17), 19949
(2016). https://doi.org/10.1364/OE.24.019949 OPEXFF 1094-4087 Google Scholar
L. Yang et al.,
“Five-lens, easy-to-implement miniature objective for a fluorescence confocal microendoscope,”
Opt. Express, 24
(1), 473
(2016). https://doi.org/10.1364/OE.24.000473 OPEXFF 1094-4087 Google Scholar
N. L. Martirosyan et al.,
“Prospective evaluation of the utility of intraoperative confocal laser endomicroscopy in patients with brain neoplasms using fluorescein sodium: experience with 74 cases,”
Neurosurg. Focus, 40
(3), E11
(2016). https://doi.org/10.3171/2016.1.FOCUS15559 Google Scholar
Z. Li et al.,
“New classification of gastric pit patterns and vessel architecture using probe-based confocal laser endomicroscopy,”
J. Clin. Gastroenterol., 50
(1), 23
–32
(2016). https://doi.org/10.1097/MCG.0000000000000298 Google Scholar
A. K. Eldaly et al.,
“Deconvolution and restoration of optical endomicroscopy images,”
IEEE Trans. Comput. Imaging, 4
(2), 194
–205
(2018). https://doi.org/10.1109/TCI.2018.2811939 Google Scholar
M. Aubreville et al.,
“Deep learning-based detection of motion artifacts in probe-based confocal laser endomicroscopy images,”
Int. J. Comput. Assist. Radiol. Surg., 14
(1), 31
–42
(2019). https://doi.org/10.1007/s11548-018-1836-1 Google Scholar
H. Li et al.,
“ field-of-view probe-based confocal microendoscope for large-area visualization in the gastrointestinal tract,”
Photonucs Res., 9
(9), 1829
(2021). https://doi.org/10.1364/PRJ.431767 Google Scholar
N. Ghatwary et al.,
“Automatic grade classification of Barretts esophagus through feature enhancement,”
Proc. SPIE, 10134 1013433
(2017). https://doi.org/10.1117/12.2250364 PSISDG 0277-786X Google Scholar
J. Wang et al.,
“Near-infrared probe-based confocal microendoscope for deep-tissue imaging,”
Biomed. Opt. Express, 9
(10), 5011
(2018). https://doi.org/10.1364/BOE.9.005011 BOEICL 2156-7085 Google Scholar
A. Mittal, A. K. Moorthy and A. C. Bovik,
“No-reference image quality assessment in the spatial domain,”
IEEE Trans. Image Process., 21
(12), 4695
–4708
(2012). https://doi.org/10.1109/TIP.2012.2214050 IIPRE4 1057-7149 Google Scholar
F.-Z. Ou, Y.-G. Wang and G. Zhu,
“A novel blind image quality assessment method based on refined natural scene statistics,”
in IEEE Int. Conf. Image Process.,
1004
–1008
(2019). https://doi.org/10.1109/ICIP.2019.8803047 Google Scholar
W. Xue et al.,
“Blind image quality assessment using joint statistics of gradient magnitude and Laplacian features,”
IEEE Trans. Image Process., 23
(11), 4850
–4862
(2014). https://doi.org/10.1109/TIP.2014.2355716 IIPRE4 1057-7149 Google Scholar
L. Liu et al.,
“No-reference image quality assessment based on spatial and spectral entropies,”
Signal Process. Image Commun., 29
(8), 856
–863
(2014). https://doi.org/10.1016/j.image.2014.06.006 SPICEF 0923-5965 Google Scholar
L. Liu et al.,
“No-reference image quality assessment in curvelet domain,”
Signal Process. Image Commun., 29
(4), 494
–505
(2014). https://doi.org/10.1016/j.image.2014.02.004 SPICEF 0923-5965 Google Scholar
M. A. Saad, A. C. Bovik and C. Charrier,
“DCT statistics model-based blind image quality assessment,”
in 18th IEEE Int. Conf. Image Process.,
3093
–3096
(2011). https://doi.org/10.1109/ICIP.2011.6116319 Google Scholar
K. Gu et al.,
“Using free energy principle for blind image quality assessment,”
IEEE Trans. Multimedia, 17
(1), 50
–63
(2015). https://doi.org/10.1109/TMM.2014.2373812 Google Scholar
C. Li, A. C. Bovik and X. Wu,
“Blind image quality assessment using a general regression neural network,”
IEEE Trans. Neural Network, 22
(5), 793
–799
(2011). https://doi.org/10.1109/TNN.2011.2120620 ITNNEP 1045-9227 Google Scholar
J. Q. Li, W. Lin and Y. Fang,
“No-reference quality assessment for multiply-distorted images in gradient domain,”
IEEE Signal Process. Lett., 23
(4), 541
–545
(2016). https://doi.org/10.1109/LSP.2016.2537321 IESPEJ 1070-9908 Google Scholar
Q. Li et al.,
“Blind image quality assessment using statistical structural and luminance features,”
IEEE Trans. Multimedia, 18
(12), 2457
–2469
(2016). https://doi.org/10.1109/TMM.2016.2601028 Google Scholar
M. Oszust,
“Local feature descriptor and derivative filters for blind image quality assessment,”
IEEE Signal Process. Lett., 26
(2), 322
–326
(2019). https://doi.org/10.1109/LSP.2019.2891416 IESPEJ 1070-9908 Google Scholar
M. Oszust,
“No-reference image quality assessment using image statistics and robust feature descriptors,”
IEEE Signal Process. Lett., 24
(11), 1656
–1660
(2017). https://doi.org/10.1109/LSP.2017.2754539 IESPEJ 1070-9908 Google Scholar
Y. Chen et al.,
“No-reference image quality assessment based on differential excitation,”
Acta Autom. Sin., 46
(8), 1727
–1737
(2020). THHPAY 0254-4156 Google Scholar
D. Ghadiyaram and A. C. Bovik,
“Massive online crowdsourced study of subjective and objective picture quality,”
IEEE Trans. Image Process., 25
(1), 372
–387
(2016). https://doi.org/10.1109/TIP.2015.2500021 IIPRE4 1057-7149 Google Scholar
X. Yang, F. Li and H. Liu,
“A survey of DNN methods for blind image quality assessment,”
IEEE Access, 7 123788
–123806
(2019). https://doi.org/10.1109/ACCESS.2019.2938900 Google Scholar
S. Bianco et al.,
“On the use of deep learning for blind image quality assessment,”
Signal Image Video Process., 12
(2), 355
–362
(2018). https://doi.org/10.1007/s11760-017-1166-8 Google Scholar
X. Liu, J. Van De Weijer and A. D. Bagdanov,
“RankIQA: learning from rankings for no-reference image quality assessment,”
in IEEE Int. Conf. Comput. Vision,
1040
–1049
(2017). https://doi.org/10.1109/ICCV.2017.118 Google Scholar
K. Ma et al.,
“End-to-end blind image quality assessment using deep neural networks,”
IEEE Trans. Image Process., 27
(3), 1202
–1213
(2018). https://doi.org/10.1109/TIP.2017.2774045 IIPRE4 1057-7149 Google Scholar
H. Zhu et al.,
“MetaIQA: deep meta-learning for no-reference image quality assessment,”
in IEEE/CVF Conf. Comput. Vision Pattern Recognit.,
14131
–14140
(2020). https://doi.org/10.1109/CVPR42600.2020.01415 Google Scholar
S. Zhang et al.,
“CNN-based medical ultrasound image quality assessment,”
Complexity, 2021 9938367
(2021). https://doi.org/10.1155/2021/9938367 COMPFS 1076-2787 Google Scholar
S. Liu et al.,
“Real-time quality assessment of pediatric MRI via semi-supervised deep nonlocal residual neural networks,”
IEEE Trans. Image Process., 29 7697
–7706
(2020). https://doi.org/10.1109/TIP.2020.2992079 IIPRE4 1057-7149 Google Scholar
J. Wang et al.,
“Deep learning for quality assessment of retinal OCT images,”
Biomed. Opt. Express, 10
(12), 6057
(2019). https://doi.org/10.1364/BOE.10.006057 BOEICL 2156-7085 Google Scholar
L. S. Chow and R. Paramesran,
“Review of medical image quality assessment,”
Biomed. Signal Process. Control, 27 145
–154
(2016). https://doi.org/10.1016/j.bspc.2016.02.006 Google Scholar
A. Kamen et al.,
“Automatic tissue differentiation based on confocal endomicroscopic images for intraoperative guidance in neurosurgery,”
Biomed. Res. Int., 2016 6183218
(2016). https://doi.org/10.1155/2016/6183218 Google Scholar
M. Izadyyazdanabadi et al.,
“Improving utility of brain tumor confocal laser endomicroscopy: objective value assessment and diagnostic frame detection with convolutional neural networks,”
Orlando, Florida
(2017). Google Scholar
M. Izadyyazdanabadi et al.,
“Convolutional neural networks: ensemble modeling, fine-tuning and unsupervised semantic localization for neurosurgical CLE images,”
J. Vis. Commun. Image Represent., 54 10
–20
(2018). https://doi.org/10.1016/j.jvcir.2018.04.004 JVCRE7 1047-3203 Google Scholar
V. Becker et al.,
“Intravenous application of fluorescein for confocal laser scanning microscopy: evaluation of contrast dynamics and image quality with increasing injection-to-imaging time,”
Gastrointest. Endosc., 68
(2), 319
–323
(2008). https://doi.org/10.1016/j.gie.2008.01.033 Google Scholar
A. K. Jain, Fundamentals of Digital Image Processing, Prentice-Hall, Englewood Cliffs, New Jersey
(1989). Google Scholar
J. Chen et al.,
“WLD: a robust local image descriptor,”
IEEE Trans. Pattern Anal. Mach. Intell., 32
(9), 1705
–1720
(2010). https://doi.org/10.1109/TPAMI.2009.155 ITPIDJ 0162-8828 Google Scholar
T. Ojala et al.,
“Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,”
IEEE Trans. Pattern Anal. Mach. Intell., 24
(7), 971
–987
(2002). https://doi.org/10.1109/TPAMI.2002.1017623 ITPIDJ 0162-8828 Google Scholar
W.-H. Liao,
“Region description using extended local ternary patterns,”
in 20th Int. Conf. Pattern Recognit.,
1003
–1006
(2010). https://doi.org/10.1109/ICPR.2010.251 Google Scholar
C. Chang and C. Lin,
“LIBSVM: a library for support vector machines,”
ACM Trans. Intell. Syst. Technol., 2
(3), 1
–27
(2011). https://doi.org/10.1145/1961189.1961199 Google Scholar
L. Liu et al.,
“Binocular spatial activity and reverse saliency driven no-reference stereopair quality assessment,”
Signal Process. Image Commun., 58 287
–299
(2017). https://doi.org/10.1016/j.image.2017.08.011 SPICEF 0923-5965 Google Scholar
M. Oszust,
“Optimized filtering with binary descriptor for blind image quality assessment,”
IEEE Access, 6 42917
–42929
(2018). https://doi.org/10.1109/ACCESS.2018.2860127 Google Scholar
M. Oszust,
“No-reference image quality assessment with local gradient orientations,”
Symmetry-Basel., 11
(1), 95
(2019). https://doi.org/10.3390/sym11010095 Google Scholar
, “Final report from the video quality experts group on the validation of objective models of video quality assessment—Phase II,”
(2003) http://www.vqeg.org/ Google Scholar
C. Nadeau and Y. Bengio,
“Inference for the generalization error,”
Mach. Learn., 52 239
–281
(2003). https://doi.org/10.1023/A:1024068626366 MALEEZ 0885-6125 Google Scholar
P. G. Freitas, W. Y. L. Akamine and M. C. Q. Farias,
“No-reference image quality assessment based on statistics of local ternary pattern,”
in Eighth Int. Conf. Quality Multimedia Experience,
1
–6
(2016). Google Scholar
BiographyXiangjiang Dong is a master student at Huazhong University of Science and Technology. He received his BS degree in physics from Huazhong University of Science and Technology in 2016. His current research interests include medical image processing and image quality assessment. |