Open Access
25 July 2022 PADAr: physician-oriented artificial intelligence-facilitating diagnosis aid for retinal diseases
Po-Kang Lin, Yu-Hsien Chiu, Chiu-Jung Huang, Chien-Yao Wang, Mei-Lien Pan, Da-Wei Wang, Hong-Yuan Mark Liao, Yong-Sheng Chen, Chieh-Hsiung Kuan, Shih-Yen Lin, Li-Fen Chen
Author Affiliations +
Funded by: Ministry of Science and Technology, Taiwan (MOST)
Abstract

Purpose: Retinopathy screening via digital imaging is promising for early detection and timely treatment, and tracking retinopathic abnormality over time can help to reveal the risk of disease progression. We developed an innovative physician-oriented artificial intelligence-facilitating diagnosis aid system for retinal diseases for screening multiple retinopathies and monitoring the regions of potential abnormality over time.

Approach: Our dataset contains 4908 fundus images from 304 eyes with image-level annotations, including diabetic retinopathy, age-related macular degeneration, cellophane maculopathy, pathological myopia, and healthy control (HC). The screening model utilized a VGG-based feature extractor and multiple-binary convolutional neural network-based classifiers. Images in time series were aligned via affine transforms estimated through speeded-up robust features. Heatmaps of retinopathy were generated from the feature extractor using gradient-weighted class activation mapping++, and individual candidate retinopathy sites were identified from the heatmaps using clustering algorithm. Nested cross-validation with a train-to-test split of 80% to 20% was used to evaluate the performance of the screening model.

Results: Our screening model achieved 99% accuracy, 93% sensitivity, and 97% specificity in discriminating between patients with retinopathy and HCs. For discriminating between types of retinopathy, our model achieved an averaged performance of 80% accuracy, 78% sensitivity, 94% specificity, 79% F1-score, and Cohen’s kappa coefficient of 0.70. Moreover, visualization results were also shown to provide reasonable candidate sites of retinopathy.

Conclusions: Our results demonstrated the capability of the proposed model for extracting diagnostic information of the abnormality and lesion locations, which allows clinicians to focus on patient-centered treatment and untangles the pathological plausibility hidden in deep learning models.

1.

Introduction

Retinopathy is an important cause of visual impairment, which is generally irreversible in its later stages. The resulting presentation of drusen, cellophane, exudate, hemorrhage, or chorioretinal scarring can have a profound effect on the vision of its victims, in which the most common causes may be diabetic retinopathy (DR),1 age-related macular degeneration (AMD),2 cellophane maculopathy (CM),3 and pathological myopia (PM).4 The asymptomatic nature of retinopathy in the initial stages means that regular screening via digital imaging is promising for early detection and timely treatment.5

Color fundus imaging is a non-invasive cost-effective tool for ophthalmological examinations.6 A number of models based on convolutional neural networks (CNNs) have been developed to facilitate the classification of retinopathies based on color fundus images.712 One recent CNN-based study reported that salient regions obtained from gradient-weighted class activation mapping (Grad-CAM++)13 closely matched the regions identified by ophthalmologists.14 Retinopathic changes over time can be used to monitor disease progression and evaluate therapeutic outcomes.15 Clinical ophthalmologists rely heavily on digital imaging for diagnostics; however, manual tracking can be arduous and time-consuming. Clinicians require user-friendly computer-aided diagnostic tools to automate the process of identifying regions with retinopathic abnormalities, and to monitor changes in those areas over time in order to facilitate decision-making and thereby alleviate their workload.

In the current study, we developed an artificial intelligent (AI) diagnostics platform for screening multiple retinopathies and monitoring regions of potential abnormality over time. A schematic illustration of the proposed system, referred to as the physician-oriented AI-facilitating diagnosis aid system for retinal diseases (PADAr), is shown in Fig. 1. We employed machine learning techniques based on fundus images from 304 eyes affected by AMD, DR, CM, or PM, as well as healthy controls (HCs). It is worth noting that all training data had previously been labeled by a retina specialist (Dr. P.K. Lin). The proposed framework performs two fundamental operations: screening and monitoring. The screening model applies a shared-weight feature extractor to fundus images and then uses multiple-binary CNN-based classifiers to formulate outcome predictions. A corresponding heatmap was obtained from the last convolutional layer of the trained feature extractor using Grad-CAM++13 to highlight regions of potential abnormality, and thereby, differentiate HCs from cases requiring attention. In the second stage (i.e., monitoring model in Fig. 1 blue box), the heatmaps are registered over time using affine transforms estimated using a speeded-up robust features (SURF) descriptor1517 based on the corresponding fundus image. We applied lesion-site estimation on each transformed heatmap to visualize change in retinopathic abnormalities over time. This study proposed a novel hybrid machine learning architecture by combining CNN, SURF descriptors, and clustering to automate the process of visualizing potential lesions over time. Our findings suggest that this type of algorithm could facilitate early diagnosis and the tracking of disease progression, contingent on the development of larger, more diverse datasets.

Fig. 1

Overview of the PADA system.

JMI_9_4_044501_f001.png

2.

Materials and Methods

2.1.

Data Acquisition and Preparation

This study was approved by the Ethics Committee of the Institutional Review Board of Taipei Veterans General Hospital, Taiwan (2018-08-003CC accepted November 26, 2018). Participants provided written informed consent allowing the retrospective collection of their retinal images. Participants were included if they were diagnosed with a major retinopathy in either eye, such as AMD, DR, CM, and PM. A total of 200 participants were selected for inclusion by a retina specialist from the Department of Ophthalmology of Taipei Veterans General Hospital in Taiwan. Sampling covered the period from 2002 to 2019.

Color fundus images of multiple fields were captured using multiple cameras equipped with lenses covering a field-of-view of 35 deg to 55 deg. The multiple fields were indicated using the seven fields designated in the general early treatment of diabetic retinopathy study (ETDRS) protocol: optic disc centered field (F1); macular centered field (F2); and all peripheral fields (F3, F4, F5, F6, and F7). Images lacking anatomical landmarks (e.g., optic disc, vessels, and macula) were removed. The images were cropped to 2201×2201 pixels. For visualization, all images were resized to 512×512 pixels. The 304 eyes (N=4908) included in the study were labeled as follows: HC (25 eyes, N=367), AMD (120 eyes, N=2029), DR (77 eyes, N=1681), CM (51 eyes, N=436), and PM (31 eyes, N=395). The dataset was divided into two subsets using an 80 to 20 split; that is, 80% of images were used as training validation data (N=4082) and 20% were used as test data (N=826). Participants who underwent more than two examinations (N=160) were selected to assess abnormalities over time.

2.2.

Screening Model

The present study proposed two models. The screening model (Fig. 2) based on multi-class classification employs a shared-weight feature extractor using VGG1618 as a backbone, a sub-network with multi-binary CNN-based classifiers for generating soft-target information, and a final fully connected (FC) layer for integrating the soft-target information to predict the class and generate the corresponding heatmap. The diseases representations (14×14×512) obtained from the last convolutional layer of a shared-weight feature extractor with global average pooling. We then removed the fully connected part of the VGG16 and employed multiple binary classifiers, including a main-classifier and the six sub-classifiers, providing soft-target information to the final FC layer. Each classifier contains three FC layers with the rectified linear unit function as activation, three dropout layers with a dropout rate of 0.2, and one softmax layer.

Fig. 2

Architecture of proposed CNN-based screening model.

JMI_9_4_044501_f002.png

The main-classifier is used to discriminate between cases of retinopathy and HCs and six binary classifiers are used to differentiate cases between each pair of four types of retinopathy (AMD, DR, CM, and PM). The final FC layer integrates soft-target information obtained from all of the classifiers to predict the outcomes. We incorporated Grad-CAM++ to obtain the corresponding heatmap from the class of interest (i.e., retinopathy). Essentially, Grad-CAM++ generates the heatmap by a weighted combination of latent feature channels from the last convolutional layer. The weights for feature channels reflect their respective importance in prediction of a given class, which is estimated from the gradient of guided back-propagation. Grad-CAM++ is shown to achieve better localization compared to Grad-CAM19 by providing improved formulations for estimating the channel weights. Majority voting is used to determine the final prediction outcome for each patient.

Prior to model training, all input images were augmented by horizontal flipping, rotation [36  deg,+36  deg], and translation of width and height [10%,+10%] to resolve the problems of overfitting, small sample size, and an imbalance in available data for model training.20,21 The images (5000 in each class) were then resized to 224×224 pixels via bilinear interpolation for model training.

Training was implemented in three steps. We first replaced the last three fully connected layer of VGG16 with one binary main-classifier, utilized ImageNet22 pretrained weights to train the feature extractor from scratch and fine-tuned the network using our dataset to classify between cases of retinopathy and HCs. We then trained six binary sub-classifiers with the estimated weights of the feature extractor. Finally, we trained the final FC layer using soft-target information obtained from the trained classifiers including one binary main-classifier and six binary sub-classifiers. We used the binary cross-entropy loss function for training each binary classifier and utilized the categorical cross-entropy for training final screening model. For the hyper-parameters of all networks, we employed the Adam optimizer23 with an initial learning rate of 1×105, a final learning rate of 1×108, and batch size of 32. The learning rate decayed by a factor of ten over ten epochs showing no improvement in validation loss.

We performed 5×5-fold nested cross-validation (CV)24 to evaluate the performance of the feature extractor. No significant difference was observed among the folds from the feature extractor; therefore, we applied holdout CV for evaluating six binary sub-classifiers and the final FC layer. Model performance was measured in terms of accuracy, precision, sensitivity, specificity, F1-score, the area under curve (AUC) of receiver operating characteristic curve,25 and Cohen’s kappa coefficient.26 For each performance metric, macro-average was also calculated by the arithmetic mean of all individual classes. A retina specialist (P.K. Lin) also visually examined the candidate sites in the testing data for validating the efficacy of the proposed model.

2.3.

Visualizing Abnormalities Over Time

The second model proposed in this work was used to monitor and visualize candidate lesion sites based on results from the aforementioned screening model at various time points for each patient. Time-series image registration was adopted to align images acquired from multiple time points from a single participant. For each image, control points were automatically extracted using SURF algorithm, and the time-series images and their corresponding heatmaps were registered to the reference image. Subsequently, a clustering algorithm was used to identify candidate sites based on their relevance to identified abnormalities.

2.3.1.

Time-series image registration

The schema of the proposed time-series image registration method is shown in Fig. 3, including image selection, control point extraction, and control point matching. For each image, we first detected the location of the optic disc (Xdisc,Ydisc) using pixel-wise distance regression based optic disc detection approach.27 The region of interest (ROI) was defined as (Xic±0.3×Imagewidth,Yic±0.25×Imageheight), where ic refers to the image center. Images with the disc located within the ROI were selected as macula-center fundus images. For each patient, the macula-center image with the shortest distance between disc location and the center of ROI was then selected as a reference for registration.

Fig. 3

Illustration of proposed time-series image registration. The mosaic image indicates registration performance by combining reference and target images.

JMI_9_4_044501_f003.png

Second, a green channel image extracted from each macula-center image was enhanced using the contrast limited adaptive histogram equalization filtering algorithm,28 whereupon the intensity was normalized to [0, 1] and resampled to 512×512 pixels. The field-of-view binary mask was derived using Otsu’s thresholding,29 followed by an erosion operator with 5 mm around the edge of the mask. Control points were extracted using the SURF algorithm.1517

Third, the correspondence between control points SX in the reference image (X) and control points SY in every macular-centered image (Y) was estimated using the efficient approximate nearest neighbor search,30 which computes the pairwise Euclidean distance between SX and SY. The affine transformation matrix of each SX and SY pair was then estimated from the predicted correspondence and applied to the corresponding macular-center image using the robust m-estimator sample consensus algorithm.31 Finally, each candidate lesion site in the reference image was aligned to specific candidates in each of the transformed macular-center images (acquired at different time points) by calculating the shortest distance between the reference and the target candidates.

2.3.2.

Identifying candidate lesion sites

An adaptive clustering algorithm was used to locate potential regions of abnormality (i.e., candidate lesion sites) on the heatmap derived from the screening model. The pipeline of our algorithm is shown in Fig. 4. The heatmap was first up-sampled to 512×512 pixels via bilinear interpolation. To visualize the abnormalities, we followed the standard procedure of Grad-CAM++13 and up-sampled the heatmaps to match the display image resolution (512×512 pixels) via bilinear interpolation. The intensity of the resulting heatmaps was then normalized to [0,1], followed by thresholding using the following Eq. (1):

Eq. (1)

Threshold=E(H)+σ,
where E(H) and σ refer to the mean and standard deviation of heatmap H intensity, respectively. We then determined the optimal number of clusters (K) with maximum silhouette coefficient.32,33 Finally, we utilized a Gaussian mixture model34 to group pixels into clusters, each of which represented one candidate lesion site.

Fig. 4

Pipeline of adaptive clustering algorithm. GMM: Gaussian mixture model.

JMI_9_4_044501_f004.png

3.

Results

3.1.

Screening Performance

In terms of screening, the proposed multi-binary-classifier model achieved macro-average accuracy of 0.80, precision of 0.81, sensitivity of 0.78, specificity of 0.94, F1 score of 0.79 (Table 1), AUC of 0.94, and Cohen’s kappa coefficient of 0.70. These results were obtained from uncleaned images captured during funduscopic examinations. A confusion matrix is presented in Fig. 5(a). As shown in Table 1 and Fig. 5(b), the removal of poor-quality images improved average accuracy by 5.4% and precision as follows: AMD (4%), DR (5%), CM (8%), and PM (10%).

Table 1

Performance of our proposed model. Numbers in parentheses indicate results based on recalculations following the removal of poor-quality images. AMD: age-related macular degeneration; DR: diabetic retinopathy; CM: cellophane maculopathy; PM: pathological myopia; and HC: healthy control.

PrecisionSensitivitySpecificityF1-scoreAUCCohen’s Kappa
AMD0.79 (0.83)0.78 (0.86)0.87 (0.90)0.79 (0.85)0.89
DR0.78 (0.86)0.84 (0.89)0.85 (0.90)0.81 (0.87)0.92
CM0.68 (0.73)0.45 (0.52)0.98 (0.98)0.55 (0.61)0.90
PM0.79 (0.89)0.82 (0.85)0.98 (0.99)0.81 (0.87)0.98
HC1.00 (1.00)1.00 (1.00)1.00 (1.00)1.00 (1.00)1.00
Macro average0.81 (0.86)0.78 (0.85)0.94 (0.96)0.79 (0.85)0.9390.701 (0.785)

Fig. 5

Confusion matrix of the screening model (a) before and (b) after removal of poor-quality images. AMD: age-related macular degeneration; DR: diabetic retinopathy; CM: cellophane maculopathy; PM: pathological myopia; and HC: health control.

JMI_9_4_044501_f005.png

The performance of the main classifier was assessed using repeated nested CV with an outer five-fold CV and an inner five-fold CV. The main-classifier for discriminating between patients and HCs achieved high mean-macro-average accuracy of 0.99±0.003 (precision, 0.99±0.005; recall, 0.93±0.017; F1 score, 0.96±0.012; and Cohen’s kappa coefficient, 0.91±0.023).

3.2.

Candidate Regions of Disease

The proposed model provides diagnostic information from heatmaps pertaining to the retina for use in identifying candidate locations of disease. Figures 6(a) and 6(b) show images showing examples of regional retinopathy in the AMD, including drusen and edema. Figure 6(c) illustrates instances of hemorrhage and exudate in a case of DR. The heatmap of CM in Fig. 6(d) focuses on the optic disc extending to the macula. The heatmap of PM in Fig. 6(e) focuses on the crescent near the disc and macular degeneration. Figure 6(f) highlights drusen and exudate. Note that most of the heatmaps highlighted the optic disk.

Fig. 6

Heatmaps of five retinopathy images generated from screening model: (a) dry-type AMD; (b) wet-type AMD; (c) DR; (d) CM; (e) PM; and (f) both AMD and DR. The original fundus images and corresponding heatmaps are respectively presented in the first and third columns. The second column displays the original images overlaid with their corresponding heatmaps. The fourth column displays the original images overlaid with their corresponding heatmap and candidate lesion-sites (in red), highlighting potential regions of abnormality. AMD: age-related macular degeneration; DR: diabetic retinopathy; CM: cellophane maculopathy; and PM: pathological myopia.

JMI_9_4_044501_f006.png

Figure 7 shows two examples of prediction error, in which an image of the AMD was misclassified as DR [Fig. 7(a)] and an image of the DR was misclassified as AMD [Fig. 7(b)]. Regardless, the heatmaps provided reasonable candidate sites of retinopathy, including sites around the macula and optic disc (third column in the panel).

Fig. 7

Heatmaps of misclassified cases. (a) AMD case misclassified as DR and (b) DR misclassified as AMD (AMD: age-related macular degeneration; DR: diabetic retinopathy).

JMI_9_4_044501_f007.png

3.3.

Visualizing Candidate Regions of Abnormality Over Time

Figure 8 shows two cases illustrating changes in retinopathy along time. In Fig. 8(a), the proposed system highlighted candidate retinopathic abnormalities in the AMD (e.g., drusen), in which the condition remained stable in subsequent yearly follow-up examinations. In Fig. 8(b), the system highlighted the progress of exudate and hemorrhage in the DR in monthly follow-up examinations, in which the severity of the conditions gradually decreased. These results demonstrate the effectiveness of the system in tracking retinopathies via funduscopic examination.

Fig. 8

Visualization results from two fundus images obtained at different timepoints, where (1), (2), and (3) denote the first-, second-, and third-time scans, respectively. (a) and (b) the left column displays the original color fundus images. The middle column displays lesion-site candidates over time. The right column displays close-up images from one of the lesion-site candidates indicating a potential region of abnormality. It is worth noting that color is used to differentiate specific candidates over time. Green bounding boxes indicate correctly identified regions, whereas the red bounding boxes denote the miss-detection regions.

JMI_9_4_044501_f008.png

4.

Discussion

Locating abnormalities in the retina is crucial to diagnostic decision-making. Previous studies have reported that heatmaps obtained from Grad-CAM++ can be used to highlight such abnormalities in instances of single retinopathy (e.g., AMD or DR).14,35,36 Note however that many patients suffer more than one retinopathy in either or both eyes; therefore, we proposed the use of a main-classifier to differentiate patients from HCs in order to detect all potential abnormalities within the regions identified by the retina specialist (Dr. P.K. Lin). As shown in Figs. 6 and 7, the resulting heatmap was able to locate all potential regions of abnormality, regardless of whether the prediction outcome was correct. Our weakly supervised approach to learning pixel-wise labeling directly from image-level annotation is meant to reduce the effort required to label ground-truth locations of retinopathy. Experiments demonstrated the feasibility and efficacy of the proposed method in locating potential sites of retinopathic abnormality.

Our model also demonstrated competitive classification performance when compared to the other retinopathy detection models in the literature, either in distinguishing retinopathy from HCs or discriminating between types of retinopathy. Table 2 gives the reported performance of other binary classification models. Compared to other binary retinopathy classification models, the proposed system demonstrated superior sensitivity compared to other models. It is worth noting that most binary classification models in the literature only involve detecting a single type of retinopathy. In contrast, the binary classification in our study involves distinguishing four types of retinopathy from HCs. With larger diversity in the disease characteristics, it is thereby a more difficult task compared to detecting a single disease type. Nonetheless, although studies by Gulshan et al.37 and Zhang et al.38 reported higher AUC and F1 score, respectively, the proposed model still achieved competitive performance in most respect under significant larger disease diversity.

Table 2

The reported performance of binary retinopathy classification models in the reviewed literature, compared with the proposed system. The best performance according to each metric is highlighted by boldface.

DatabaseClassificationAUCAccuracySensitivitySpecificityF1 score
Gargeya and Leng35Private datasetDR0.970.940.98
E-OphthaDR0.950.90.94
Choi et al.8STARENine diseases0.9030.8030.855
Tan et al.10Private datasetAMD0.95450.96430.9375
Gulshan et al.37Private datasetDR0.980.9210.952
Zhang et al.38Private datasetDR0.980.980.98
Zago et al.39MessidorDR0.9120.94
Das et al.40DIARETDB1 (train)DR0.9740.9760.972
Private dataset (test)
ProposedPrivate datasetFour diseases0.9390.990.980.970.85

Table 3 shows the reported performance in the literature for distinguishing between multiple retinopathy types. Compared to other models, the proposed system demonstrated superior sensitivity. It is worth noting that the classification of CM yielded a lower sensitivity compared to other types of retinopathy. As previously shown in Fig. 5, CM is occasionally confused with other types of retinopathy, such as AMD and DR. We hypothesize that this lower performance is attributable to the confounding effects by the prevalence of myopia in Taiwan. Nonetheless, the proposed system still serves as an effective screening tool with its ability to accurately detect the presence of retinopathy, despite occasional confusion between retinopathy types. In-depth examinations using other imaging techniques (such as optical coherence tomography and fluorescein angiography) can be used after the screening stage for a more accurate diagnosis of retinopathy types.

Table 3

The reported performance of multi-class retinopathy classification models in the reviewed literature, compared with the proposed system. ARIA: automated retinal image analysis database; STARE: structured analysis of the retina database; ODIR: ocular disease intelligent recognition database.

DatabaseModelClassificationAUCAccuracySensitivitySpecificityF1-scoreKappa
Arunkumar, et al.12ARIADimension reduced deep learningThree classes (AMD/DR/normal)0.96730.79320.9673
Choi, et al.8STAREVGG19 random forest10 classes (normal, BDR, PDR, Dry AMD, Wet AMD, RVO, RAO, hypertensive retinopathy, coat’s disease, and retinitis) three classes (normal, BDR, and dry AMD)0.3050.224
0.7280.577
Gour, et al.41ODIRVGG16-SGDEight classes (normal, diabetes, glaucoma, catareact, AMD, hypertension, myopia, and other)0.68880.89060.8557
Normal-0.660.770.21
Glaucoma0.670.40.6
Diabetic retinopathy0.930.050.94
AMD0.940.060.93
Hypertension0.9500.99
Cataract0.9601
Myopia0.940.110.94
Other abnormalities0.730.740.32
Rajan, et al.42STARECNN10 classes (normal, BDR, PDR, Dry AMD, Wet AMD, RVO, RAO, hypertensive retinopathy, coat’s disease, and retinitis)0.42
Proposedprivate datasetVGG16-basedFive classes (normal, AMD, DR, PM, and CM)0.930.860.850.940.790.7
Normal11111
DR0.910.890.840.850.81
AMD0.890.860.780.870.79
PM0.970.850.820.980.81
CM0.900.520.450.980.55

It is worth noting that our study incorporated real-world data with minimal data cleaning and annotations. In the literature, screening models trained in real-world clinical settings are generally outperformed by those trained in a laboratory setting with carefully selected data,43,44 due to noise or artifacts originated from sub-optimal imaging equipment, patient movement, or exposure error.45,46 Nevertheless, our comparison results demonstrate that the proposed model achieved comparable performance to models trained with carefully selected data. Additionally, the proposed system infers location information from eye-based annotation in a weakly supervised manner, by which we sought to preserve the subclinical features of fundus images and to mitigate the labor-intensive annotation process. How to improve the detection performance and localization ability under the real-world data paradigm will be one of our future focus.

Monitoring disease progression from multiple examinations performed on different days provides quantitative and qualitative information by which to monitor disease progression. This process is critical to ensuring timely treatment; however, the process is time-consuming. Recent studies have reported that the discrimination of disease stage can help to reveal the risk of disease progression, particularly in areas such as the AMD and DR.47,48 Sequential changes in retinopathic characteristics observed in fundus images can be used to detail the evolution of retinopathy progression. In the current study, we developed a novel user-friendly tool by which to obtain assessments tailored to the individual for use in pinpointing the location of abnormalities from a single fundus image and visualizing changes in the corresponding disease spot region over time.

To the best of our knowledge, this is the first attempt to automate the location and visualization of retinopathic regions in the temporal domain. Our results demonstrates the capability of the proposed PADAr to identify potential retinopathy sites and perform longitudinal follow-ups of disease progression, suggesting its feasibility for facilitating clinicians in their decision-making process and focusing on patient-centered treatment.

Disclosures

No conflicts of interest, financial or otherwise, are declared by the authors.

Acknowledgments

This work was supported in part by the grants from Ministry of Science and Technology, Taiwan (Grant Nos. MOST106-2218-E-010-004-MY3, MOST109-2327-B-010-005-(4), MOST109-2314-B-010-027, and MOST110-2314-B-A49A-529), Veterans General Hospitals and University System of Taiwan Joint Research Program, Taipei, Taiwan (Grant Nos. VGHUST108-G1-2-2 and VGHUST108-G1-2-1), Cheng Hsin General Hospital Foundation, Taipei, Taiwan (Grant No. CY11002), College of Medicine of National Yang Ming Chiao Tung University, Taipei, Taiwan (Grant No. 107F-M01-0611), and Thematic Research Program of Institute of Information Science: Digital Medicine Initiative, Institute of Information Science, Academia Sinica, Taipei, Taiwan.

Code availability

The code to develop the screening model is based on Keras using TensorFlow as backend. Custom code was specific to our computing infrastructure and mainly used for data input/output and parallelization across computers.

References

1. 

M. R. Mookiah et al., “Computer-aided diagnosis of diabetic retinopathy: a review,” Comput. Biol. Med., 43 (12), 2136 –2155 (2013). https://doi.org/10.1016/j.compbiomed.2013.10.007 CBMDAW 0010-4825 Google Scholar

2. 

A.-R. E. D. S. R. Group, “The age-related eye disease study system for classifying age-related macular degeneration from stereoscopic color fundus photographs: the age-related eye disease study report number 6,” Am. J. Ophthalmol., 132 (5), 668 –681 (2001). https://doi.org/10.1016/S0002-9394(01)01218-1 AJOPAA 0002-9394 Google Scholar

3. 

C. J. Pournaras et al., “Macular epiretinal membranes,” Semin. Ophthalmol., 15 (2), 100 –107 (2000). https://doi.org/10.3109/08820530009040000 SEOPE7 Google Scholar

4. 

K. Neelam et al., “Choroidal neovascularization in pathological myopia,” Prog. Retin. Eye Res., 31 (5), 495 –525 (2012). https://doi.org/10.1016/j.preteyeres.2012.04.001 PRTRES 1350-9462 Google Scholar

5. 

A. Dietzel et al., “Automatic detection of diabetic retinopathy and its progression in sequential fundus images of patients with diabetes,” Acta Ophthalmol., 97 (4), e667 –e669 (2019). https://doi.org/10.1111/aos.13976 Google Scholar

6. 

Y. Yan et al., “Classification of artery and vein in retinal fundus images based on the context-dependent features,” Digital Human Modeling. Applications in Health, Safety, Ergonomics, and Risk Management: Ergonomics and Design, 198 –213 Springer International Publishing.Google Scholar

7. 

M. D. Abramoff et al., “Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning,” Invest. Ophthalmol. Vis. Sci., 57 (13), 5200 –5206 (2016). https://doi.org/10.1167/Iovs.16-19964 IOVSDA 0146-0404 Google Scholar

8. 

J. Y. Choi et al., “Multi-categorical deep learning neural network to classify retinal images: a pilot study employing small database,” PLoS One, 12 (11), e0187336 (2017). https://doi.org/10.1371/journal.pone.0187336 POLNCL 1932-6203 Google Scholar

9. 

J. A. de Sousa et al., “Texture based on geostatistic for glaucoma diagnosis from fundus eye image,” Multimedia Tools Appl., 76 (18), 19173 –19190 (2017). https://doi.org/10.1007/s11042-017-4608-y Google Scholar

10. 

J. H. Tan et al., “Age-related macular degeneration detection using deep convolutional neural network,” Future Gener. Comput. Syst. Int. J. Esci., 87 127 –135 (2018). https://doi.org/10.1016/j.future.2018.05.001 Google Scholar

11. 

V. V. Kamble and R. D. Kokate, “Automated diabetic retinopathy detection using radial basis function,” in Int. Conf. Comput. Intell. and Data Sci., 799 –808 (2020). Google Scholar

12. 

R. Arunkumar and P. Karthigaikumar, “Multi-retinal disease classification by reduced deep learning features,” Neural Comput. Appl., 28 (2), 329 –334 (2015). https://doi.org/10.1007/s00521-015-2059-9 Google Scholar

13. 

A. Chattopadhay et al., “Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks,” in IEEE Winter Conf. Appl. Comput. Vision (WACV), 839 –847 (2018). Google Scholar

14. 

Q. Meng, Y. Hashimoto and S. Satoh, “Fundus image classification and retinal disease localization with limited supervision,” in Asian Conf. Pattern Recognit., 469 –482 Google Scholar

15. 

S. K. Saha et al., “A two-step approach for longitudinal registration of retinal images,” J. Med. Syst., 40 (12), 277 (2016). https://doi.org/10.1007/s10916-016-0640-0 JMSYDA 0148-5598 Google Scholar

16. 

H. Bay et al., “Speeded-up robust features (SURF),” Comput. Vision Image Understanding, 110 (3), 346 –359 (2008). https://doi.org/10.1016/j.cviu.2007.09.014 Google Scholar

17. 

C. Hernandez-Matas, X. Zabulis and A. A. Argyros, “Retinal image registration based on keypoint correspondences, spherical eye modeling and camera pose estimation,” in 37th Annu. Int. Conf. IEEE Eng. in Med. and Biol. Soc. (EMBC), 5650 –5654 (2015). https://doi.org/10.1109/EMBC.2015.7319674 Google Scholar

18. 

K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,” in 3rd Int. Conf. Learn. Represent., (2015). Google Scholar

19. 

R. R. Selvaraju et al., “Grad-CAM: visual explanations from deep networks via gradient-based localization,” Int. J. Comput. Vision, 128 (2), 336 –359 (2019). https://doi.org/10.1007/s11263-019-01228-7 IJCVEQ 0920-5691 Google Scholar

20. 

A. Krizhevsky, I. Sutskever and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Adv. Neural Inf. Process. Syst., 1097 –1105 Google Scholar

21. 

T. Zhou, S. Ruan and S. Canu, “A review: deep learning for medical image segmentation using multi-modality fusion,” Array, 3-4 100004 (2019). https://doi.org/10.1016/j.array.2019.100004 Google Scholar

22. 

J. Deng et al., “ImageNet: a large-scale hierarchical image database,” in IEEE Conf. Comput. Vision and Pattern Recognit., 248 –255 (2009). https://doi.org/10.1109/CVPR.2009.5206848 Google Scholar

23. 

D. P. Kingma and J. Ba, “Adam: a method for stochastic optimization,” in 3rd Int. Conf. Learn. Represent., (2015). Google Scholar

24. 

D. Krstajic et al., “Cross-validation pitfalls when selecting and assessing regression and classification models,” J. Cheminf., 6 (1), 10 (2014). https://doi.org/10.1186/1758-2946-6-10 Google Scholar

25. 

T. Fawcett, “An introduction to ROC analysis,” Pattern Recognit. Lett., 27 (8), 861 –874 (2006). https://doi.org/10.1016/j.patrec.2005.10.010 PRLEDG 0167-8655 Google Scholar

26. 

J. R. Landis and G. G. Koch, “The measurement of observer agreement for categorical data,” Biometrics, 33 (1), 159 –74 (1977). https://doi.org/10.2307/2529310 BIOMB6 0006-341X Google Scholar

27. 

M. I. Meyer et al., “A pixel-wise distance regression approach for joint retinal optical disc and fovea detection,” Lect. Notes Comput. Sci., 11071 39 –47 (2018). https://doi.org/10.1007/978-3-030-00934-2_5 LNCSD9 0302-9743 Google Scholar

28. 

K. Zuiderveld, Contrast Limited Adaptive Histogram Equalization, 474 –485 Academic Press Professional, Inc. (1994). Google Scholar

29. 

N. Otsu, “A threshold selection method from gray-level histograms,” IEEE Trans. Syst., Man, Cybern., 9 (1), 62 –66 (1979). https://doi.org/10.1109/TSMC.1979.4310076 Google Scholar

30. 

M. Muja and D. G. Lowe, “Fast approximate nearest neighbors with automatic algorithm configuration,” in VISAPP 2009: Proc. Fourth Int. Conf. Comput. Vision Theory and Appl., 331 –340 (2009). Google Scholar

31. 

P. H. S. Torr and A. Zisserman, “MLESAC: a new robust estimator with application to estimating image geometry,” Comput. Vision Image Understanding, 78 (1), 138 –156 (2000). https://doi.org/10.1006/cviu.1999.0832 Google Scholar

32. 

P. J. Rousseeuw, “Silhouettes—a graphical aid to the interpretation and validation of cluster-analysis,” J. Comput. Appl. Math., 20 53 –65 (1987). https://doi.org/10.1016/0377-0427(87)90125-7 JCAMDI 0377-0427 Google Scholar

33. 

L. Kaufman and P. J. Rousseeuw, Finding Groups in Data: An Introduction to Cluster Analysis, 344 John Wiley & Sons(2009). Google Scholar

34. 

J. Han, J. Pei and M. Kamber, Data Mining: Concepts and Techniques, Elsevier(2011). Google Scholar

35. 

R. Gargeya and T. Leng, “Automated identification of diabetic retinopathy using deep learning,” Ophthalmology, 124 (7), 962 –969 (2017). https://doi.org/10.1016/j.ophtha.2017.02.008 OPANEW 0743-751X Google Scholar

36. 

W. M. Gondal et al., “Weakly-supervised localization of diabetic retinopathy lesions in retinal fundus images,” in IEEE Int. Conf. Image Process., 2069 –2073 (2017). https://doi.org/10.1109/ICIP.2017.8296646 Google Scholar

37. 

V. Gulshan et al., “Performance of a deep-learning algorithm vs manual grading for detecting diabetic retinopathy in India,” JAMA Ophthalmol., 137 (9), 987 –993 (2019). https://doi.org/10.1001/jamaophthalmol.2019.2004 Google Scholar

38. 

W. Zhang et al., “Automated identification and grading system of diabetic retinopathy using deep neural networks,” Knowl.-Based Syst., 175 12 –25 (2019). https://doi.org/10.1016/j.knosys.2019.03.016 KNSYET 0950-7051 Google Scholar

39. 

G. T. Zago et al., “Diabetic retinopathy detection using red lesion localization and convolutional neural networks,” Comput. Biol. Med., 116 103537 (2020). https://doi.org/10.1016/j.compbiomed.2019.103537 CBMDAW 0010-4825 Google Scholar

40. 

S. Das et al., “Deep learning architecture based on segmented fundus image features for classification of diabetic retinopathy,” Biomed. Signal Process. Control, 68 102600 (2021). https://doi.org/10.1016/j.bspc.2021.102600 Google Scholar

41. 

N. Gour and P. Khanna, “Multi-class multi-label ophthalmological disease detection using transfer learning based convolutional neural network,” Biomed. Signal Process. Control, 66 102329 (2021). https://doi.org/10.1016/j.bspc.2020.102329 Google Scholar

42. 

K. Rajan and C. Sreejith, “Retinal Image Processing and Classification Using Convolutional Neural Networks,” Int. Conf. ISMAC in Comput. Vision and Bio-Eng., 1271 –1280 Springer(2019). Google Scholar

43. 

F. D. Verbraak et al., “Diagnostic accuracy of a device for the automated detection of diabetic retinopathy in a primary care setting,” Diabetes Care, 42 (4), 651 –656 (2019). https://doi.org/10.2337/dc18-0148 DICAD2 0149-5992 Google Scholar

44. 

Y. T. Hsieh et al., “Application of deep learning image assessment software verisee for diabetic retinopathy screening,” J. Formos Med. Assoc., 120 (1 Pt 1), 165 –171 (2021). https://doi.org/10.1016/j.jfma.2020.03.024 Google Scholar

45. 

L. Giancardo et al., “Quality assessment of retinal fundus images using elliptical local vessel density,” New Developments in Biomedical Engineering, IntechOpen(2010). Google Scholar

46. 

Z. Shen et al., “Modeling and enhancing low-quality retinal fundus images,” IEEE Trans. Med. Imaging, 40 (3), 996 –1006 (2021). https://doi.org/10.1109/TMI.2020.3043495 ITMID4 0278-0062 Google Scholar

47. 

F. Grassmann et al., “A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography,” Ophthalmology, 125 (9), 1410 –1420 (2018). https://doi.org/10.1016/j.ophtha.2018.02.037 OPANEW 0743-751X Google Scholar

48. 

F. Arcadu et al., “Deep learning algorithm predicts diabetic retinopathy progression in individual patients,” NPJ Digit. Med., 2 92 (2019). https://doi.org/10.1038/s41746-019-0172-3 Google Scholar

Biography

Po-Kang Lin graduated from National Yang Ming University, Taiwan. He had completed resident training and retina fellowship at Taipei Veterans General Hospital, Taiwan. Currently, he is an associate professor of National Yang Ming Chiao Tung University, Taiwan, and the director of retina section of Taipei Veterans General Hospital. Also, he is the director of Taiwan Society of Luminescence Science. His researches are focused on clinical ophthalmology, retinal, retinal biology, and retinal prosthesis.

Yu-Hsien Chiu is currently a research assistant at the Institute of Brain Science, National Yang Ming Chiao Tung University, Taiwan. He received his BS degree in biomedical engineering from Ming Chuan University, Taoyuan, Taiwan, in 2016, and his MS degree at the Institute of Brain Science, National Yang Ming Chiao Tung University, Taipei, Taiwan in 2018. His research interests include image processing, deep learning, and machine learning.

Chiu-Jung Huang is a research assistant at the Institute of Brain Science, National Yang Ming Chiao Tung University, Taiwan. In 2013, she received her MS degree from the Institute of Brain Science, National Yang Ming Chiao Tung University, Taipei, Taiwan.

Chien-Yao Wang is currently a postdoctoral fellow with the Institute of Information Science, Academia Sinica, Taiwan. He received his BS degree in computer science and information engineering from National Central University, Zhongli, Taiwan, in 2013, and the PhD from National Central University, Zhongli, Taiwan, in 2017. His research interests include signal processing, deep learning, and machine learning. He is an honorary member of Phi Tau Phi Scholastic Honor Society.

Mei-Lien Pan is an assistant professor of the Information Technology Service Center at National Yang Ming Chiao Tung University, Taiwan since August 2021. She received her MS and PhD degrees at the Institute of Public Health of National Yang Ming University, Taiwan in 2000 and 2012. Her research interests include medical data science, data privacy, disease simulation model, public informatics, and evaluation in medical information systems.

Da-Wei Wang received his BS and a MS degrees in information engineering and computer science from National Taiwan University in 1985 and 1987, respectively. He received his PhD in computer science from Yale University in 1992. Since December 1992, he joined the institute as an assistant research fellow. He is currently a research fellow and deputy director in Institute of Information Science.

Hong-Yuan Mark Liao received his PhD from Northwestern University, Evanston, Illinois, in 1990. He joined the Institute of Information Science, Academia Sinica, Taiwan in 1991. He received the Young Investigators’ Award from Academia Sinica in 1998; the Distinguished Research Award from the National Science Council in 2003, 2010 and 2013; the Academia Sinica Investigator Award in 2010; the TECO Award from the TECO Foundation in 2016; and the 64th Academic Award from the Ministry of Education in 2020. His professional activities include Editorial Board Member, IEEE Signal Processing Magazine (2010–2013); Associate Editor, IEEE Transactions on Image Processing (2009–2013), IEEE Transactions on Information Forensics and Security (2009–2012), IEEE Transactions on Multimedia (1998–2001), ACM Computing Surveys (2018–2021). He is now a senior associate editor of ACM Computing Surveys (2021–present). He has been a fellow of the IEEE since 2013.

Yong-Sheng Chen received a BS degree in computer and information science from National Chiao Tung University, Hsinchu, Taiwan, in 1993, and an MS and a PhD degrees in computer science and information engineering from National Taiwan University, Taipei, Taiwan, in 1995 and 2001, respectively. He is currently a professor in the Department of Computer Science, National Yang Ming Chiao Tung University, Hsinchu, Taiwan. His research interests include biomedical signal processing, medical image processing, and computer vision. He received the Best Paper Award in the 2008 Robot Vision Workshop and the Best Annual Paper Award of the Journal of Medical and Biological Engineering, 2008.

Chieh-Hsiung Kuan graduated from National Taiwan University, Taiwan in 1985, and received his PhD in electrical engineering from Princeton University, in 1994. He became a professor at the Department of Electrical Engineering in National Taiwan University, Taiwan in 2002. His researches are focused on optoelectronic device, nano-electronics, and e-beam lithography technology. He is also deeply involved in retinal disease self-therapy from energy point of view.

Shih-Yen Lin is currently a postdoctoral research fellow in the Department of Computer Science, National Yang Ming Chiao Tung University. He received his BS and PhD degrees in computer and information science from National Chiao Tung University, Hsinchu, Taiwan in 2013 and 2020, respectively. His research interests include biomedical engineering, medical image analysis, and deep neural networks.

Li-Fen Chen received her BS degree in computer science from National Chiao Tung University, Hsing-Chu, Taiwan, in 1993. Her research interests include image processing, pattern recognition, computer vision, and wavelets.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Po-Kang Lin, Yu-Hsien Chiu, Chiu-Jung Huang, Chien-Yao Wang, Mei-Lien Pan, Da-Wei Wang, Hong-Yuan Mark Liao, Yong-Sheng Chen, Chieh-Hsiung Kuan, Shih-Yen Lin, and Li-Fen Chen "PADAr: physician-oriented artificial intelligence-facilitating diagnosis aid for retinal diseases," Journal of Medical Imaging 9(4), 044501 (25 July 2022). https://doi.org/10.1117/1.JMI.9.4.044501
Received: 30 August 2021; Accepted: 1 July 2022; Published: 25 July 2022
Advertisement
Advertisement
KEYWORDS
Performance modeling

Curium

Binary data

Data modeling

Visualization

Eye models

Retina

Back to Top