Artificial Intelligence-Based Classification of Diabetic Peripheral Neuropathy From Corneal Confocal Microscopy Images

Diabetes

Diabetic peripheral neuropathy (DPN) is characterized by pain and sensory loss, affecting approximately 50% of patients (1). Early identification and risk factor management are key to limiting progression of DPN. In contrast to retinopathy (retinal fundus imaging) and nephropathy (microalbuminuria) with early disease detection, the 10-g monofilament identifies advanced DPN. Corneal confocal microscopy (CCM) is an ophthalmic imaging technique that identifies subclinical corneal nerve loss, which predicts incident DPN (2) and has good diagnostic utility for DPN (3). It also identifies corneal nerve regeneration prior to improvement in symptoms and nerve conduction studies after simultaneous pancreas and kidney transplantation (4). CCM studies have primarily used manual corneal nerve analysis (CCMetrics), which, although highly reliable, is time-consuming with limited scalability.

Here, we combine a deep learning (DL) algorithm for fully automated quantification of corneal nerves in CCM images along with an adaptive neuro-fuzzy inference system (ANFIS) to rapidly differentiate patients with (DPN+) and without (DPN−) neuropathy and healthy control subjects. Participants with type 1 diabetes (n = 87) and control subjects (n = 21) underwent detailed assessment of neuropathy (Table 1). Based on the Toronto criteria, which combine symptoms, signs, and abnormal nerve conduction, patients were subdivided into DPN+ (29%) and DPN− (71%) groups (Table 1). Participants underwent CCM, and 6–8 central corneal nerve images/subject were quantified using our established methodology (5). DL performance was assessed against ground truth derived from gold-standard 1) manual CCMetrics and 2) automated (ACCMetrics) methods to quantify corneal nerve fiber length (CNFL).

Table 1

Demographic and clinical characteristics based on Toronto classification of DPN

The DL algorithm uses a U-Net–based convolutional neural network, which requires smaller training sets for more precise segmentation. The algorithm was initially trained on 25% (n = 174 images) of the data set (affecting approximately 50% epochs/0.0001 learning rate) for 30 h (Intel Core i3–6100) and then validated on the remaining images (n = 534). To optimize quality, input training images were cropped to 256 × 256 pixels, binarized, and skeletonized, resulting in a segmented output image, where nerve pixels corresponded to a value of 1 and nonnerve pixels to 0.

DL-estimated CNFL was comparable to CCMetrics (1,326 ± 459 vs. 1,269 ± 444 pixels, P > 0.05) and significantly outperformed ACCMetrics (1,036 ± 385 pixels, P < 0.0001), with higher pixel detection sensitivity (85% vs. 70%) and lower false negative rate (15% vs. 30%) than ground truth. The intraclass correlation coefficient indicated excellent reproducibility for DL segmentation (0.98, P = 0.0001) and lower but high reproducibility for ACCMetrics (0.85, P = 0.0001).

ANFIS harnesses the learning ability of convolutional neural networks and the logic-based reasoning of fuzzy systems and was deployed to classify participants into DPN+, DPN−, or healthy control subjects (Fig. 1A and B). A multiclass model for ANFIS was constructed using “one versus all” and “one versus one” classifiers. The first classifier was trained to differentiate control (labeled 0) from DPN− (labeled 1) and DPN+ (labeled 2) participants using a one-versus-all approach. The second classifier was trained to differentiate DPN− from DPN+ participants using a one-versus-one approach. Βoth classifiers were trained at 20 epochs (step size, 0.1; step increase size, 1.01). Based on CCM images, ANFIS classified 43% of participants as DPN+ with excellent reliability (Cohen κ = 0.86, P < 0.0001). Receiver operating characteristic curve analysis showed the following capacity for discriminating 1) DPN− from control subjects: area under the curve (AUC) = 0.86 (95% CI 0.77–0.94, P < 0.0001) with 84% sensitivity/71% specificity; 2) DPN− from DPN+: AUC = 0.95 (95% CI 0.91–0.99, P < 0.0001) with 92% sensitivity/80% specificity; and 3) control subjects from DPN+: AUC = 1.0 (95% CI 0.99–1.0, P < 0.0001) with 100% sensitivity/95% specificity (Fig. 1C and D). Model size analysis showed that the combined segmentation (DL)–classification (ANFIS) system occupied 38 megabytes.

Figure 1
Figure 1

A: The hierarchical ANFIS classification model. CCM images from the training subset are fed into an initial neuro-fuzzy inference system (INFIS), and the results are evaluated on the first validation data set. The best-performing cutoff points are determined by subsequent fuzzy inference systems (FIS), and performance of the final system is evaluated on the second validation set. B: The hierarchical ANFIS prediction workflow is then applied to classify DPN based on the extent of CNFL loss. C: Representative CCM images from a healthy control and DPN− and DPN+ patients. D: Receiver operating characteristic curve analysis for identification of DPN by ANFIS based on DL segmented CCM images.

We show that automated DL image segmentation has excellent agreement with manual expert analysis for the quantification of corneal nerves. This is important, as rapid and accurate automated quantification is key for successful deployment of CCM in large-scale, multicenter studies and clinical trials of disease-modifying therapies in DPN (4). Furthermore, we show that DPN classification with ANFIS was highly accurate for discriminating control from DPN− and particularly DPN+ subjects, surpassing previous diagnostic outcomes (3). Indeed, the ANFIS model identified corneal nerve loss in patients deemed to have no DPN based on the Toronto criteria, which relies on nerve conduction and misses early small nerve fiber involvement (5). Patients with DPN were older, and given that increasing age is associated with corneal nerve loss, that may have contributed to the outcomes of this study. In conclusion, this artificial intelligence model could discriminate patients with DPN from control subjects with almost perfect diagnostic outcomes, indicating the considerable potential of CCM in screening for DPN.

Article Information

Acknowledgments. M. Tavakoli (University of Exeter) undertook corneal confocal microscopy and H. Fadavi (Imperial College) undertook neuropathy assessments and quantitative sensory testing in a portion of study participants.

Funding. This research was funded by awards from the National Institutes of Health (R105991) and the JDRF (27-2008-362).

Duality of Interest. No potential conflicts of interest relevant to this article were reported.

Author Contributions. T.S. and I.N.P. prepared the first draft. T.S., I.N.P., M.F., G.P., O.A., U.A., and Z.R.M. acquired and analyzed data. All authors assisted in interpretation of results. S.K., N.E., R.A.M., and U.A.Q. critically revised the manuscript and approved the final version of the article. R.A.M. and U.A.Q. conceptualized and designed the study and supervised the analysis. R.A.M. and U.A.Q. are the guarantors of this work and, as such, had full access to all the data in the study and take responsibility for the integrity of the data and the accuracy of the data analysis.

  • Received August 13, 2020.
  • Accepted April 25, 2021.

Products You May Like

Leave a Reply

Your email address will not be published. Required fields are marked *