954
Views
4
CrossRef citations to date
0
Altmetric
Review

Machine Learning and Deep Learning Techniques for Optic Disc and Cup Segmentation – A Review

, ORCID Icon, , ORCID Icon, ORCID Icon, & ORCID Icon show all
Pages 747-764 | Published online: 11 Mar 2022

Figures & data

Table 1 Summary of Glaucoma Prevalence Studies Around the World

Table 2 Current Available Public Datasets with Labelling and Manual Annotation for Glaucoma Fundus Images

Table 3 Comparison of Optic Cup (OC) and Optic Disc (OD) Segmentation Methods on REFUGE Test Set Using Different Metrics

Table 4 Deep Learning and Machine Learning Methods for Optic Cup and Disc Segmentation

Figure 1 Tan et al proposed CNN architecture.

Notes: Reprinted from: Tan JH, Acharya UR, Bhandary SV, Chua KC, Sivaprasad S. Segmentation of optic disc, fovea and retinal vasculature using a single convolutional neural network. J Comput Sci. 2017;20:70–79.Citation94 Copyright 2017, with permission from Elsevier.
Figure 1 Tan et al proposed CNN architecture.

Figure 2 Spatial pyramid pooling layer: pooling features extracted using different window sizes on the feature maps.

Notes: © 2015 IEEE. Reprinted, with permission, from: He K, Zhang X, Ren S, Sun J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell. 2015;37(9):1904–1916.Citation98.
Figure 2 Spatial pyramid pooling layer: pooling features extracted using different window sizes on the feature maps.

Figure 3 U-net architecture (example for 32×32 pixels in the lowest resolution). Each blue box corresponds to a multi-channel feature map. The number of channels is denoted on top of the box. The x-y-size is provided at the lower left edge of the box. White boxes represent copied feature maps. The arrows denote the different operation.

Notes: Reprinted by permission from Springer Nature from: Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells W, Frangi A (eds). Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. MICCAI 2015. Lecture Notes in Computer Science, vol 9351. Springer, Cham; 2015:234–241.Citation102 Copyright © Springer International Publishing Switzerland 2015. Available from: https://link.springer.com/book/10.1007/978-3-319-24574-4.
Figure 3 U-net architecture (example for 32×32 pixels in the lowest resolution). Each blue box corresponds to a multi-channel feature map. The number of channels is denoted on top of the box. The x-y-size is provided at the lower left edge of the box. White boxes represent copied feature maps. The arrows denote the different operation.

Figure 4 GAN general architecture consists of Generator (G) which output a synthetic sample given a noise variable input and a Discriminator (D) which estimate the probability of a given sample coming from real dataset. Both components are built based on neural network.

Notes: © 2018 IEEE. Reprinted, with permission, from: Creswell A, White T, Dumoulin V, Arulkumaran K, Sengupta B, Bharath AA. Generative adversarial networks: An overview. IEEE Signal Process Mag. 2018;35(1):53–65.Citation108.
Figure 4 GAN general architecture consists of Generator (G) which output a synthetic sample given a noise variable input and a Discriminator (D) which estimate the probability of a given sample coming from real dataset. Both components are built based on neural network.