Abstract
Background
Deep learning systems have improved performance of devices through more accurate object detection in a significant number of areas, for medical aid in general, and also for navigational aids for the visually impaired. Systems addressing different needs are available, and many manage effectively the detection of static obstacles.
Purpose
This research provides a review of deep learning systems used with navigational tools for the visually Impaired and a framework for guidance for future research.
Methods
We compare current deep learning systems used with navigational tools for the visually impaired and compile a taxonomy of indispensable features for systems.
Results
Challenges to detection. Our taxonomy of improved navigational systems shows that it is sufficiently robust to be generally applied.
Conclusion
This critical analysis is, to the best of our knowledge, the first of its kind and will provide a much-needed overview of the field.
Deep learning systems can provide lost cost solutions for the visually impaired.
Of these, convolutional neural networks (CNN) and fully convolutional neural networks (FCN) show great promise in terms of the development of multifunctional technology for the visually impaired (i.e., being less specific task oriented).
CNN have also potential for overcoming challenges caused by moving and occluded objects.
This work has also highlighted a need for greater emphasis on feedback to the visually impaired which for many technologies is limited.
Implication for Rehabilitation
Disclosure statement
No potential conflict of interest was reported by the authors.
Table 1. Common system components and their main attributes with instances of their use.
Table 2. Deep learning architecture used for the object detection and recognition process.