Abstract
The visual impaired people requires being as much independent as possible to navigate indoor environments. The mobility of this kind of people is difficult and several times requires a companion. Therefore there are several methods for object detection and recognition based on computer vision techniques. This work proposes the use of these methods in the context of accessibility for people with vision problems.
You have full access to this open access chapter, Download conference paper PDF
1 Introduction
According to WHO (World Health Organization) [1], in 2010, 39.365 million of people suffered blindness and there were 264,024 million people with low vision in area of the Americas (Central and South America) 3,211 million (8%) were blind and 23,401 million (9.5%) with low vision. Vision is one of the main ways humans develop most of their daily activities, due to that, people with affected eyesight suffer various difficulties dealing with everyday activities. Although governments and public agencies in different cities have implemented systems that include Braille and RFID (Radio Frequency Identification) labels in main sites to orient blind people, these methods are not widespread and not effective in all cases. As a result in most cases blind people depend on a companion to move towards their work, home or any indoor environment. Relying other methods in many cases is not feasible or practicable, for lack of time or resources. For this reason these people seek to be as independent as possible in order to integrate and become part of society. There are several applications that help navigation and mobility of people with visual disabilities [2,3,4,5], that can be divided into navigation support indoor and outdoor spaces.
Additionally, there are other methods based on the generation of mental maps for navigation using step to step actions in a virtual environment using portable devices [6].
2 Developed Process
A system for support of people with visual disabilities must consider a set of different kinds of sensors, which provides as much as possible information about the environment. Such as a unit of processing (i.e. a companion computer), a method for feedback (tactile or auditory) and a correctly developed interface that receives the user commands and manage the inputs of sensors to return the required information to the user. In this work, a part of the whole system is presented, this part is composed of a visual system based on computer vision techniques. The computer vision techniques are in charge of detect objects of interest in the environment as signs that indicate which is the place where the user is located. For this purpose a machine learning method is used for object detection using a dataset obtained in first instance from a camera. In this stage a ROI (region of interest) is obtained. After that, a color based segmentation is used to improve the detection. Since the vision system obtain images with perspective distortion, a homograpy is computed in order to get a rectangular sign. Then, an implementation of OCR (optical character recognition) using neural networks is employed to obtain the data of the sign. After that, the information is processed in order to give a response to the user using a text to speech component.
3 Methodology
In the first stage a set of 50 images with resolution of 8 megapixels were obtained. These images allowed to test the system. The dataset different features that includes, images taken from different point of view, background and sign color as can be shown in Fig. 1. For the implementation of the system, the OpenCV library was used.
3.1 Object Detection
An object method based on HOG descriptors [7], that are computed by a sliding window is implemented for sign detection; the classifier is a support vector machine as is shown in Fig. 2.
3.2 Irregular Quadrilateral Recognition
After the sign detection, it is necessary to find the corners of the sign in order to apply a homography. This is done by computing the closest distance from the detected corners to the frontiers of a bounding rectangle as it is shown in Fig. 3.
The corner detection is necessary to specify a set of points that permits to create homography, the results are shown in Fig. 4.
3.3 OCR Stage
It was necessary to use a database of characters in order to train a classifier that is able to recognize the characters that appears in the rooms signs. Before the optical character recognition (OCR) it is necessary to process the image in order to improve the contrast and binarize, obtaining a simplified input to the system as is shown in Fig. 5.
4 Results
After the object detection stage, the system is able to isolate the image of the sign using an image processing method, due to corner detection method a segmented rectangular sign is produced. A method for perspective correction based on homography is employed with good results. The proposed process is shown in Fig. 6.
5 Conclusions and Future Work
The actual system allows to distinguish the area where the signs are located inside the walls of the building. Moreover, it permits to detect the characters included in the sign.
In addition, a connection with a text to speech system is able to communicate the information to the blind user in an auditory feedback.
The study and development of an interaction method that improve the usability of the system is proposed as a future work.
Also the semantic relations that can be present in internal environments could be explored as a future work in order to improve the system.
References
Shakespeare, T.: World report on disability. World Health Organization (2011). Open Access Content
Chumkamon, S., Tuvaphanthaphiphat, P., Keeratiwintakorn, P.: A blind navigation system using RFID for indoor environments. In: 2008 5th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, vol. 2, pp. 765–768 (2008)
Wang, H.C., Katzschmann, R.K., Teng, S., Araki, B., Giarré, L., Rus, D.: Enabling independent navigation for visually impaired people through a wearable vision-based feedback system. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 6533–6540. IEEE (2017)
Faria, J., Lopes, S., Fernandes, H., Martins, P., Barroso, J.: Electronic white cane for blind people navigation assistance. In: World Automation Congress (WAC), pp. 1–7. IEEE (2010)
D’Atri, E., Medaglia, C.M., Serbanati, A., Ceipidor, U.B., Panizzi, E., D’Atri, A.: A system to aid blind people in the mobility: a usability test and its results. In: ICONS 2007. Second International Conference on Systems, p. 35. IEEE (2007)
Guerreiro, J., Ahmetovic, D., Kitani, K.M., Asakawa, C.: Virtual navigation for blind people: building sequential representations of the real-world. In: Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility. ASSETS 2017, pp. 280–289. ACM, New York (2017)
Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: Computer Vision and Pattern Recognition, CVPR (2005)
Acknowledgments
This work was supported by the Vicerrectoria de Investigaciones Universidad Militar Nueva Granada through the PIC-ING-2681 project.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Mendieta, C., Ramos, C., Cerón, A. (2018). Towards the Development of a System for the Support of People with Visual Disabilities Using Computer Vision. In: Stephanidis, C. (eds) HCI International 2018 – Posters' Extended Abstracts. HCI 2018. Communications in Computer and Information Science, vol 851. Springer, Cham. https://doi.org/10.1007/978-3-319-92279-9_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-92279-9_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-92278-2
Online ISBN: 978-3-319-92279-9
eBook Packages: Computer ScienceComputer Science (R0)