A Novel Algorithm of Distance Calculation Based-on Grid-Edge-Depth-Map and Gyroscope for Visually-Impaired
This paper presented a new algorithm for determining the distance of an object in front of a stereo camera placed on a helmet. By using a stereo camera with a Sum of Absolute Difference with a Sobel edge detector, our previous Grid-Edge-Depth map algorithm could calculate the objects’ distance up to 500 cm. The problem started when a vision disability person used the device with an unfixed stereo camera angle. The unspecified angle caused by the helmet’s movement influenced the distance calculation result. This novel process started with calculating the distance from a Grid-Edge-Depth map considering unfixed angle data of the x-axis from a gyroscope sensor placed on the stereo camera using the trigonometry formula. The angle data used was the x-axis data. The distance measurement results by the system were then computed based on the unfixed angle compared to the actual distance. The test was carried out with three scenarios which required the user to stand at a distance of 100 cm, 125 cm, and 150 cm from a table, chair, or wall, with 30 tests for each scenario. The test results showed an average accuracy of 96.05% with three experimental scenarios, which meant that this machine was feasible to implement.
on Instrumentation and Measurement, vol. 70, 2021.
 E. J. Anthony and R. A. Kusnadi, “Computer Vision for Supporting Visually Impaired People: A Systematic Review,” Engineering,
MAthematics and Computer Science (EMACS) Journal, vol. 3, no. 2, pp. 65–71, 2021.
 S. Caraiman, O. Zvoristeanu, A. Burlacu, and P. Herghelegiu, “Stereo Vision Based Sensory Substitution for the Visually
Impaired,” Sensors, vol. 19, no. 2771, pp. 1–18, 2019.
 J. K. Mahendran, D. T. Barry, A. K. Nivedha, and S. M. Bhandarkar, “Computer Vision-Based Assistance System for the
Visually Impaired Using Mobile Edge Artificial Intelligence,” in IEEE Computer Society Conference on Computer Vision and
Pattern Recognition Workshops. IEEE Explore, 2021, pp. 2418–2427.
 J. Supriyanto and A. Fadlil, “Qualitative Testing of Earthquake Information Applications in The Form of Sound for the Visually
Impaired,” Matrik, vol. 21, no. 3, pp. 619–628, 2022.
 A. Sen, K. Sen, and J. Das, “Ultrasonic Blind Stick for Completely Blind People to Avoid Any Kind of Obstacles,” in Proceedings
of IEEE Sensors, vol. October. IEEE, 2018, pp. 1–4.
 A. A. Di´az-Toro, S. E. Campan˜a-Bastidas, and E. F. Caicedo-Bravo, “Vision-Based System for Assisting Blind People to
Wander Unknown Environments in a Safe Way,” Journal of Sensors, vol. 2021, pp. 1–18, 2021.
 N. Loganathan, K. Lakshmi, N. Chandrasekaran, S. R. Cibisakaravarthi, R. H. Priyanga, and K. H. Varthini, “Smart Stick for
Blind People,” in 2020 6th International Conference on Advanced Computing and Communication Systems, ICACCS 2020,
2020, pp. 65–67.
 S. Budilaksono, B. Bertino, I. G. A. Suwartane, A. Rosadi, M. A. Suwarno, S.W. Purtiningrum, Y. Sari, E. Suhandono, E. M. S.
Sakti, D. Gustina, and A. A. Riyadi, “Designing an Ultrasonic Sensor Stick Prototype for Blind People,” in 1st Bukittinggi
International Conference on Education IOP Conf. Series: Journal of Physics: Conf. Series, vol. 1471, no. 1. IOP Publioshing,
2020, pp. 1–7.
 R. Mehmood, “LidSonic V2.0: A LiDAR and Deep-Learning-Based Green Assistive Edge Device to Enhance Mobility for the
Visually Impaired,” Sensors, vol. 22, no. 19, pp. 1–55, 2022.
 J. Tong, J. Huang, V. Khou, J. Martin, M. Kalloniatis, and A. Ly, “Topical Review: Assessment of Binocular Sensory Processes
in Low Vision,” Optometry and Vision Science, vol. 98, no. 4, pp. 310–325, 2021.
 M. D. Messaoudi, B. A. J. Menelas, and H. Mcheick, “Review of Navigation Assistive Tools and Technologies for the Visually
Impaired,” Sensors, vol. 22, no. 20, 2022.
 B. Kuriakose, R. Shrestha, and F. E. Sandnes, “Tools and Technologies for Blind and Visually Impaired Navigation Support: A
Review,” IETE Technical Review (Institution of Electronics and Telecommunication Engineers, India), vol. 39, no. 1, pp. 3–18,
 M. Mashiata, T. Ali, P. Das, Z. Tasneem, F. R. Badal, S. K. Sarker, M. Hasan, S. H. Abhi, R. Islam, F. Ali, H. Ahamed,
M. Islam, and S. K. Das, “Towards Assisting Visually Impaired Individuals: A Review on Current Status and Future Prospects,”
Biosensors and Bioelectronics: X, vol. 12, no. August, p. 100265, 2022.
 B. Rahmani, A. Harjoko, T. K. Priyambodo, and H. Aprilianto, “Early Model of Vision-Based Obstacle Mapping Utilizing
Grid-Edge-Depth Map,” International Journal of Engineering and Advanced Technology (IJEAT), vol. 9, no. 2, pp. 4519–4523,
 A. Jalal, A. Ahmed, A. A. Rafique, and K. Kim, “Scene Semantic Recognition Based on Modified Fuzzy C-Mean and Maximum
Entropy Using Object-to-Object Relations,” IEEE Access, vol. 9, 2021.
 A. A. Diaz, S. E. Campana, and E. F. Caicedo, “Methodology to Build a Wearable System for Assisting Blind People in
Purposeful Navigation,” in Proceedings - 3rd International Conference on Information and Computer Technologies, ICICT
2020, vol. 1, no. 1, 2020, pp. 205–212.
 J. Ganesan, A. T. Azar, S. Alsenan, N. A. Kamal, B. Qureshi, and A. E. Hassanien, “Deep Learning Reader for Visually
Impaired,” Electronics, vol. 11, no. 2225, pp. 1–22, 2022.
 P. Chandankhede, “Visually Impaired Aid Using Computer Vision to Read the Obstacles,” Journal of Algebraics Statistics,
vol. 13, no. 3, pp. 4467–4481, 2022.
 I. P. Adi, H. Kusuma, and M. Attamimi, “Blind People Guidance System Using Stereo Camera,” in 2019 International Seminar
on Intelligent Technology and Its Applications (ISITIA), 2019, pp. 298–303.
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.