Publication:
Step Length Estimation Using Sensor Fusion

dc.contributor.authorsSevinc H., Ayvaz U., Ozlem K., Elmoughni H., Atalay A., Atalay O., Ince G.
dc.date.accessioned2022-03-15T02:15:22Z
dc.date.accessioned2026-01-11T15:41:14Z
dc.date.available2022-03-15T02:15:22Z
dc.date.issued2020
dc.description.abstractOne of the main challenges of navigation systems is the inability of orientation and insufficient localization accuracy in indoor spaces. There are situations where navigation is required to function indoors with high accuracy. One such example is the task of safely guiding visually impaired people from one place to another indoors. In this study, to increase localization performance indoors, a novel method was proposed that estimates the step length of the visually impaired person using machine learning models. Thereby, once the initial position of the person is known, it is possible to predict their new position by measuring the length of their steps. The step length estimation system was trained using the data from three separate devices; capacitive bend sensors, a smart phone, and WeWALK, a smartcane developed to assist visually impaired people. Out of the various machine learning models used, the best result obtained using the K Nearest Neighbor model, with a score of 0.945 R2. These results support that indoor navigation will be possible through step length estimation. © 2020 IEEE.
dc.identifier.doi10.1109/FLEPS49123.2020.9239441
dc.identifier.isbn9781728152783
dc.identifier.urihttps://hdl.handle.net/11424/248117
dc.language.isoeng
dc.publisherInstitute of Electrical and Electronics Engineers Inc.
dc.relation.ispartofFLEPS 2020 - IEEE International Conference on Flexible and Printable Sensors and Systems
dc.rightsinfo:eu-repo/semantics/closedAccess
dc.titleStep Length Estimation Using Sensor Fusion
dc.typeconferenceObject
dspace.entity.typePublication
oaire.citation.titleFLEPS 2020 - IEEE International Conference on Flexible and Printable Sensors and Systems

Files