Real-time indoor tracking for augmented reality using computer vision technique
Abstract
In recent times, there has been an increase in the stability and integration of augmented reality (AR) technology in everyday applications. AR relies on tracking techniques to capture the characteristics of the surrounding environment. Tracking falls into two categories: outdoor and indoor. While outdoor tracking predominantly relies on the global positioning system (GPS), it is performance indoors is hindered by imprecise GPS signals. Indoor tracking offers a solution for navigating complex indoor environments. This paper introduces an indoor tracking system that combines smartphone sensor data and computer vision using the oriented features from accelerated and segments test and rotated binary robust independent elementary features (ORB) algorithm for feature extraction, along with brute force match (BFM) and k-nearest neighbor (KNN) for matching. This approach outperforms previous systems, offering efficient navigation without relying on pre-existing maps. The system uses the A* algorithm to find the shortest path and cloud computing for data storage. Experimental results demonstrate an impressive 99% average accuracy within a 7-10 cm error range, even in scenarios with varying distances. Moreover, all users successfully reached their destinations during the experiments. This innovative model presents a promising advancement in indoor tracking, enhancing the accuracy and effectiveness of navigation in complex indoor spaces
Keywords
Augmented reality; Camera tracking; Global positioning system; Indoor tracking; Navigation ORB algorithm; Smartphone sensors
Full Text:
PDFDOI: http://doi.org/10.11591/ijai.v13.i2.pp1845-1857
Refbacks
- There are currently no refbacks.
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
IAES International Journal of Artificial Intelligence (IJ-AI)
ISSN/e-ISSN 2089-4872/2252-8938
This journal is published by the Institute of Advanced Engineering and Science (IAES) in collaboration with Intelektual Pustaka Media Utama (IPMU).