Enabling Immersive Indoor Navigation and Control through Computer Vision and Augmented Reality | ARoBotX (Final Year Research Project)

Enabling Immersive Indoor Navigation and Control through Computer Vision and Augmented Reality | ARoBotX (Final Year Research Project)

Created
Jun 25, 2025 09:58 AM
Technologies
Unity Engine (C#) | Vuforia Engine | Augmented Reality | AWS IoT | Kinect V2 | Yolo V8 | OWL-VIT | Python | OpenCV | C++ | IPC (Inter-Process Communication) Named Pipes | CMake | MQTT
Tags
Augmented Reality
Mobile Robotics
Computer Vision
Cloud Services
Network Protocols
Status
Done
This project integrates computer vision, Augmented Reality (AR), and cloud-based communication to create a real-time 3D map of the robot’s surroundings. The outcome is an engaging and immersive control experience for users in indoor robot monitoring and controlling. Furthermore, combining an image semantic segmentation model with the Microsoft Kinect V2 depth camera introduces a novel method for achieving 1 cm accuracy in estimating the distance to a particular object. This precise depth estimation is a significant benefit for tasks that require accurate object identification, segmentation, and depth information to generate a virtual representation of the real-world environment of a robot. We have enabled remote communication with the robot’s environment by incorporating AWS cloud services, guaranteeing minimal delay. The commercial viability of this project can be identified in a wide range of applications that find the system useful due to its easy-to-use interface, remote access capability, and precise object rendering capabilities. Examples of such applications include surgical robotics, instructional robotics platforms in classrooms, and automated guided vehicles in warehouses.