The ability to navigate unstructured environments is an essential task for intelligent systems.Autonomous navigation by ground vehicles requires developing an internal representation of space, trained by recognizable landmarks, robust visual processing, computer vision and image processing.A mobile robot needs a platform enabling it to operate in an environment autonomously, recognize the objects, and avoid Shelf Bracket obstacles in its path.In this study, an open-source ground robot called SROBO was designed to accurately identify its position and navigate certain areas using a deep convolutional neural network and transfer learning.
The framework uses an RGB-D MYNTEYE camera, a 2D laser scanner and inertial Bike Parts - Wheel Parts measurement units (IMU) operating through an embedded system capable of deep learning.The real-time decision-making process and experiments were conducted while the onboard signal processing and image capturing system enabled continuous information analysis.State-of-the-art Real-Time Graph-Based SLAM (RTAB-Map) was adopted to create a map of indoor environments while benefiting from deep convolutional neural network (Deep-CNN) capability.Enforcing Deep-CNN improved the performance quality of the RTAB-Map SLAM.
The proposed setting equipped the robot with more insight into its surroundings.The robustness of the SROBO increased by 35% using the proposed system compared to the conventional RTAB-Map SLAM.