Vision and Simultaneous Localization and Mapping based Autonomous Indoor Mobile Robot

DOI : 10.17577/IJERTV10IS050354

Download Full-Text PDF Cite this Publication

Text Only Version

Vision and Simultaneous Localization and Mapping based Autonomous Indoor Mobile Robot

Akshay Baburao Mane

Dept. of Electronics and Telecommunication Padmabhushan Vasantdada Patil Institute of Technology, Bavdhan Pune, India

Prof. S. A. Gandhi

Dept. of Electronics and Telecommunication Padmabhushan Vasantdada Patil Institute of Technology,

Bavdhan Pune, India

Abstract The Simultaneous Localization and Mapping (SLAM) has been one of the most studied subjects in Robotics. It is a fundamental process in mobile robotics, since it is responsible for building maps and at same time estimate the robot position in the environment. In this paper, a real-time implementation of simultaneous localization and mapping (SLAM) is described based on a mobile robot platform, which consists of two driving wheels, one caster wheel, an ultrasonic sensor and a Kinect sensor camera. The robot used for this purpose is differential drive robot. The main algorithm used is extended Kalman Filter (EKF) which is combined with Visual SLAM to handle data association problems and to map the trajectory with more accuracy. A mathematical equations of state space model and observation model is used to represents the motion of a robotic system from one timestep to the next and to infer what the sensor measurement would be at that timestep, respectively. The robots built-in ability allows it to map the interiors of unknown environment and run autonomously using its self ruling and pilot feature. Robot Operating System is used for data processing and analysis of mobile robot platform. It is intended to implement the proposed SLAM by Extended Kalman Filter and VSLAM using low cost, energy efficient Raspberry PI as embedded system, IMU, laptop and ROS to verify the utility of proposed methods as a indoor wireless mobile robot.

Keywords Extended Kalman filter, State space model, Observation model, Robot operating system, differential drive robot

mentioned as feature-based map, and this problem is called CML (Concurrent Mapping and Localization), but now more people would like to use SLAM (Simultaneous Localization and Mapping) [5]. Since then, a probabilistic approach [3] [4]

  1. has become a standard way of solving the SLAM problem.

    Kalman filters are an efficient algorithm and widely introduced to solve the SLAM [6]. However, the regular Kalman filter only can cope with linear problem, extended Kalman filter is a nonlinear version of the Kalman filter which linearizes about an estimate of the current mean and covariance [7].Particle filter algorithm, also accomplish the target successfully [11]. However, in the presented paper the focus is given to the experimental results of an EKF-based localization of a mobile robot, which investigates a scalable algorithm for the simultaneous mapping and localization (SLAM) problem, and evaluates it in the context of unknown environment.

    This paper is organized as follows. In Section II proposed robot system is described. Section III presents how the feature extraction and data association done with extended Kalman filter. Section IV describes Differential drive concept plan and mapping. Section V describes experimental results of robot in an indoor environment. The paper is concluded in Section VI.

    1. PROPOSED ROBOT SYSTEM

      1. INTRODUCTION

        Wireless Mobile robots are very promising application of robotics. They should help people to deal with various tasks. The SLAM problem is the problem of acquiring a map of an unknown environment with a moving robot, while simultaneously localizing the robot relative to this map [1][2]. To solve this problem, there are several simultaneous localization and mapping (SLAM) approaches, such as Kalman filterbased algorithm [1][5][6], Particle filter[11], Graph based SLAM[12], G-mapping algorithm[12].

        The idea of solving both of the mapping and localization problems simultaneously, firstly introduced by Smith et al. (1990) [1]. They developed a probabilistic method to indicate the spatial relationship between landmarks in an environment when robot is estimating its pose, currently. The map is represented as a set of landmark position and a covariance matrix is used to present the uncertainty of either the landmark or the robot's pose. This kind of map today is usually

        The main task of proposed system is to collect indoor environment information and build environment map simultaneously without collisions in the whole process. Our developed Indoor mobile robot platform is illustrated in Fig. 1. It equips with sensors, Kinect Camera to grasp external information. The two wheels are driven by servo motors. Ultrasonic sensors are also called ping sensors, and are mainly used to measure the robot's distance from an object. The main application of ping sensors is to avoid obstacles. Proposed system includes Robot Operating System (ROS) to build map. ROS is a framework that is widely used in robotics. It provides a set of tools, libraries and drivers in order to help develop robot applications with hardware abstraction. In ROS, we are

        Fig. 1 Block Diagram Indoor Mobile robot

        going to use two sensors of Kinect Camera namely the RGB camera and the depth sensor for vision and robotics tasks.

    2. SIMULTANEOUS LOCALIZATION AND MAPPING Simultaneous localization and mapping (SLAM) is a

      process which estimates the robots pose and landmarks position at the same time, namely location and mapping. The fundamental step is to obtain external information to search proper features (landmarks). When navigating autonomously through an unknown environment, the robot keeps a track of its estimated (probabilistic) pose with reference to a global static frame usually located at its start point or a previous pose frame. The heart of our SLAM process is correlating the mean and covariance matrix of system state through EKF technique.

      1. Extended Kalman Filter Algorithm for Localization

        In estimation theory, the extended Kalman filter (EKF) is the nonlinear version of the Kalman filter which linearizes about an estimate of the current mean and covariance. In order to understand what an EKF is, we should know what a state space model and an observation model are.

        The State Space Model: The state space model for our differential drive robot is as follows:

        where vk-1 is the linear velocity of the robot in the robots reference frame and k-1 is the angular velocity in the robots reference frame.

        Observation Model: An observation model is a mathematical equation that represents a vector of predicted sensor measurements y at time t as a function of the state of a robotic system x at time k, plus some sensor at time k, denoted by vector k. In vector format, the observation model is:

        Formal equation

        Where, k = current time, y vector = n predicted sensor observations at time k, n = number of sensor observations, H

        = measurement matrix, w = the noise of each of the n sensor observations.

        The Extended Kalman Filter is an algorithm that leverages our knowledge of the physics of motion of the system (i.e. the state space model) to make small adjustments to (i.e. to filter) the actual sensor measurements (i.e. what the robots sensors actually observed) to reduce the amount of noise, and as a result, generate a better estimate of the state of a system.

        Predict Step

        • Using the state space model of the robotics system, predict the state estimate at time k based on the state estimate at time k-1 and the control input applied at time k-1.

        • This is exactly what we did in my state spac modeling.

        • Predict the state covariance estimate based on the previous covariance and some noise.

          Update (Correct) Step

        • Calculate the difference between the actual sensor measurements at time k minus what the measurement model predicted the sensor measurements would be for the current timestep k.

        • Calculate the measurement residual covariance.

        • Calculate the near-optimal Kalman gain.

        • Calculate an updated state estimate for time k.

        • Update the state covariance estimate for time k.

          This research work, proposes a robot limitation framework taking into account the extended Kalman filter to beat the issues of restriction in mobile robot. The exploratory consequence of this paper delineates the strong and the precision of the proposed framework.

      2. Simulation Result

      Python simulation of mobile robot is performed using Extended Kalman Filter.

      Fig. 3 Mapping & Localization

      VI. CONCLUSION

      Fig. 2 EKF Localization

    3. DIFFERENTIAL DRIVE AND MAPPING Differential Drive

      Proposed mobile robots use a drive mechanism known as Differential drive. It consists of 2 drive wheels mounted on a common axis, And each wheel can independently being driven either forward or backward. While we can vary the velocity of each wheel, for the robot to perform rolling motion, the robot must rotate about a point that lies along their common left and right wheel axis. At any instance in time we can solve for R and :

      where l is the distance between the centers of the two wheels, Vr ; Vl are the right and left wheel velocities along the ground , and R is the signed distance from the instantaneous Center of Curvature (ICC) to the midpoint between the wheels.

      Mapping

      The ROS software in robots was connected to UI program in PC. Some main messages (topics) were communicated each other, for example, relative locations from robot to blocks, robots location Depth data of Kinect sensor in robots was transformed to scan data at Ordroid of robot body. But RGB image and depth image were transmitted to PC and processed.

    4. EXPERIMENT RESULTS

Following map is plotted using Kinect and ROS using g- mapping package of living room and passage between living room and bed room of our house.

As we can see from the results, system acts as we expected of it in real indoor environment. In SLAM (Simultaneous Localization and Mapping), by processing the detected features with landmarks which set previously in global coordinate system, we utilize Extended Kalman Filter to estimate the pose of robot. In the end, precise and usable map is provided to the end user. So, we expect that autonomous robots in unknown indoor environment would be possible easily without complex path planning.

REFERENCES

  1. Smith, R., Self, M. & Cheeseman, P. (1990), Estimating uncertain spatial relationships in robotics, Autonomous Robot Vehicles 8, 167- 193.

  2. Csorba, M. (1997), Simultaneous localization and map building, PhD thesis, University of Oxford.

  3. Fox, D., Thrun, S., Dellaert, F. & Burgard, W. (2000), Particle filters for mobile robot localization, in A. Doucet, N. de Freitas & N. Gordon, eds, Sequential Monte Carlo Methods in Prctice, Springer Verlag, New York.

  4. Montemerlo, M., Thrun, S., Koller, D. & Wegbreit, B. (2002), FastSLAM: A factored solution to the simultaneous localization and mapping problem, in The AAAI National Conference on Artificial Intelligence, AAAI, Edmonton, Canada, pp. 593-598.

  5. Liu, Y. & Thrun, S. (2002), Results for outdoor-SLAM using sparse extended information filters, in IEEE International Conference on Robotics and Automation ICRA, pp. 1227-1233.

  6. Marron, M.; Garcia, J.C. ; Sotelo, M.A. ; Cabello, M. ; Pizarro, D. ; Huerta, F. ; Cerro, J. (2007), ' Comparing a Kalman Filter and a Particle Filter in a Multiple Objects Tracking Application', Intelligent Signal Processing . WISP 2007. IEEE International Symposium on pp 1-6

  7. Brooks, A. & Bailey, T. (2009), HybridSLAM: Combining FastSLAM and EKF-SLAM for reliable mapping, in Algorithm Foundation of Robotics VIII, Springer Berlin/ Heidelberg.

  8. Teslic, L., Krjanc, I. & Klanar, G. (2011), EKF-based localization of a wheeled mobile robot in structured environments, Journal of Intelligent and Robotic Systems 62 (2),187-203.

  9. Mingyang Li and Anastasios I. Mourikis (2012), 'Vision-aided Inertial Navigation for Resource-constrained Systems', IEEE/RSJ International Conference on Intelligent Robots and Systems

  10. Zoran Miljkovi, Najdan Vukovi, Marko Miti & Bojan Babi (2013), New hybrid vision-based control approach for automated guided vehicles, Springer Berlin Heidelberg – Int J Adv Manuf Technol pp 231249.

  11. Jihong Min, Jungho Kim, Hyeongwoo Kim, Kiho Kwak and In So Kweon (2014), ' Hybrid Vision-based SLAM Coupled with Moving Object Tracking', Robotics and Automation (ICRA), IEEE International Conference on pp 867-874

  12. T. R. Gayathri et. al Feature based simultaneous localization and mapping, 2017, IEEE.

  13. Hameem Shanavas et. al. Design of an Autonomous Surveillance Robot using Simultaneous Localization And Mapping, 2018, IEEE.

Leave a Reply