Preview

Mekhatronika, Avtomatizatsiya, Upravlenie

Advanced search

Scan Matching for Navigation of a Mobile Robot in Semi-Structured Terrain Conditions

https://doi.org/10.17587/mau.22.246-253

Abstract

To ensure unmanned autonomous movement of ground robotic means, it is required to accurately determine the position and orientation of the robot. The present study is related to the estimation of coordinates by comparing the scans of a laser scanning rangefinder in conditions of semi-structed infrastructure and the absence of a global satellite communications signal. The existing methods of comparing scans have significant drawbacks in the conditions of movement over a semi-structured terrain, associated both with the processing time of data from the laser scanning rangefinder, and with the quality of the results obtained. The scan is preliminarily placed in a map consisting of cells. Each cell of around point scan is described by forces represented by the laws of physics or probability theory. In the cells of the map, we take into account the mutual influence of all forces from each point of the scan and thus we obtain the resulting artificial potential field of the scan. The position of the robot is estimated by the change in the number of acting forces of one scan per points of the next scan taking into account their direction. We estimate the orientation of the robot based on the sum of the vector products of the forces and distances to the given forces acting on the points of the next scan. This method allows you to calculate the displacement of the robot between scans regardless of road conditions and terrain. This article presents the results of an experimental verification of the method on a mock-up of a mobile robot equipped with a Velodyne HDL-32 LIDAR. We indicate the operating conditions of the method for a given LIDAR, as well as the time spent on calculating the bias estimate. Given the peculiarities of the LIDAR, we present a method for eliminating the Doppler Effect (distortion) for the original point cloud. A comparative analysis of the developed method for integrating wheel odometry data, inertial and satellite navigation using the Extended Kalman Filter shows the applicability of this method to assess the position and orientation of the robot in conditions of its movement over rough terrain.

About the Author

N. A. Buzlov
Bauman Moscow Technical University
Russian Federation

Buzlov Nikita A., Postgraduate 

Moscow, 105005 



References

1. Brands and companies. Top 50 developers of autonomous cars and technologies in the world, Cars section of the drone website, available at: https://bespilot.com/marki-i-kompanii.

2. Sokolov S. M., Boguslavsky A. A. Unified layout of integrated intelligent systems for information support of autopilots of mobile devices, Izvestia of the Southern Federal University. Technical Science, 2016, no. 2 (175), pp. 200—213 (in Russian).

3. Gerasimov V. N., Mikhailov B. B. Solving the problem of controlling the movement of a mobile robot in the presence of dynamic obstacles, Vestnik BMSTU. Instrumentation. Special issue "Robotic systems", 2012, no. 6, pp. 83—92 (in Russian).

4. Reutov E. V., Golovina L. S. Problematic aspects of the operation and design of unmanned vehicles, International Scientific Research Journal. 2019. no. 5-2 (83), pp. 14—18 (in Russian).

5. Adamova A. D., Zhukabaeva T. K., Mukanova Zh. A. Review of methods for localization and construction of maps of the surrounding area of mobile robots, Automation. Computer Science, 2018, no. 2 (43), pp. 57—62 (in Russian).

6. Kaess M., Dellaert F. Probabilistic structure matching for visual SLAM with a multi-camera rig, Computer Vision and Image Understanding, 2010, vol. 114, pp. 286—296.

7. Carrera G., Angeli A., Davison A. J. SLAM-based automatic extrinsic calibration of a multi-camera rig, 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 2011, pp. 2652—2659.

8. Davison A., Gonz lez Y., Kita N. Real-time 3D SLAM with wide-angle vision, IFAC Proceedings Volumes, 2004, vol. 37, no. 8, pp. 868—873.

9. Altukhov V. G. Overview of omnidirectional vSLAM mobile robot navigation technology, Collection of scientific papers of the Novosibirsk State Technical University, 2018, no. 2, pp. 81—92 (in Russian).

10. Borisov A. G., Gol S. A., Luksha S. S. Analysis of the efficiency of algorithms for processing three-dimensional laser scans in the problem of building maps for navigation of mobile robots, Bulletin of the Ryazan State Radio Engineering University, 2013, no. 46 (3), pp. 34—42 (in Russian).

11. Besl P. J. and McKay N. D. A method for registration of 3-D shapes, IEEE Transactions on Pattern Analysis and Machine Intelligence, 1992. Vol. 14, pp. 239—256.

12. Removing outliers using a Conditional or Radius Outlier removal, PointCloud.org, available at: http://pointclouds.org/documentation/tutorials/remove_outliers.php (дата обращения 20.12.2019).

13. Eggert D. W., Fitzgibbon A. W., Fisher R. B. Simultaneous registration of multiple range views for use in reverse engineering of CAD models, Comput. Vis. Image Underst., 1998, vol. 69, pp. 253—272.

14. Bergevin R., Soucy M., Gagnon H., Laurendeau D. Towards a general multi-view registration technique, IEEE Trans. Pattern Anal. Mach. Intell., 1996, vol. 18, no. 5 pp. 540—547.

15. Lu F. Robot Pose Estimation in Unknown Environments by Matching 2D Range Scans, Journal of Intelligent and Robotic Systems, 1997, no. 18 (3), pp. 249—275.

16. B dkowski J, Mas owski A. GPGPU implementation of On-Line point to plane 3D data registration, Proceedings of the 2011 International Conference on Electrical Engineering and Informatics, 2011, Jul 17, pp. 1—6.

17. Minguez J., Montesano L., Lamiraux F. Metric-Based Iterative Closest Point Scan Matching for Sensor Displacement Estimation, IEEE Transactions on Robotics, 2006, vol. 22, iss. 5, pp. 1047—1054

18. Biber P., Strasser W. The normal distributions transform a new approach to laser scan matchin, Proceedings of Intelligent Robots and Systems, 2003. Vol. 3, pp. 2743—2748.

19. Magnusson M. The three-dimensional normal-distributions transform: an efficient representation for registration, surface analysis, and loop detection Philosophy doctor, Örebro universitet, 2009.

20. Steder B., Rusu R. B., Burgard W. Point feature extraction on 3D range scans taking into account object boundaries, Computer Science Published in IEEE International Conference, 2011, DOI:10.1109/ICRA.2011.5980187.

21. Steder B., Rusu R. B., Konolige K., Burgard W. NARF: 3D range image features for object recognition, Workshop on Defining and Solving Realistic Perception Problems in Personal Robotics at the IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), 2010, Oct 8. Vol. 44.

22. Radu B. R., Blodow N., Beetz M. Fast Point Feature Histograms (FPFH) for 3D registration, 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 2009.

23. Robot Operating System, available at: http://www.ros.org/wiki/ (date of access 04.03.2018).

24. Merriaux P., Dupuis Y., Boutteau R., Vasseur P., Savatier X. LiDAR point clouds correction acquired from a moving car based on CAN-bus data, arXiv preprint arXiv:1706.05886. 2017 Jun 19.

25. Moore T., Stouch D. A. Generalized extended Kalman filter implementation for the robot operating system, Intelligent autonomous systems, 2016, no. 13, pp. 335—348.


Review

For citations:


Buzlov N.A. Scan Matching for Navigation of a Mobile Robot in Semi-Structured Terrain Conditions. Mekhatronika, Avtomatizatsiya, Upravlenie. 2021;22(5):246-253. https://doi.org/10.17587/mau.22.246-253

Views: 496


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 1684-6427 (Print)
ISSN 2619-1253 (Online)