Abstract:Aiming at the high positioning accuracy of automatic guided vehicle (AGV) in industrial scene, an improved visual simultaneous localization and mapping (VSLAM) algorithm is proposed. In the front end of the algorithm, the stereo images are captured using a binocular camera and sub-pixel-level matching point pairs are obtained through binocular matching algorithm. And the 3D information of these feature points is calculated in the camera coordinate system. Then the pose transformation is acquired using RansacPnP algorithm based on the 3D-2D matching point pairs. And the transformation is used as the initial value to do the local optimization by minimizing re-projection errors. Additionally, an uncertainty model is proposed based on the Hamming distance between the descriptors of matching point pairs. The model provides the information matrix for the constraints in the local optimization, which can improve the positioning accuracy. In the back end of the algorithm, reliable closed-loop information of artificial landmarks is detected using the monocular camera shooting vertically upwards, and then the global pose optimization is conducted. According to the motion model and working scene of the AGV, an optimization method based on global plane constraint is proposed to reduce the error of SLAM system. In the experiment, the accuracy of binocular odometer in the front end of the proposed algorithm is compared with the odometers in ORB-SLAM2 and libviso2 algorithm using the offline KITTI dataset. The online experiments in the factory environment are conducted to test the real localization accuracy and robustness of the whole system. The results show that the algorithm is of good overall performance, and its positioning errors are within 10cm, while the positioning frequency can reach 20Hz, which can meet the industrial requirements.
[1] 林辉灿,吕强,张洋,等.稀疏和稠密的VSLAM的研究进展[J].机器人,2016,38(5):621-631.Lin H C, Lü Q, Zhang Y, et al. The sparse and dense VSLAM:A survey[J]. Robot, 2016, 38(5):621-631.
[2] 赵洋,刘国良,田国会,等.基于深度学习的视觉SLAM综述[J].机器人,2017,39(6):889-896.Zhao Y, Liu G L, Tian G H, et al. A survey of visual SLAM based on deep learning[J]. Robot, 2017, 39(6):889-896.
[3] 刘浩敏,章国锋,鲍虎军.基于单目视觉的同时定位与地图构建方法综述[J].计算机辅助设计与图形学学报,2016,28(6):855-868.Liu H M, Zhang G F, Bao H J. A survey of monocular simultaneous localization and mapping[J]. Journal of Computer-Aided Design & Computer Graphics, 2016, 28(6):855-868.
[4] Davison A J, Reid I D, Molton N D, et al. MonoSLAM:Real-time single camera SLAM[J]. IEEE Transactions on Pattern Analysis & Machine Intelligence, 2007, 29(6):1052-1067.
[5] Mourikis A I, Roumeliotis S I. A multi-state constraint Kalman filter for vision-aided inertial navigation[C]//IEEE International Conference on Robotics and Automation. Piscataway, USA:IEEE, 2007:3565-3572.
[6] Klein G, Murray D. Parallel tracking and mapping for small AR workspaces[C]//IEEE International Symposium on Mixed and Augmented Reality. Piscataway, USA:IEEE, 2008:1-10.
[7] Mur-Artal R, Montiel J M M, Tardos J D. ORB-SLAM:A versatile and accurate monocular SLAM system[J]. IEEE Transactions on Robotics, 2017, 31(5):1147-1163.
[8] Tan W, Liu H, Dong Z, et al. Robust monocular SLAM in dynamic environments[C]//IEEE International Symposium on Mixed and Augmented Reality. Piscataway, USA:IEEE, 2013:209-218.
[9] Engel J, Schops T, Cremers D. LSD-SLAM:Large-scale direct monocular SLAM[M]//Lecture Notes in Computer Science, vol. 8690. Berlin, Germany:Springer-Verlag, 2014:834-849.
[10] Newcombe R A, Lovegrove S J, Davison A J. DTAM:Dense tracking and mapping in real-time[C]//IEEE International Conference on Computer Vision. Piscataway, USA:IEEE, 2011:2320-2327.
[11] Hwang S Y, Song J B. Clustering and probabilistic matching of arbitrarily shaped ceiling features for monocular vision-based SLAM[J]. Advanced Robotics, 2013, 27(10):739-747.
[12] Choi H, Kim R, Kim E, et al. An efficient ceiling-view SLAM using relational constraints between landmarks[J]. International Journal of Advanced Robotic Systems, 2014, 11(1):1-11.
[13] Wu H, Tian G H, Duan P, et al. The design of a novel artificial label for robot navigation[M]//Lecture Notes in Electrical Engineering, vol.255. Berlin, Germany:Springer-Verlag, 2013:479-487.
[14] Okuyama K, Kawasaki T, Kroumov V. Localization and position correction for mobile robot using artificial visual landmarks[C]//International Conference on Advanced Mechatronic Systems. Piscataway, USA:IEEE, 2011:414-418.
[15] Rublee E, Rabaud V, Konolige K, et al. ORB:An efficient alternative to SIFT or SURF[C]//IEEE International Conference on Computer Vision. Piscataway, USA:IEEE, 2011:2564-2571.
[16] Fischler M A, Bolles R C. Random sample consensus:A paradigm for model fitting with applications to image analysis and automated cartography[J]. Readings in Computer Vision, 1987, 24(6):726-740.
[17] Kummerle R, Grisetti G, Strasdat H, et al. g2o:A general framework for graph optimization[C]//IEEE International Conference on Robotics and Automation. Piscataway, USA:IEEE, 2011:3607-3613.
[18] Gao X S, Hou X R, Tang J, et al. Complete solution classification for the perspective-three-point problem[J]. IEEE Transactions on Pattern Analysis & Machine Intelligence, 2003, 25(8):930-943.
[19] Lepetit V, Moreno-Noguer F, Fua P. EPnP:An accurate O(n) solution to the PnP problem[J]. International Journal of Computer Vision, 2009, 81(2):155-166.
[20] Li Y H, Zhu S Q, Yu Y Q, et al. An improved graph-based visual localization system for indoor mobile robot using newly designed markers[J]. International Journal of Advanced Robotic Systems, 2018, 15(2). DOI:10.1177/1729881418769191.
[21] Geiger A, Lenz P, Stiller C, et al. Vision meets robotics:The KITTI dataset[J]. International Journal of Robotics Research, 2013, 32(11):1231-1237.