Demonstration Programming and Optimization Method ofCooperative Robot Based on Multi-Source Information Fusion
WANG Fei1, QI Huan2, ZHOU Xingqun2, WANG Jianhui2
1. Faculty of Robot Science and Engineering, Northeastern University, Shenyang 110819, China;
2. College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
Abstract:In order to solve the problems of complex assembly and learning process of robot and high requirements for the programming technology, an implicit interaction method based on the fusion of forearm sEMG (surface electromyography) and inertial multi-source information is proposed to realize robot demonstration programming. Based on the assembly experience obtained from the demonstration learning by presenters, a multiple deep deterministic policy gradient (M-DDPG) algorithm is proposed to modify the assembly parameters in order to improve the adaptability to the changes of assembly objects and environment. On the basis of demonstration programming, reinforcement learning is carried out to ensure that the robot performs its tasks stably. In demonstration programming experiment, an improved PCNN (parallel convolution neural network) is proposed, named 1-D PCNN (1D-PCNN), which automatically extracts feature inertia and EMG through 1-dimensional convolution and pooling and enhances the generalization performance and accuracy of gesture recognition. In the demonstration reproduction experiment, the Gaussian mixture model (GMM) is used to statistically encode the demo data, and Gaussian mixture regression (GMR) is used to reproduce the robot's trajectory and eliminate noise points. Finally, the environmental changes in the X-axis and Y-axis directions are acquired respectively with Primesense Carmine camera by using the fusion tracking algorithm based on the frame difference method and the multiple kernel correlation filter (MKCF) algorithm. Two identical network structures are used to concurrently carry out the deep reinforcement learning of continuous processes. When the relative position of the peg and hole changes, the robotic arm can automatically adjust the end-effector position according to the generalization strategy model learned by reinforcement learning so as to realize the demonstration learning of the peg-in-hole assembly.
[1] 马静,李宇.工业机器人自动装配生产线的研制[J].电子技术与软件工程,2017(8):132. Ma J, Li Y. Development of an automatic assembly line based industrial robots[J]. Electronic Technology & Software Engineering, 2017(8):132.
[2] 潘立,鲍官军,胥芳,等.六自由度装配机器人的动态柔顺性控制[J].浙江大学学报(工学版),2018,52(1):125-132. Pan L, Bao G J, Xu F, et al. Dynamic compliant control of six DOF assembly robot[J]. Journal of Zhejiang University (Engineering Science), 2018, 52(1):125-132.
[3] 党宏社,候金良,强华,等.基于视觉引导的SCARA机器人自动装配系统[J].电子技术应用,2017,43(5):21-24. Dang H S, Hou J L, Qiang H, et al. SCARA automatic assembly system based on vision guided[J]. Application of Electronic Technique, 2017, 43(5):21-24.
[4] 沈程慧,白瑞林,李新.视觉引导的装配机器人平面定位补偿方法[J].激光技术,2017,41(1):79-84. Shen C H, Bai R L, Li X. Plane positioning compensation method for assembly robot with visual guiding[J]. Laser Technology, 2017, 41(1):79-84.
[5] Ngeo J G, Tamei T, Shibata T. Continuous and simultaneous estimation of finger kinematics using inputs from an EMG-to-muscle activation model[J]. Journal of Neuroengineering & Rehabilitation, 2014, 11:No.122.
[6] 佟丽娜,侯增广,彭亮,等.基于多路sEMG时序分析的人体运动模式识别方法[J].自动化学报,2014,40(5):810-821. Tong L N, Hou Z G, Peng L, et al. Multi-channel sEMG time series analysis based human motion recognition method[J]. Acta Automatica Sinica, 2014, 40(5):810-821.
[7] Schwung D, Csaplar F, Schwung A, et al. An application of reinforcement learning algorithms to industrial multi-robot stations for cooperative handling operation[C]//IEEE International Conference on Industrial Informatics. Piscataway, USA:IEEE, 2017:194-199.
[8] Rozo L, Jiménez P, Torras C. Robot learning from demonstration of force-based tasks with multiple solution trajectories[C]//IEEE 15th International Conference on Advanced Robotics. Piscataway, USA:IEEE, 2011:124-129.
[9] Rozo L, Jiménez P, Torras C. A robot learning from demonstration framework to perform force-based manipulation tasks[J]. Intelligent Service Robotics, 2013, 6(1):33-51.
[10] Toris R, Suay H B, Chernova S. A practical comparison of three robot learning from demonstration algorithms[C]//ACM/IEEE International Conference on Human-Robot Interaction. NewYork, USA:ACM, 2012:261-262.
[11] Wang F, Zhou J, Lin J, et al. A comparative study on sign recognition using sEMG and inertial sensors[C]//6th lEEE Annual International Conference on Cyber Technology in Automation, Control, and Intelligent Systems. Piscataway, USA:IEEE, 2016:290-295.
[12] Krizhevsky A, Sutskever I, Hinton G E. ImageNet classification with deep convolutional neural networks[C]//26th Annual Conference on Neural Information Processing Systems. Canada:Neural Information Processing System Foundation, 2012:1097-1105.
[13] LeCun Y, Bengio Y, Hinton G. Deep learning[J]. Nature, 2015, 521(7553):436-444.
[14] Goodfellow I, Bengio Y, Courville A. Deep learning[M]. Cambridge, USA:MIT Press, 2016.
[15] Shen X J, Zhu X C, Jiang X P, et al. Visualization of non-metric relationships by adaptive learning multiple maps t-SNE regularization[C]//IEEE International Conference on Big Data. Piscataway, USA:IEEE, 2017:3882-3887.
[16] Zhang R, Gong W, Grzeda V, et al. An adaptive learning rate method for improving adaptability of background models[J]. IEEE Signal Processing Letters, 2013, 20(12):1266-1269.
[17] Holding T, Lestas I. On the convergence to saddle points of concave-convex functions, the gradient method and emergence of oscillations[C]//53rd IEEE Annual Conference on Decision and Control. Piscataway, USA:IEEE, 2014:1143-1148.
[18] Chen J, Lau H Y K, Xu W, et al. Towards transferring skills to flexible surgical robots with programming by demonstration and reinforcement learning[C]//8th International Conference on Advanced Computational Intelligence. Piscataway, USA:IEEE, 2016:378-384.
[19] Sheng W, Thobbi A, Ye G. An integrated framework for human-robot collaborative manipulation[J]. IEEE Transactions on Cybernetics, 2015, 45(10):2030-2041.
[20] Gu S, Holly E, Lillicrap T, et al. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates[C]//IEEE International Conference on Robotics and Automation. Piscataway, USA:IEEE, 2016:3389-3396.
[21] Silver D, Lever G, Heess N, et al. Deterministic policy gradient algorithms[C]//31st International Conference on International Conference on Machine Learning. International Machine Learning Society, 2014:605-619.