Robot Three-Finger Grasping Strategy Based on DeeplabV3+
Abstract
:1. Introduction
2. Principal Analysis
2.1. Atrous Convolution Kernel
2.2. Encoder and Decoder
2.3. Overall Structure of DeeplabV3+
3. Improvement of Model Structure
- (1)
- The SPP with large and small atrous value is adopted on high-resolution image and low-resolution image, respectively, which can alleviate the grid effect and ensure the ability of the model to obtain multi-scale object information at the same time;
- (2)
- Sawtooth structure and loop structure have their own advantages and disadvantages. In principle, the complexity is reduced as much as possible while ensuring the performance of the model;
- (3)
- The difference between atrous values should not jump too large, which basically meets the distribution of arithmetic sequence.
4. Experiment
4.1. Introduction to Grasp Strategy
4.2. Introduction of Datasets
4.3. Experimental Environment
5. Result and Discussion
5.1. Training Process Analysis
5.2. Visualization of Data
5.3. Application Verification
5.3.1. Platform Introduction
5.3.2. System Architecture
5.3.3. Result Analysis
6. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Xie, R.; Liu, J.; Cao, R.; Qiu, C.S.; Duan, J.; Garibaldi, J.; Qiu, G. End-to-End Fovea Localisation in Colour Fundus Images With a Hierarchical Deep Regression Network. IEEE Trans. Med. Imaging 2021, 40, 116–128. [Google Scholar] [CrossRef] [PubMed]
- Guo, D.; Sun, F.; Liu, H.; Kong, T.; Fang, B.; Xi, N. A hybrid deep architecture for robotic grasp detection. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 1609–1614. [Google Scholar]
- Lenz, I.; Lee, H.; Saxena, A. Deep learning for detecting robotic grasps. Int. J. Robot. Res. 2015, 34, 705–724. [Google Scholar] [CrossRef] [Green Version]
- Xu, Y.; Wang, L.; Yang, A.; Chen, L.; Ynag, A. GraspCNN: Real-Time Grasp Detection Using a New Oriented Diameter Circle Representation. IEEE Access 2019, 7, 159322–159331. [Google Scholar] [CrossRef]
- Morrison, D.; Corke, P.; Leitner, J. Learning robust, real-time, reactive robotic grasping. Int. J. Robot. Res. 2019, 39, 183–201. [Google Scholar] [CrossRef]
- Redmon, J.; Angelova, A. Real-time grasp detection using convolutional neural networks. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 1316–1322. [Google Scholar]
- Kumra, S.; Kanan, C. Robotic grasp detection using deep convolutional neural networks. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 769–776. [Google Scholar]
- Avella, D.S.; Tripicchio, P.; Avizzano, C.A. A study on picking objects in cluttered environments: Exploiting depth features for a custom low-cost universal jamming gripper. Robot. Comput.-Integr. Manuf. 2020, 63, 101888. [Google Scholar]
- Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
- Liu, W.; Lee, J. A 3-D Atrous Convolution Neural Network for Hyperspectral Image Denoising. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5701–5715. [Google Scholar] [CrossRef]
- Wang, Z.; Wang, Z.; Dong, B.; Chen, C.; Yang, Y.; Yu, Z. Accelerating Atrous Convolution with Fetch-and-Jump Architecture for Activation Positioning. In Proceedings of the 2020 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA), IEEE, Nanjing, China, 23–25 November 2020; pp. 151–152. [Google Scholar]
- Liu, Y.; Zhu, X.; Zhao, X.; Cao, Y. Adversarial Learning for Constrained Image Splicing Detection and Localization Based on Atrous Convolution. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2551–2566. [Google Scholar] [CrossRef]
- Pan, X.; Li, L.; Yang, D.; He, Y.; Liu, Z.; Yang, H. An Accurate Nuclei Segmentation Algorithm in Pathological Image Based on Deep Semantic Network. IEEE Access 2019, 7, 110674–110686. [Google Scholar] [CrossRef]
- Lv, L.; Li, X.; Jin, J.; Li, X. Image Semantic Segmentation Method Based on Atrous Algorithm and Convolution CRF. In Proceedings of the 2019 IEEE 7th International Conference on Computer Science and Network Technology (ICCSNT), Dalian, China, 19–25 October 2019; pp. 160–165. [Google Scholar]
- Pan, D.; Yu, F.; Li, C.; Yang, L. Multi-oriented Scene Text Detector with Atrous Convolution. In Proceedings of the 2020 IEEE Information Communication Technologies Conference (ICTC), Nanjing, China, 29–31 May 2020; pp. 346–350. [Google Scholar]
- Jie, D.; Yu, H.; Li, C.; Wu, H.; Ni, F. Research of Teleoperation Grasping Control Method Based on Three-fingered Dexterous Hand. In Proceedings of the 2020 IEEE 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; pp. 3812–3816. [Google Scholar]
- Wang, D. SGDN: Segmentation-Based Grasp Detection Network For Unsymmetrical Three-Finger Gripper. arXiv 2020, arXiv:2005.08222. [Google Scholar]
- Jiang, Y.; Moseson, S.; Saxena, A. Efficient grasping from RGBD images: Learning using a new rectangle representation. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3304–3311. [Google Scholar]
- Chu, F.-J.; Xu, R.; Vela, P.A. Real-World Multiobject, Multigrasp Detection. IEEE Robot. Autom. Lett. 2018, 3, 3355–3362. [Google Scholar] [CrossRef] [Green Version]
- Pinto, L.; Gupta, A. Supersizing self-supervision: Learning to grasp from 50 K tries and 700 robot hours. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 3406–3413. [Google Scholar]
- Na, Y.-H.; Jo, H.; Song, J.-B. Learning to grasp objects based on ensemble learning combining simulation data and real data. In Proceedings of the 2017 IEEE 17th International Conference on Control, Automation and Systems (ICCAS), Jeju, Korea, 18–21 October 2017; pp. 1030–1034. [Google Scholar]
- Shang, W.; Song, F.; Zhao, Z.; Gao, H.; Cong, S.; Li, Z. Deep Learning Method for Grasping Novel Objects Using Dexterous Hands. IEEE Trans. Cybern. 2020, 10, 1–13. [Google Scholar] [CrossRef] [PubMed]
SPP_Number | 4 | 5 | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
Dilated_rate | 1-2-3-4 | 1-2-4-6 | 1-3-5-7 | 1-3-6-9 | 1-6-12-18 | 1-2-3-4-5 | 1-2-4-6-8 | 1-3-5-7-9 | 1-3-6-9-12 | 1-6-12-18-24 |
Structure | Loss | Loss_Able | Loss_Angle | Loss_Width | Valida Tion_IOU | Validation Graspable | Accuracy_ Image_Wise | Accuracy_ Object_Wise |
---|---|---|---|---|---|---|---|---|
1-2-3-4 | 0.01923 | 0.00184 | 0.00187 | 0.01552 | 0.97 | 0.920935 | 92.39% | 90.25% |
1-2-4-6 | 0.02021 | 0.00200 | 0.00202 | 0.01618 | 0.97 | 0.923298 | 97.83% | 97.04% |
1-3-6-9 | 0.02030 | 0.00198 | 0.00196 | 0.01636 | 0.97 | 0.928626 | 94.57% | 93.27% |
1-2-3-4-5 | 0.01995 | 0.00197 | 0.00195 | 0.01604 | 0.96 | 0.922368 | 95.65% | 94.67% |
1-2-4-6-8 | 0.01973 | 0.00187 | 0.00187 | 0.01600 | 0.96 | 0.927294 | 88.89% | 86.76% |
1-3-5-7-9 | 0.01918 | 0.00191 | 0.00192 | 0.01535 | 0.96 | 0.919595 | 95.65% | 96.98% |
1-6-12-18 [9] | 0.02026 | 0.00201 | 0.00198 | 0.01627 | 0.95 | 0.919623 | 92.39% | 92.05% |
Author | Year | Representation | Datasets | Algorithm | Accuracy (%) | Speed (ms) | |
---|---|---|---|---|---|---|---|
IW | OW | ||||||
Yun Jiang [18] | 2011 | Rectangle | Self-made | Two-step proces | 60.5 | 58.3 | 5000 |
Ian Lenz [3] | 2013 | Rectangle | Cornell grasp dataset | A two-step cascaded system | 88.4 | 88.7 | — |
Joseph Redmon [6] | 2015 | Rectangle | Cornell grasp dataset | Single-stage regression | 88.0% | 87.1% | 76 |
Sulabh Kumra [7] | 2017 | Rectangle | Cornell grasp dataset | ResNet-50 * 2 | 89.2 | 88.9 | 16 |
Di Guo [2] | 2017 | Rectangle | Cornell grasp dataset | Hybrid architecture | 93.2 | 89.1 | — |
Fu-Jen Chu [19] | 2018 | Rectangle | Cornell grasp dataset | ResNet-50 | 96.5 | 96.1 | 20 |
YULIN XU [4] | 2019 | Oriented diameter circle | Cornell grasp dataset | GraspCNN | 96.5% | — | 50 |
Douglas Morrison [5] | 2020 | Rectangle | Cornell grasp dataset | GG-CNN | 88.0% | — | 20 |
Wang Dexin [17] | 2020 | Triangle | Cornell grasp dataset | SGDN | 96.8% | 92.3% | 19 |
Ours | 2021 | Triangle | Cornell grasp dataset | SSGP | 97.83% | 97.04% | 19 |
Author | Accuracy on Common Objects (%) | Accuracy on Uncommon Objects (%) | Overall Accuracy (%) | Two/Three Fingers | Year |
---|---|---|---|---|---|
Ian Lenz [3] | 89 (89/100) | - | 89 | Two fingers (rectangle) | 2015 |
Pinto Lerrel [20] | 73 (109/150) | - | 73 | Two fingers (rectangle) | 2015 |
Na Yong-Ho [21] | 72 | 69 | 70.5 | Two fingers (rectangle) | 2017 |
Chu Fu-Jen [19] | 89 (89/100) | - | 89 | Two fingers (rectangle) | 2018 |
Morrison [5] | 92 (110/120) | 84 (67/80) | 88.5 | Two fingers (rectangle) | 2019 |
Shang Weiwei [22] | 92 (276/300) | - | 92 | Five fingers (rectangle) | 2020 |
Ours | 95 (95/100) | 90 (45/50) | 93.3 | Three fingers (triangle) | 2021 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Bai, Q.; Li, S.; Yang, J.; Shen, M.; Jiang, S.; Zhang, X. Robot Three-Finger Grasping Strategy Based on DeeplabV3+. Actuators 2021, 10, 328. https://doi.org/10.3390/act10120328
Bai Q, Li S, Yang J, Shen M, Jiang S, Zhang X. Robot Three-Finger Grasping Strategy Based on DeeplabV3+. Actuators. 2021; 10(12):328. https://doi.org/10.3390/act10120328
Chicago/Turabian StyleBai, Qiang, Shaobo Li, Jing Yang, Mingming Shen, Sanlong Jiang, and Xingxing Zhang. 2021. "Robot Three-Finger Grasping Strategy Based on DeeplabV3+" Actuators 10, no. 12: 328. https://doi.org/10.3390/act10120328
APA StyleBai, Q., Li, S., Yang, J., Shen, M., Jiang, S., & Zhang, X. (2021). Robot Three-Finger Grasping Strategy Based on DeeplabV3+. Actuators, 10(12), 328. https://doi.org/10.3390/act10120328