Determination of Angular Status and Dimensional Properties of Objects for Grasping with Robot Arm

Authors

Keywords:

Learning and Adaptive Systems, Grasping, Kinematics, Recognition.

Abstract

In any task, robot arms can work more effectively without human control. With components such as imaging devices, it is possible to program robots to control autonomously. In this study, the problem of grasping moving objects with the robot arm is realized fully automatically. Deep learning-based You Only Look Once(YOLO) recognizes the objects moving at unknown speeds on a conveyor belt. The velocities of the detected objects are calculated by image processing methods with 2D camera frames. An Artificial Neural Network (ANN) was trained to output the angles required for the robot arm to grasp. According to the got values, the robot arm waits for the object to arrive and then realizes the grip. In the trials, the robot arm achieved successful gripping of over 93% without knowing the sizes, speeds, and positions of the objects.

Downloads

Download data is not yet available.

Author Biographies

Kürşad Uçar, Selçuk University, Konya, Turkey

Kür¸sad UÇAR received the Master degree from Department of Electric –Electronic Engineering, Institute of Science, Selçuk University, Konya, Turkey (2018). He continues his Ph. at same university. He is a research assistant at Selçuk University since 2016. His research interests are in image processing, robotic, automation and control, and machine learning.

Hasan Erdinç Koçer, Selçuk University, Konya, Turkey

Hasan Erdinç KOÇER received the Ph.D. degree from Department of Electric–Electronic Engineering, Institute of Science, Selçuk University, Konya, Turkey (2007), where he is an associate professor
since 2018. His research interests are in machine vision and pattern recognition, industrial automation
on camera, medical image processing and, automation and control

References

E. J. Van Henten, J. Hemming, B. Van Tuijl, J. Kornet, J. Meuleman,

J. Bontsema, and E. Van Os, “An autonomous robot for harvesting

cucumbers in greenhouses,” Autonomous robots, vol. 13, no. 3, pp. 241–

, 2002.

C. Zhihong, Z. Hebin, W. Yanbo, L. Binyan, and L. Yu, “A vision-based

robotic grasping system using deep learning for garbage sorting,” in 2017

th Chinese control conference (CCC), pp. 11223–11226, IEEE, 2017.

Y. Liu, Y. Zhong, X. Chen, P. Wang, H. Lu, J. Xiao, and H. Zhang, “The

design of a fully autonomous robot system for urban search and rescue,”

in 2016 IEEE International Conference on Information and Automation

(ICIA), pp. 1206–1211, IEEE, 2016.

M. J. Shafiee, B. Chywl, F. Li, and A. Wong, “Fast yolo: A fast you

only look once system for real-time embedded object detection in video,”

arXiv preprint arXiv:1709.05943, 2017.

W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and A. C.

Berg, “Ssd: Single shot multibox detector,” in European conference on

computer vision, pp. 21–37, Springer, 2016.

S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time

object detection with region proposal networks,” Advances in neural

information processing systems, vol. 28, 2015.

I. V. Zoev, A. P. Beresnev, and N. G. Markov, “Convolutional neural

networks of the yolo class in computer vision systems for mobile robotic

complexes,” in 2019 International Siberian Conference on Control and

Communications (SIBCON), pp. 1–5, IEEE, 2019.

A. C´ orovic´, V. Ilic´, S. Ðuric´, M. Marijan, and B. Pavkovic´, “The realtime

detection of traffic participants using yolo algorithm,” in 2018 26th

Telecommunications Forum (TELFOR), pp. 1–4, IEEE, 2018.

J.-W. Chang, R.-J. Wang, W.-J. Wang, and C.-H. Huang, “Implementation

of an object-grasping robot arm using stereo vision measurement

and fuzzy control,” International Journal of Fuzzy Systems, vol. 17,

no. 2, pp. 193–205, 2015.

A. Saxena, J. Driemeyer, and A. Y. Ng, “Robotic grasping of novel

objects using vision,” The International Journal of Robotics Research,

vol. 27, no. 2, pp. 157–173, 2008.

K.-T. Song and S.-C. Tsai, “Vision-based adaptive grasping of a humanoid

robot arm,” in 2012 IEEE International Conference on Automation

and Logistics, pp. 155–160, IEEE, 2012.

B. Wang, L. Jiang, J. Li, H. Cai, and H. Liu, “Grasping unknown

objects based on 3d model reconstruction,” in Proceedings, 2005

IEEE/ASME International Conference on Advanced Intelligent Mechatronics.,

pp. 461–466, IEEE, 2005.

A. Bicchi and V. Kumar, “Robotic grasping and contact: A review,”

in Proceedings 2000 ICRA. Millennium conference. IEEE international

conference on robotics and automation. Symposia proceedings (Cat. No.

CH37065), vol. 1, pp. 348–353, IEEE, 2000.

O.-L. Ouabi, P. Pomarede, N. Declercq, N. Zeghidour, M. Geist, and

C. Pradalier, “Learning the propagation properties of plate-like structures

for lamb wave-based mapping,” 2021.

P. K. Allen, A. Timcenko, B. Yoshimi, and P. Michelman, “Automated

tracking and grasping of a moving object with a robotic hand-eye

system,” IEEE Transactions on Robotics and Automation, vol. 9, no. 2,

pp. 152–165, 1993.

Y. Zhang, L. Li, M. Ripperger, J. Nicho, M. Veeraraghavan, and

A. Fumagalli, “Gilbreth: A conveyor-belt based pick-and-sort industrial

robotics application,” in 2018 Second IEEE International Conference on

Robotic Computing (IRC), pp. 17–24, IEEE, 2018.

D. K. Reddy et al., “Sorting of objects based on colour by pick and

place robotic arm and with conveyor belt arrangement,” Int. J. Mech.

Eng. & Rob. Res, vol. 3, no. 1, p. 3, 2014.

W. T. Abbood, O. I. Abdullah, and E. A. Khalid, “A real-time automated

sorting of robotic vision system based on the interactive design approach,”

International Journal on Interactive Design and Manufacturing

(IJIDeM), vol. 14, no. 1, pp. 201–209, 2020.

S. Kim, A. Shukla, and A. Billard, “Catching objects in flight,” IEEE

Transactions on Robotics, vol. 30, no. 5, pp. 1049–1065, 2014.

A. Menon, B. Cohen, and M. Likhachev, “Motion planning for smooth

pickup of moving objects,” in 2014 IEEE International Conference on

Robotics and Automation (ICRA), pp. 453–460, IEEE, 2014.

N. Marturi, M. Kopicki, A. Rastegarpanah, V. Rajasekaran, M. Adjigble,

R. Stolkin, A. Leonardis, and Y. Bekiroglu, “Dynamic grasp and

trajectory planning for moving objects,” Autonomous Robots, vol. 43,

no. 5, pp. 1241–1256, 2019.

C. De Farias, M. Adjigble, B. Tamadazte, R. Stolkin, and N. Marturi,

“Dual quaternion-based visual servoing for grasping moving objects,” in

IEEE 17th International Conference on Automation Science and

Engineering (CASE), pp. 151–158, IEEE, 2021.

D. Wu, S. Lv, M. Jiang, and H. Song, “Using channel pruning-based

yolo v4 deep learning algorithm for the real-time and accurate detection

of apple flowers in natural environments,” Computers and Electronics

in Agriculture, vol. 178, p. 105742, 2020.

A. Daffertshofer, C. J. Lamoth, O. G. Meijer, and P. J. Beek, “Pca in

studying coordination and variability: a tutorial,” Clinical biomechanics,

vol. 19, no. 4, pp. 415–428, 2004.

M. A. Hossain and S. Afrin, “Optical character recognition based on

template matching,” Global Journal of Computer Science and Technology,

Published

2022-10-21

How to Cite

Uçar, K., & Koçer, H. E. (2022). Determination of Angular Status and Dimensional Properties of Objects for Grasping with Robot Arm. IEEE Latin America Transactions, 21(2), 335–343. Retrieved from https://latamt.ieeer9.org/index.php/transactions/article/view/6999