학술논문

Partial Depth Estimation with Single Image Using YOLO and CNN for Robot Arm Control
Document Type
Conference
Source
2022 IEEE International Conference on Mechatronics and Automation (ICMA) Mechatronics and Automation (ICMA), 2022 IEEE International Conference on. :1727-1731 Aug, 2022
Subject
Bioengineering
Components, Circuits, Devices and Systems
Computing and Processing
Robotics and Control Systems
Signal Processing and Analysis
Deep learning
Automation
Mechatronics
Image recognition
Estimation
Object detection
Manipulators
Depth Estimation
Object Detection
YOLO
Deep Learning
Robot Arm
Pick and Place
Language
ISSN
2152-744X
Abstract
This paper presents experimental results of partial object detection using YOLO (You Only Look Once) and partial depth estimation using CNN (Convolutional Neural Network) for application to a robot arm control. In recent years, image recognition and its application automation, as an alternative to advanced work, are attracting attention in various fields. In order for robot arms used in factories to perform high-value-added and flexible work, it is necessary to control them by object detection and recognition by deep learning. In this study, the authors propose a new approach for estimating the depth of partial images using YOLO and uses it to control the robot arm. In the experiments, both ends of a pen detected by YOLO are used for the input to a CNN. The detected parts are saved to images with a size of about 60 × 60 pixels, and the depths are estimated by giving the cropped images to the CNN. A desktop-sized robot with 4DOFs can successfully pick the pen by referring the depths. The effectiveness of the proposed method is demonstrated through experiments.