Depth Image Assisted Aiming for Scoring Goal in Wheeled Soccer Robot
DOI:
https://doi.org/10.12962/jaree.v8i2.268Abstrak
Wheeled soccer robot, as an automatic robot, required to have an advanced decision making system based on information it grasp from its surrounding. One of the most crucial decision making ability is to determine aiming angle when it is scoring goal.
This research will enhance the aiming ability for scoring goal by predicting unguarded area of goal. Combination of depth image and RGB image information will be used to predict the position of unguarded space in goal. This position will be converted into aiming angle for the robot. Intel Realsense D435i depth camera will be used to get RGB and depth image simultaneously
By using this method, the system is capable to predict unguarded area in all of 60 test points, with 1.3% average error for the predicted coordinate.
Referensi
S. E. Butner and M. Ghodoussi, “Transforming a surgical robot for human telesurgery,” IEEE Trans. Robot. Autom., vol. 19, no. 5, pp. 818–824, Oct. 2003, doi: 10.1109/TRA.2003.817214.
K. Severinson-Eklundh, A. Green, and H. Hüttenrauch, “Social and collaborative aspects of interaction with a service robot,” Robot. Auton. Syst., vol. 42, no. 3–4, pp. 223–234, Mar. 2003, doi: 10.1016/S0921-8890 (02) 00377-9.
Balkeshwar Singh, , N. Sellappan, and , Kumaradhas P, “Evolution of Industrial Robots and their Applications,” Int. J. Emerg. Technol. Adv. Eng.
P. Cano, Y. Tsutsumi, C. Villegas, and J. Ruiz-del-Solar, “Robust Detection of White Goals,” in RoboCup 2015: Robot World Cup XIX, Cham, 2015, pp. 229–238.
A. K. Mulya, F. Ardilla, and D. Pramadihanto, “Ball tracking and goal detection for middle size soccer robot using omnidirectional camera,” in 2016 International Electronics Symposium (IES), Denpasar, Indonesia, Sep. 2016, pp. 432–437, doi: 10.1109/ELECSYM.2016.7861045.
M. S. Ahn, H. Chae, D. Noh, H. Nam, and D. Hong, “Analysis and Noise Modeling of the Intel RealSense D435 for Mobile Robots,” in 2019 16th International Conference on Ubiquitous Robots (UR), Jeju, Korea (South), Jun. 2019, pp. 707–711, doi: 10.1109/URAI.2019.8768489.
J. D. Mejia-Trujillo et al., “KinectTM and Intel RealSenseTM D435 comparison: a preliminary study for motion analysis,” p. 4, 2019.
S. Sayyar-Roudsari, S. A. Hamoush, T. M. V. Szeto, and S. Yi, “Using a 3D Computer Vision System for Inspection of Reinforced Concrete Structures,” in Advances in Computer Vision, vol. 944, K. Arai and S. Kapoor, Eds. Cham: Springer International Publishing, 2020, pp. 608–618.
A. R. Tinkar et al., “Team Description Paper : IRIS Team 2020,” p.8.
M. Quigley et al., “ROS: an open-source Robot Operating System,” p. 6.
A. Grunnet-Jepsen, J. N. Sweetser, and J. Woodfill, “Best-Known-Methods for Tuning Intel® RealSenseTM D400 Depth Cameras for Best Performance,” p. 10.
##submission.downloads##
Diterbitkan
Terbitan
Bagian
Lisensi
Copyright
Submission of a manuscript implies that the submitted work has not been published before (except as part of a thesis or report, or abstract); that it is not under consideration for publication elsewhere; that its publication has been approved by all co-authors. If and when the manuscript is accepted for publication, the author(s) still hold the copyright and retain publishing rights without restrictions. Authors or others are allowed to multiply article as long as not for commercial purposes. For the new invention, authors are suggested to manage its patent before published. The license type is CC-BY-NC 4.0.
Disclaimer
No responsibility is assumed by publisher and co-publishers, nor by the editors for any injury and/or damage to persons or property as a result of any actual or alleged libelous statements, infringement of intellectual property or privacy rights, or products liability, whether resulting from negligence or otherwise, or from any use or operation of any ideas, instructions, procedures, products or methods contained in the material therein.


