JRM Vol.34 No.4 pp. 877-886
doi: 10.20965/jrm.2022.p0877


Localization Method Using Camera and LiDAR and its Application to Autonomous Mowing in Orchards

Hiroki Kurita*, Masaki Oku*, Takeshi Nakamura**, Takeshi Yoshida*, and Takanori Fukao*

*Ritsumeikan University
1-1-1 Noji-higashi, Kusatsu, Shiga 525-8577, Japan

**OREC Co., Ltd.
548-22 Hiyoshi, Hirokawa-cho, Yame-gun, Fukuoka 834-0195, Japan

July 25, 2020
June 1, 2022
August 20, 2022
deep learning, LiDAR, sensor fusion, UGV, autonomous mowing
Localization Method Using Camera and LiDAR and its Application to Autonomous Mowing in Orchards

Identified pear trunks

This paper presents a localization method using deep learning and light detection and ranging (LiDAR) for unmanned ground vehicle (UGV) in field environment. We develop a sensor fusion algorithm that UGV recognizes natural objects from RGB camera using deep learning and measures the distance to the recognized objects with LiDAR. UGV calculates its position relative to the objects, creates a reference path, and then executes path following control. The method is applied to autonomous mowing operation in orchard. A mower is tracked by UGV. UGV needs to run along a tree row keeping an appropriate distance from tree trunks, by which the mowing arm of the tracked mower properly touches the trunks. Field experiments are conducted in pear and apple orchards. UGV localizes self position relative to trees and performs autonomous mowing successfully. The results show that the presented method is effective.

Cite this article as:
H. Kurita, M. Oku, T. Nakamura, T. Yoshida, and T. Fukao, “Localization Method Using Camera and LiDAR and its Application to Autonomous Mowing in Orchards,” J. Robot. Mechatron., Vol.34, No.4, pp. 877-886, 2022.
Data files:
  1. [1] United Nations, Department of Economic and Social Affairs, Population Division, “World Population Prospects 2019 Highlights,” 2019.
  2. [2] Y. Nagasaka, N. Umeda, Y. Kanetani, K. Taniwaki, and Y. Sasaki, “Automated rice transplanter using global positioning and gyroscopes,” Computers and Electronics in Agriculture, Vol.43, No.3, pp. 222-234, 2004.
  3. [3] R. Takai, O. Barawid Jr., K. Ishii, and N. Noguchi, “Development of crawler-type robot tractor based on GPS and IMU,” Proc. of the 3rd IFAC Int. Conf. AGRICONTROL 2010, A3-5, 2010.
  4. [4] H. Kurita, M. Iida, W. Cho, and M. Suguri, “Rice autonomous Harvesting: Operation Framework,” J. of Field Robotics, Vol.34, No.6, pp. 1084-1099, 2017.
  5. [5] Q. Wang, J. Zhang, Y. Liu, and X. Zhang, “High-Precision and Fast LiDAR Odometry and Mapping Algorithm,” J. Adv. Comput. Intell. Intell. Inform., Vol.26, No.2, pp. 206-216, 2022.
  6. [6] L. Sun, R. P. Singh, and F. Kanehiro, “Visual SLAM Framework Based on Segmentation with the Improvement of Loop Closure Detection in Dynamic Environments,” J. Robot. Mechatron., Vol.33, No.6, pp. 1385-1397, 2021.
  7. [7] A. Sujiwo, T. Ando, E. Takeuchi, Y. Ninomiya, and M. Edahiro, “Monocular Vision-Based Localization Using ORB-SLAM with LIDAR-Aided Mapping in Real-World Robot Challenge,” J. Robot. Mechatron., Vol.28, No.4, pp. 479-490, 2016.
  8. [8] S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal network,” Proc. of the 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 1-9, 2016.
  9. [9] J. Redmon, S. Diwala, R. Girshick, and A. Farhadi, “You Only Look Once: Unified, real-time object detection,” Proc. of the 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 779-788, 2016.
  10. [10] J. Redmon and A. Farhadi, “YOLO9000: Better, faster, stronger,” Proc. of the 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 6517-6525, 2017.
  11. [11] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Y. Fu, and A. C. Berg, “SSD: Single shot multibox detector,” Proc. of the European Conf. on Computer Vision, pp. 21-37, 2016.
  12. [12] S. Kato, S. Tokunaga, Y. Maruyama, S. Maeda, M. Hirabayashi, Y. Kitsukawa, A. Monrroy, T. Ando, Y. Fujii, and T. Azumi, “Autoware on Board: Enabling Autonomous Vehicles with Embedded Systems,” Proc. of the 9th ACM/IEEE Int. Conf. on Cyber-Physical Systems (ICCPS2018), pp. 287-296, 2018.
  13. [13] J. Yoshida, T. Sugimachi, and T. Fukao, “Autonomous Driving fo a Truck Based on Path Following Control,” Trans. of the JSME, C, Vol.77, pp. 4125-4135, 2011.
  14. [14] Y. Kanayama, Y. Kimura, F. Miyazaki, and T. Noguchi, “A stable tracking control method for an autonomous mobile robot,” Proc. of IEEE Int. Conf. on Robotics and Automation, pp. 384-389, 1990.

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Sep. 22, 2022