single-rb.php

JRM Vol.36 No.5 pp. 1055-1064
doi: 10.20965/jrm.2024.p1055
(2024)

Paper:

Drone System Remotely Controlled by Human Eyes: A Consideration of its Effectiveness When Remotely Controlling a Robot

Yoshihiro Kai* ORCID Icon, Yuki Seki*, Yuze Wu*, Allan Paulo Blaquera** ORCID Icon, and Tetsuya Tanioka*** ORCID Icon

*Tokai University
4-1-1 Kitakaname, Hiratsuka, Kanagawa 259-1292, Japan

**St. Paul University Philippines
Mabini St., Tuguegarao City, Cagayan 3500, Philippines

***Tokushima University
3-18-15 Kuramoto-cho, Tokushima 770-8509, Japan

Received:
April 1, 2024
Accepted:
August 5, 2024
Published:
October 20, 2024
Keywords:
robot, drone, control system, eye-tracking device, visual support
Abstract

In recent years, Japan has experienced numerous natural disasters, such as typhoons and earthquakes. Teleoperated ground robots (including construction equipment) are effective tools for restoration work at disaster sites and other locations that are dangerous and inaccessible to humans. Using visual information obtained from various viewpoints by a drone can allow for more effective remote control of a teleoperated ground robot, making it easier for the robot to perform a task. We previously proposed and developed a remote-controlled drone system using only human eyes. However, the effectiveness of using this drone system during the remote control of a robot has never been verified. In this paper, as the first step in verifying the effectiveness of the remote-controlled drone system using only the eyes when remote-controlling a robot, we consider its effectiveness in a simple task based on the task times, subjects’ eye fatigue, and subjective evaluations of subjects. First, the previously proposed drone system is briefly described. Next, we describe an experiment in which a drone was controlled by the eyes using the drone system while a robot was controlled by hand, and an experiment in which both the drone and robot were controlled by hand without using the drone system. Based on the experimental results, we evaluate the effectiveness of the remote-controlled drone system using only the eyes when remote-controlling a robot.

Drone system remotely controlled by eyes

Drone system remotely controlled by eyes

Cite this article as:
Y. Kai, Y. Seki, Y. Wu, A. Blaquera, and T. Tanioka, “Drone System Remotely Controlled by Human Eyes: A Consideration of its Effectiveness When Remotely Controlling a Robot,” J. Robot. Mechatron., Vol.36 No.5, pp. 1055-1064, 2024.
Data files:
References
  1. [1] Cabinet Office, “White Paper on Disaster Management in Japan 2022,” Cabinet Office, Government of Japan, pp. 1-29, 2022.
  2. [2] Policy Bureau, Ministry of Land, Infrastructure, Transport and Tourism (MLIT), “Summary of the white paper on land, infrastructure, transport and tourism in Japan, 2023,” Ministry of Land, Infrastructure, Transport and Tourism (MLIT), Government of Japan, p. 29, 2023.
  3. [3] N. Michael, S. Shen, K. Mohta, Y. Mulgaonkar, V. Kumar, K. Nagatani, Y. Okada, S. Kiribayashi, K. Otake, K. Yoshida, K. Ohno, E. Takeuchi, and S. Tadokoro, “Collaborative mapping of an earthquake-damaged building via ground and aerial robots,” Field and Service Robotics, pp. 33-47, 2013. https://doi.org/10.1007/978-3-642-40686-7_3
  4. [4] S. Kiribayashi and K. Nagatani, “Development of collaborative system of multicopter and unmanned-ground-vehicle for exploration of the high altitude area in the building (foldable-arm-multicopter and tethered-landing-system),” Jpn. Soc. Mech. Eng., Vol.82, No.837, Article No.15-00563, 2016 (in Japanese). https://doi.org/10.1299/transjsme.15-00563
  5. [5] K. Nagatani, S. Kiribayashi, Y. Okada, K. Otake, K. Yoshida, S. Tadokoro, T. Nishimura, T. Yoshida, E. Koyanagi, M. Fukushima, and S. Kawatsuma, “Emergency response to the nuclear accident at the Fukushima daiichi nuclear power plants using mobile rescue robots,” J. of Field Robotics, Vol.30, No.1, pp. 44-63, 2013. https://doi.org/10.1002/rob.21439
  6. [6] W. Sun, S. Iwataki, R. Komatsu, H. Fujii, A. Yamashita, and H. Asama, “Simultaneous televisualization of construction machine and environment using body mounted cameras,” Proc. of the 2016 IEEE Int. Conf. on Robotics and Biomimetics (ROBIO2016), pp. 382-387, 2016. https://doi.org/10.1109/ROBIO.2016.7866352
  7. [7] A. Gawel, Y. Lin, T. Koutros, R. Siegwart, and C. Cadena, “Aerial-ground collaborative sensing: Third-person view for teleoperation,” Proc. of the 2018 IEEE Int. Symp. on Safety, Security, and Rescue Robotics (SSRR), 2018. https://doi.org/10.1109/SSRR.2018.8468657
  8. [8] S. Kiribayashi, K. Yakushigawa, S. Igarashi, and K. Nagatani, “Evaluation of aerial video for unmanned construction machine teleoperation assistance,” Proc. of the 2018 JSME Conf. on Robotics and Mechatronics, 2A1-K01(4), 2018 (in Japanese). https://doi.org/10.1299/jsmermd.2018.2A1-K01
  9. [9] S. Kiribayashi, K. Yakushigawa, and K. Nagatani, “Design and development of tether-powered multirotor micro unmanned aerial vehicle system for remote-controlled construction machine,” Field and Service Robotics, pp. 637-648, 2017. https://doi.org/10.1007/978-3-319-67361-5_41
  10. [10] X. Xiao, J. Dufek, and R. R. Murphy, “Tethered aerial visual assistance,” Field and Service Robotics, pp. 1-26, 2020. https://doi.org/10.48550/arXiv.2001.06347
  11. [11] X. Xiao, J. Dufek, and R. R. Murphy, “Autonomous visual assistance for robot operations using a tethered UAV,” Field and Service Robotics, pp. 15-29, 2021. https://doi.org/10.1007/978-981-15-9460-1_2
  12. [12] H. Yamada, N. Bando, K. Ootsubo, and Y. Hattori, “Teleoperated construction robot using visual support with drones,” J. Robot. Mechatron., Vol.30, No.3, pp. 406-415, 2018. https://doi.org/10.20965/jrm.2018.p0406
  13. [13] J. Dufek, X. Xiao, and R. R. Murphy, “Best viewpoints for external robots or sensors assisting other robots,” IEEE Trans. on Human-Machine System, Vol.51, No.4, pp. 324-334, 2021. https://doi.org/10.1109/THMS.2021.3090765
  14. [14] M. Kamezaki, M. Miyata, and S. Sugano, “Video presentation based on multiple-flying camera to provide continuous and complementary images for teleoperation,” Automation in Construction, Vol.159, Article No.105285, 2024. https://doi.org/10.1016/j.autcon.2024.105285
  15. [15] T. Ikeda, K. Niihara, S. Shiraki, S. Ueki, and H. Yamada, “Visual assist by drone-based wrap-around moving viewpoint display for teleoperated construction robot—Comparison of work accuracy and mental load by methods of moving viewpoints—,” J. Robot. Soc. Jpn., Vol.41, No.5, pp. 485-488, 2023 (in Japanese). https://doi.org/10.7210/jrsj.41.485
  16. [16] D. Saakes, V. Choudhary, D. Sakamoto, M. Inami, and T. Igarashi, “A teleoperating interface for ground vehicles using autonomous flying cameras,” Proc. of the 23rd Int. Conf. on Artificial Reality and Telexistence (ICAT), pp. 13-19, 2013. https://doi.org/10.1109/ICAT.2013.6728900
  17. [17] X. Xiao, J. Dufek, and R. Murphy, “Visual servoing for teleoperation using a tethered UAV,” Proc. of the 2017 IEEE Int. Symp. on Safety, Security and Rescue Robotics (SSRR), pp. 147-152, 2017. https://doi.org/10.1109/SSRR.2017.8088155
  18. [18] T. Ikeda, N. Bando, and H. Yamada, “Semi-automatic visual support system with drone for teleoperated construction robot,” J. Robot. Mechatron., Vol.33, No.2, pp. 313-321, 2021. https://doi.org/10.20965/jrm.2021.p0313
  19. [19] W. Dang, M. Li, D. Lv, X. Sun, and Z. Gao, “MHLCNN: Multi-harmonic linkage CNN model for SSVEP and SSMVEP signal classification,” IEEE Trans. on Circuits and Systems II: Express Briefs, Vol.69, No.1, pp. 244-248, 2022. https://doi.org/10.1109/TCSII.2021.3091803
  20. [20] M. McCoggle, S. Wilson, A. Rivera, R. Alba-Flores, and V. Soloiu, “Biosensors based controller for small unmanned aerial vehicle navigation,” Proc. of the IEEE Southeastcon 2022, pp. 380-384, 2022. https://doi.org/10.1109/SoutheastCon48659.2022.9764015
  21. [21] M. Yamamoto, T. Ikeda, and Y. Sasaki, “Real-time analog input device using breath pressure for the operation of powered wheelchair,” Proc. of the 2008 IEEE Int. Conf. on Robotics and Automation (ICRA), pp. 3914-3919, 2008. https://doi.org/10.1109/ROBOT.2008.4543812
  22. [22] C. Yapicioğlu, Z. Dokur, and T. Ölmez, “Voice command recognition for drone control by deep neural networks on embedded system,” Proc. of the 2021 8th Int. Conf. on Electrical and Electronics Engineering (ICEEE), pp. 65-72, 2021. https://doi.org/10.1109/ICEEE52452.2021.9415964
  23. [23] B. N. P. Kumar, A. Balasubramanyam, A. K. Patil, B. Chethana, and Y. H. Chai, “GazeGuide: An eye-gaze-guided active immersive UAV camera,” Applied Sciences, Vol.10, No.5, Article No.1668, 2020. https://doi.org/10.3390/app10051668
  24. [24] J. P. Hansen, A. Alapetite, I. S. MacKenzie, and E. Møllenbach, “The use of gaze to control drones,” Proc. of the Symp. on Eye Tracking Research and Applications, pp. 27-34, 2014. https://doi.org/10.1145/2578153.2578156
  25. [25] A. Kogawa, M. Onda, and Y. Kai, “Development of a remote-controlled drone system by using only eye movements: Design of a control screen considering operability and microsaccades,” J. Robot. Mechatron., Vol.33, No.2, pp. 301-312, 2021. https://doi.org/10.20965/jrm.2021.p0301
  26. [26] Y. Kai, Y. Seki, R. Suzuki, A. Kogawa, R. Tanioka, K. Osaka, Y. Zhao, and T. Tanioka, “Evaluation of a remote-controlled drone system for bedridden patients using their eyes based on clinical experiment,” Technologies, Vol.11, No.1, Article No.15, 2023. https://doi.org/10.3390/technologies11010015
  27. [27] Q. Gu, Q. Zhang, Y. Han, P. Li, Z. Gao, and M. Shen, “Microsaccades reflect attention shifts: A mini review of 20 years of microsaccade research,” Frontiers in Psychology, Article No.1364939, 2024. https://doi.org/10.3389/fpsyg.2024.1364939
  28. [28] N. Mizukami, “Measuring human fatigue,” RRR, Vol.65, No.4, pp. 26-29, 2008 (in Japanese).
  29. [29] X. Xiao, J. Dufek, M. Suhail, and R. Murphy, “Motion planning for a UAV with a straight or kinked tether,” Proc. of the 2018 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), pp. 8486-8492, 2018. https://doi.org/10.1109/IROS.2018.8594461

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Apr. 15, 2025