JRM Vol.21 No.2 pp. 267-276
doi: 10.20965/jrm.2009.p0267


Functions of Mobile-Robot Step-On Interface

Takafumi Matsumaru and Kosuke Akai

Graduate School of Engineering, Shizuoka University,
3-5-1 Johoku, Naka, Hamamatsu, 432-8561 Japan

September 30, 2008
January 29, 2009
April 20, 2009
step-on interface, mobile robot, projector, range sensor, human interaction

To improve HFAMRO-1 mobile robot maneuverability and safety, we added a step-on interface (SOI) to direct robotic or mechatronic tasks and operations (HFAMRO: “human-friendly amusing” mobile robot). To do so, a projector displays a direction screen on a floor or other surface, and an operator specifies a button showing the selected movement by stepping or pointing. We modified the direction screen so that among buttons displayed on two lines, stepped-on buttons directing movement are displayed only on the lower line. We also shortened retention time and had selected movement executed only when the foot was removed from the stepped-on button. The robot has 2 SOIs and multiple projection screens, and can be controlled from either direction for the same function. We synchronized direction and preliminary-announcement screens to inform passers-by which way the robot would move. Using range scanner data, the robot distinguishes feet from other objects based on size and autonomous movement fusion control to avoid obstacles is implemented.

Cite this article as:
Takafumi Matsumaru and Kosuke Akai, “Functions of Mobile-Robot Step-On Interface,” J. Robot. Mechatron., Vol.21, No.2, pp. 267-276, 2009.
Data files:
  1. [1] S. Honda, K. Akai, M.T. Lin, and T. Matsumaru, “Teleoperation of Human-Friendly Robot (41st report) -Examination of hardware constitution of HFAMRO-,” SICE-Chubu Shizuoka 2006, 2D1, 2006.
  2. [2] K. Akai, S. Honda, M. T.Lin, and T. Matsumaru, “Teleoperation of Human-Friendly Robot (43th report) -Examination of software constitution of HFAMRO-,” SICE-Chubu Shizuoka 2006, 2D3, 2006.
  3. [3] K. Ishikawa, H. Aoyama, J. Seki, M. Okamura, S. Ishimura, K. Ozaki, K. Yokota, and S. Yamamoto, “Practical Design and Use of Autonomous Robot for Outdoor Cleaning,” J. RSJ, Vol.25, No.6, pp. 802-812, 2007.
  4. [4] T. Sakai, D. Nizhimura, H. Uematsu, R. Murai, T. Nakahara, and Y. Kitano, “Autonomous Cleaning Robot SuiPPi,” 2006 JSME Conf. on Robotics and Mechatronics, 1A1-E02, 2006.
  5. [5] T.Kouno and S.Kanda, “Robot for carrying food trays to the aged and disabled,” J. RSJ, Vol.16, No.3, pp. 317-320, 1998.
  6. [6] T. Sakai, H. Nakajima, D. Nishimura, H. Uematsu, and Y. Kitano, “Autonomous Mobile Robot System for Delivery in Hospital,” Panasonic Electric Works Technical Report, Vol.53, No.2, pp. 62-67, 2005.
  7. [7] Y. Shimosasa, K. Wakabayashi, T. Ariki, T. Omachi, D. Suetomi, M. Sugiura, and K. Kotani, “Development and Safety Policy of the Outdoor Security Robot ALSOK Guardrobo i,” J. RSJ, Vol.24, No.2, pp. 156-158, 2006.
  8. [8] S. Kanda, Y. Murase, K. Okabayashi, and T. Asada, “Development of ”enon” Service Robot,” Fujitsu, Vol.57, No.3, pp. 307-313, 2006.
  9. [9] Y. Hosoda, “An approach to development of a human-symbiotic robot,” J. RSJ, Vol.24, No.3, pp. 296-299, 2006.
  10. [10] K. Matsukuma, “Service robot applications,” Yasukawa technical review, Vol.72, No.2, pp. 99-103, 2008.
  11. [11] K. Onishi, “wakamaru, the robot for your home,” J. JSME, Vol.109, No.1051, pp. 448-449, 2006.
  12. [12] H. Mizoguchi, T. Yoshimi, and N. Matsuhira, “Person Following Robot, ApriAttenda,” J. RSJ, Vol.24, No.2, p. 173, 2006.
  13. [13] T. Umezaki, “Playing tag robot ASKA,” J. RSJ, Vol.24, No.2, p. 202, 2006.
  14. [14] A. Holzinger, “Finger Instead of Mouse: Touch Screens as a means of enhancing Universal Access,” in Universal Access, Theoretical Perspectives, Practice, and Experience N. Carbonell, C. Stephanidis (Eds), Springer, pp. 387-397, 2003.
  15. [15] S. Kanda, M. Yuichi, N. Sawasaki, and T. Asada, “Development of the Service Robot ”enon',” J. RSJ, Vol.24, No.3, pp. 288-291, 2006. bibitemOffice-Conversant H. Asoh, I. Hara, and T. Matsui, “Dynamic Structured Multi-Agent Architecture for Controlling Office-Conversant Mobile Robot,” Proc. IEEE ICRA'98, pp. 1552-1557, 1998.
  16. [16] K. Nakadai, D. Matsuura, H.G. Okuno, and H. Kitano, “Applying scattering theory to robot audition system: Robust sound source localization and extraction,” Proc. IEEE/RSJ IROS 2003, Vol.2, pp. 1157-1162, 2003.
  17. [17] T. Shibata, Y. Matsumoto, T. Kuwahara, M. Inaba, and H. Inoue, “Hyper Scooter: A mobile Robot sharing visual information with a human,” Proc. 1995 IEEE ICRA, Vol.1, pp. 1074-1079, 1995.
  18. [18] S.Esaki, Y.Ebisawa, A.Sugioka, and M.Konishi, “Quick menu selection using eye blink for eye-slaved nonverbalcommunicator with video-based eye-gaze detection,” Proc. 19th Annual Int. Conf. IEEE EMBS 1997, Vol.5, pp. 2322-2325, 1997.
  19. [19] Center for Art and Media in Karlsruhe, Small Fish,
  20. [20] Solidray Co., Taptalk2,
  21. [21] H.E. Korth, “Method and device for optical input of commands or data,” EP patent 0554492, 1992.02.07.
  22. [22] S. Hiura, K. Tojo, and S. Inokuchi, “3-D Tele-direction Interface using Video Projector,” ACM SIGGRAPH 2003 Sketches & Applications, 2003.
  23. [23] T. Machino, Y. Nanjo, Y. Yanagihara, H. Kawata, S. Iwaki, and K. Shimokura, “Robot-augmented communication: a remote-collaboration system based on a shared field of view in real space,” IEEE/RSJ IROS 2005, pp. 2203-2209, 2005.
  24. [24] R. Raskar, P. Beardsley, J.v. Baar, Y. Wang, P. Dietz, J. Lee, D. Leigh, and T. Willwacher, “RFIG lamps: interacting with a self-describing world via photosensing wireless tags and projectors,” ACM Trans. on Graphics (TOG), Vol.23, No.3, pp. 406-415, 2003.
  25. [25] M. Terashima and S. Sakane, “A Human-Robot Interface Using an Extended Digital Desk,” Proc. IEEE ICRA'99, pp. 2874-2880, 1999.
  26. [26] S. Sato and S. Sakane, “A human-robot interface using an interactive hand pointer that projects a mark in the real work space,” Proc. IEEE ICRA'00, pp. 589-595, 2000.
  27. [27] M. Yamashita and S. Sakane, “Adaptive Annotation Using a Human-Robot Interface System PARTNER,” Proc. IEEE 2001 ICRA, pp. 2661-2667, 2001.
  28. [28] Y. Wakita, S. Hirai, T. Suehiro, T. Hori, and K. Fujiwara, “Information Sharing via Projection Function for Coexistence of Robot and Human,” Autonomous Robots, Vol.10, No.3, pp. 267-277, 2001.
  29. [29] T. Matsumaru, “Mobile Robot with Preliminary-announcement and Display Function of Following Motion using Projection Equipment,” IEEE RO-MAN 06, pp. 443-450, 2006.
  30. [30] S. Honda, K. Akai, M.T. Lin, and T. Matsumaru, “Teleoperation of Human-Friendly Robot (42nd report) -Examination of foot position sensing with a URG sensor-,” SICE-Chubu Shizuoka 2006, 2D2, 2006.
  31. [31] National Institute of Advanced Industrial Science and Technology (AIST) and National Institute of Technology and Evaluation (NITE), Anthropometric Database 1997-98, 2001.
  32. [32] O. Hikosaka, “Neural Mechanism of Attention,” in M. Ito, Y. Anzai, M. Kawato, S. Ichikawa, H. Nakashima, and K. Hashida (Eds.), “Cognitive Science Vol.9 Attention and Consciousness,” Iwanami Shoten, pp. 90-168, 1996.
  33. [33] K. Goldberg, B. Chen, R. Solomon, S. Bui, B. Farzin, J. Heitler, D. Poon, and G. Smith, “Collaborative teleoperation via the Internet,” Proc. IEEE ICRA '00, Vol.2, pp. 2019-2024, 2000.
  34. [34] T. Matsumaru, “Evaluation Experiment in Simulated Interactive Situation between People and Mobile Robot with Preliminary-Announcement and Indication Function of Upcoming Operation,” Trans. Human Interface Society, Vol.10, No.1, pp. 11-20, 2008.
  35. [35] T. Matsumaru, S. Kudo, H. Endo, and T. Ito, “Examination on a Software Simulation of the Method and Effect of Preliminary-announcement and Display of Human-friendly Robot's Following Action,” Trans. SICE, Vol.40, No.2, pp. 189-198, 2004.
  36. [36] T. Matsumaru, K. Hagiwara, and T. Ito, “Examination on The Combination Control of Manual Operation and Autonomous Motion for Teleoperation of Mobile Robot Using a Software Simulation,” Trans. SICE, Vol.41, No.2, pp. 157-166, 2005.

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Mar. 05, 2021