single-jc.php

JACIII Vol.21 No.2 pp. 310-320
doi: 10.20965/jaciii.2017.p0310
(2017)

Development Report:

Development of Web-Based Remote Speech-to-Text Interpretation System captiOnline

Daisuke Wakatsuki*,†, Nobuko Kato*, Takeaki Shionome**, Sumihiro Kawano*, Tomoyuki Nishioka*, and Ichiro Naito*

*Faculty of Industrial Technology, Tsukuba University of Technology
4-3-15 Amakubo, Tsukuba-shi, Ibaraki 305-8520, Japan

**Faculty of Science and Technology, Tohoku Bunka Gakuen University
6-45-1 Kunimi, Aoba-ku, Sendai 981-8551, Japan

Corresponding author

Received:
August 30, 2016
Accepted:
November 15, 2016
Online released:
March 15, 2017
Published:
March 20, 2017
Keywords:
remote speech-to-text interpretation, information support, hearing impaired, HTML5, WebSocket
Abstract
For ensuring information accessibility for hearing impaired, speech-to-text interpretation captioning is used. Usually, a text interpreter at the site interprets the speech of a speaker and enters a text into a PC. The text is then displayed through a projector or other devices to the users. For Japanese typing, in 1998, a method was developed for a pair of interpreters to collaboratively enter a sentence in which they used dedicated software to convert speech to text and present the resulting text. In the recent years, remote speech-to-text interpretation using the internet has been studied to overcome place and time restrictions of interpreters and users. However, conventional remote speech-to-text systems necessitated ensuring the availability of a network and preparing and operating devices for the system establishment and, hence, could be operated only by a limited number of organizations and groups. To overcome these limitations, we developed a system called captiOnline through which one can perform remote speech-to-text interpretation by simply accessing it through a web browser. In this paper, we explain how to implement captiOnline, evaluate its use in actual situations, and compare it with conventional systems. In addition, we show that captiOnline can realize remote speech-to-text interpretation in a simple configuration requiring less preparatory works.
Cite this article as:
D. Wakatsuki, N. Kato, T. Shionome, S. Kawano, T. Nishioka, and I. Naito, “Development of Web-Based Remote Speech-to-Text Interpretation System captiOnline,” J. Adv. Comput. Intell. Intell. Inform., Vol.21 No.2, pp. 310-320, 2017.
Data files:
References
  1. [1] D. Wakatsuki, “A Study on Remote Communication Support System for Hearing Impaired,” J. of the Japanese Society for Wellbeing Science and Assistive Technology, Vol.11, No.1, pp. 8-13, 2011 (in Japanese).
  2. [2] D. Wakatsuki, N. Kato, T. Shionome, S. Kawano, T. Nishioka, and I. Naito, “captiOnline: A Web-based Online Captioning and Speech-to-text Interpretation System for Deaf and Hard of Hearing,” Proc. of HCG Symposium 2014, pp. 517-522, 2014 (in Japanese).
  3. [3] D. Wakatsuki, “captiOnline,” http://capti.info.a.tsukuba-tech.ac.jp/. [Accessed Mar. 10, 2016] (in Japanese).
  4. [4] J.E. Preminger, and H. Levitt, “Computer-Assisted Remote Transcription (CART): A Tool To Aid People Who Are Deaf or Hard of Hearing in the Workplace,” Volta Review, Vol.99, No.4 pp. 219-230, 1997.
  5. [5] L.B. Elliot, M.S. Stinson, B.G. McKee, V.S. Everhart, and P.J. Francis, “College Students’ Perceptions of the C-Print Speech-to-Text Transcription System,” J. of Deaf Studies and Deaf Education, Vol.6, No.4, pp. 285-298, 2001.
  6. [6] M. Marschark, G. Leigh, P. Sapere, D. Burnham, C. Convertino, M. Stinson, H. Knoors, M.P.J Vervloed, and W. Noble, “Benefits of sign language interpreting and text alternatives to classroom learning by deaf students,” J. of Deaf Studies and Deaf Education, Vol.11, No.4, pp. 421-437, 2006.
  7. [7] L.B. Elliot and M.S. Stinson, “Supporting Deaf and Hard of Hearing Undergraduate Students in STEM Field Settings with Remote Speech-to-Text Services,” RIT/NTID JAM 2010 Poster, https://www.rit.edu/ntid/cprint/sites/rit.edu.ntid.cprint/filesnewline /presentations/06_2010_JAM_Poster.pdf. [Accessed Oct. 27, 2016].
  8. [8] NCRA’s Broadcast and CART Captioning Committee, “Guidelines for CART Captioners (Communication Access Realtime Translation Captioners),” National Court Reporters Association, pp. 21-24, http://www.ncra.org/files/GovernmentRelationsnewline /Guidelines%20for%20CART%20Captioners.pdf. [Accessed Oct. 27, 2016].
  9. [9] L.B. Elliot, M.S. Stinson, D. Easton, and J. Bourgeois, “College Students Leaning with C-Print’s Education Software and Automatic Speech Recognition,” American Ed. Research Assoc. Annual Meeting, 2008
  10. [10] W.S. Lasecki, C.D. Miller, A. Sadilek, A. Abumoussa, D. Borrello, R. Kushalinagar, and J.P. Bigham, “Real-Time Captioning by Groups of Non Experts,” Proc. of the 25th Annual ACM Symp. on User Interface Software and Technology, pp. 23-34, 2012.
  11. [11] W. S. Lasecki, C. D. Miller, and J. P. Bigham, “Warping Time for More Effective Real-Time Crowdsourcing,” Proc. of the International ACM Conference on Human Factors in Computing Systems (CHI 2013), pp. 2033-2036, 2013.
  12. [12] PEPNet-Japan, “PC note-taking skill up! teaching material collection. Let’s Try It! Combination Input”, M. Shirasawa, and A. Nakajima, eds, Research and Support Center, Tsukuba University of Technology, 2012 (in Japanese).
  13. [13] S. Kurita, “IPtalk,” http://www.geocities.jp/shigeaki_kurita/. [Accessed Mar. 10, 2016] (in Japanese).
  14. [14] S. Kurita, S. Kawano, and K. Kondo, “The remote computer assisted speech-to-text interpreter system for reducing operational costs,” Correspondences on Human Interface, Vol.15, No.8, pp. 13-20, 2013 (in Japanese)
  15. [15] S. Kawano, S. Miyoshi, K. Isoda, M. Hasuike, and M. Shirasawa, “ An Evaluation for Mobile Type Remote-captioning System in the Case of Plant Tour,” IEICE Technical Report, Vol.110, No.164, pp. 65-68, 2010 (in Japanese).
  16. [16] S. Miyoshi, S. Kawano, M. Shirasawa, K. Isoda, M. Shirasawa, M. Kobayashi, E. Ogasawara, M. Umehara, T. Kanazawa, S. Nakano, and T. Ifukube, “Proposal of Mobile Type Remote-captioning System for Deaf or Hard of Hearing People and Evaluation by Captionists,” J. of Life Support Engineering, Vol.22, No.4, pp. 11-16, 2010 (in Japanese).
  17. [17] S. Miyoshi, “Remote Communication Support System T-TAC Caption, Technical Manual Ver.3 R,” PEPNet-Japan, http://www.tsukuba-tech.ac.jp/ce/xoops/file/enkaku/tech-manual-ttaccaption.pdf. [Accessed Oct. 27, 2016], (in Japanese).
  18. [18] Information-technology Promotion Agency, Japan, “1.4.2 Vulnerability situation,” Information Security White Paper 2015, Information-technology Promotion Agency, Japan, pp. 47-58, 2015 (in Japanese).
  19. [19] Node.js Developers, Joyent, Github Node.js Contributors, “node.js,” https://nodejs.org/. [Accessed Mar. 10, 2016].
  20. [20] Guillermo Rauch, Github Socket.IO Contributors, “Socket.IO,” http://socket.io/. [Accessed Mar. 10, 2016].
  21. [21] Telenor Digital, “appear.in,” https://appear.in/. [Accessed Mar. 10, 2016].
  22. [22] W3C, Google Inc, “WebRTC,” http://www.webrtc.org/. [Accessed Mar. 10, 2016].

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Apr. 18, 2024