Toward A Tactile Language For Human-Robot Interaction: Two Studies Of Tacton Learning And Performance
Keywords
human-robot interaction; multitasking; tactile communication; tactons; workload
Abstract
Objective: Two experiments were performed to investigate the feasibility for robot-to-human communication of a tactile language using a lexicon of standardized tactons (tactile icons) within a sentence. Background: Improvements in autonomous systems technology and a growing demand within military operations are spurring interest in communication via vibrotactile displays. Tactile communication may become an important element of humanrobot interaction (HRI), but it requires the development of messaging capabilities approaching the communication power of the speech and visual signals used in the military. Method: In Experiment 1 (N = 38), we trained participants to identify sets of directional, dynamic, and static tactons and tested performance and workload following training. In Experiment 2 (N = 76), we introduced an extended training procedure and tested participants ability to correctly identify two-tacton phrases. We also investigated the impact of multitasking on performance and workload. Individual difference factors were assessed. Results: Experiment 1 showed that participants found dynamic and static tactons difficult to learn, but the enhanced training procedure in Experiment 2 produced competency in performance for all tacton categories. Participants in the latter study also performed well on twotacton phrases and when multitasking. However, some deficits in performance and elevation of workload were observed. Spatial ability predicted some aspects of performance in both studies. Conclusions: Participants may be trained to identify both single tactons and tacton phrases, demonstrating the feasibility of developing a tactile language for HRI. Application: Tactile communication may be incorporated into multi-modal communication systems for HRI. It also has potential for humanhuman communication in challenging environments.
Publication Date
5-23-2015
Publication Title
Human Factors
Volume
57
Issue
3
Number of Pages
471-490
Document Type
Article
Personal Identifier
scopus
DOI Link
https://doi.org/10.1177/0018720814548063
Copyright Status
Unknown
Socpus ID
84928334062 (Scopus)
Source API URL
https://api.elsevier.com/content/abstract/scopus_id/84928334062
STARS Citation
Barber, Daniel J.; Reinerman-Jones, Lauren E.; and Matthews, Gerald, "Toward A Tactile Language For Human-Robot Interaction: Two Studies Of Tacton Learning And Performance" (2015). Scopus Export 2015-2019. 899.
https://stars.library.ucf.edu/scopus2015/899