NASA Logo

NTRS

NTRS - NASA Technical Reports Server

Back to Results
Human-Robot Interaction Directed Research ProjectHuman-robot interaction (HRI) is about understanding and shaping the interactions between humans and robots (Goodrich & Schultz, 2007). It is important to evaluate how the design of interfaces and command modalities affect the human s ability to perform tasks accurately, efficiently, and effectively (Crandall, Goodrich, Olsen Jr., & Nielsen, 2005) It is also critical to evaluate the effects of human-robot interfaces and command modalities on operator mental workload (Sheridan, 1992) and situation awareness (Endsley, Bolt , & Jones, 2003). By understanding the effects of interface design on human performance, workload, and situation awareness, interfaces can be developed that support the human in performing tasks with minimal errors and with appropriate interaction time and effort. Thus, the results of research on human-robot interfaces have direct implications for design. Because the factors associated with interfaces and command modalities in HRI are too numerous to address in 3 years of research, the proposed research concentrates on three manageable areas applicable to National Aeronautics and Space Administration (NASA) robot systems. These topic areas emerged from the Fiscal Year (FY) 2011 work that included extensive literature reviews and observations of NASA systems. The three topic areas are: 1) video overlays, 2) camera views, and 3) command modalities. Each area is described in detail below, along with relevance to existing NASA human-robot systems. In addition to studies in these three topic areas, a workshop is proposed for FY12. The workshop will bring together experts in human-robot interaction and robotics to discuss the state of the practice as applicable to research in space robotics. Studies proposed in the area of video overlays consider two factors in the implementation of augmented reality (AR) for operator displays during teleoperation. The first of these factors is the type of navigational guidance provided by AR symbology. In the proposed studies, participants performance during teleoperation of a robot arm will be compared when they are provided with command-guidance symbology (that is, directing the operator what commands to make) or situation-guidance symbology (that is, providing natural cues so that the operator can infer what commands to make). The second factor for AR symbology is the effects of overlays that are either superimposed or integrated into the external view of the world. A study is proposed in which the effects of superimposed and integrated overlays on operator task performance during teleoperated driving tasks are compared
Document ID
20110023273
Acquisition Source
Johnson Space Center
Document Type
Conference Paper
Authors
Rochlis, Jennifer
(NASA Johnson Space Center Houston, TX, United States)
Ezer, Neta
(Futron Corp. Houston, TX, United States)
Sandor, Aniko
(Lockheed Martin Corp. Houston, TX, United States)
Date Acquired
August 25, 2013
Publication Date
January 1, 2011
Subject Category
Cybernetics, Artificial Intelligence And Robotics
Report/Patent Number
JSC-CN-25271
Meeting Information
Meeting: Human Research Program Investigators'' Workshop
Location: Houston, TX
Country: United States
Start Date: February 14, 2012
End Date: February 16, 2012
Sponsors: NASA Headquarters
Distribution Limits
Public
Copyright
Public Use Permitted.
No Preview Available