Thousands of exciting and novel applications of augmented reality and robotic technologies have emerged in recent years, but the potential for networking these two technologies and using them in conjunction with each other has gone relatively unexplored. However, two papers published by the ATLAS Iron Lab last week for the ACM/IEEE International Conference on Human Robot Interaction in Chicago open the door to this promising area of research, paving the way for more seamless integration of robots in modern life.
听
Recognizing the value of their innovative work, conference organizers awarded the IRON Lab teams best paper and runner-up best paper in the design category.听Assistant Professor Dan Szafir, who directs the IRON Lab, explains that both papers examine the potential for transmitting real-time visual information from drones to people with AR. In the first study, research participants completing an assembly task while sharing a workspace with a drone听were more efficient when informed of the drone's听flightpath using AR, versus tracking its path without assistance.听In the second study, drone photography proved safer and more accurate when a drone camera鈥檚 field of view was streamed to operators' AR displays instead of tablet screens, as is the norm today.
听
To conduct the first study, researchers set up an environment similar to a small warehouse, where participants were assigned the task of stringing beads in a specific color order, requiring them to move between six assembly stations, remaining at a safe distance from the drone at all times. Their goal was to assemble as many beaded strings as possible in eight minutes. When the drone approached, the participants听had to stop work and move to a different workstation.
听
Results found that when the drone鈥檚 imminent flightpath was communicated with AR, participants were more efficient. Furthermore, the study evaluated tradeoffs in a variety of different graphical approaches to communicating the drone鈥檚 flightpath, which may help guide the design of future AR interfaces.
听
The second study found that AR technology helped drone operators take photos more safely and with more accuracy. Using a drone-mounted camera, research subjects were asked to photograph framed targets on a wall as quickly and precisely as possible. The drone camera鈥檚 field of view was visible to operators using a handheld tablet and using AR, in a variety of graphical configurations.
听
Results were judged by how fast subjects completed the task, the accuracy of their photos and the number of times the drones听crashed. Once again, the study found AR significantly improved performance, increasing accuracy and reducing the number of crashes, with some AR graphical approaches proving more effective than others.
As the world moves towards integrating humans and robots in the baby直播app, effective collaboration depends on the ability of team members to rapidly understand and predict a robot鈥檚 behavior, something that human workers do through facial expressions, gestures and speech,鈥 says Szafir, who directs the IRON Lab. 鈥淗uman workers want to know explicitly when and where their robot coworker intends to move next, and they perform best when they can anticipate those movements. We are excited to be exploring how to leverage augmented reality to communicate this information in new and more effective ways.鈥
听
Taking place over a 12-month period, the two studies听Szafir supervised听were conducted by PhD students Michael Walker and Hooman Hedayati, along with master鈥檚 student Jennifer Lee. After being launched in January 2016 and听Szafir making the Forbes听"30 Under 30: Science" list in January 2017, the IRON Lab's latest commendations from the world's preeminent HRI conference sets expectations high for this ambitious and growing group of researchers.
by Michael Walker, Hooman Hedayati, Jennifer Lee, and听Daniel Szafir听(Best Paper鈥擠esign,听ACM/IEEE International Conference on Human Robot Interaction, 2018)
听by Hooman Hedayati, Michael Walker and Daniel Szafir听(Runner-Up Best Paper鈥擠esign,听ACM/IEEE International Conference on Human Robot Interaction, 2018)