The future of robots: A psychological approach to human-automation interaction

author avatar

20 Apr, 2020

BagoGames via Flickr

BagoGames via Flickr

It’s called the uncanny valley. Those who are fans of the HBO show "Westworld" or who have seen the movie "Ex Machina" may already be familiar with the phenomenon.

It’s called the uncanny valley. Those who are fans of the HBO show "Westworld" or who have seen the movie "Ex Machina" may already be familiar with the phenomenon. But for those who are not, it’s essentially the idea that humans are comfortable with robots who have humanoid features, but become very uncomfortable when the robot looks almost but not exactly like a human.

For Dr. Nathan Tenhundfeld, however, the uncanny valley is just one of many factors he must take into account while researching human-automation interaction as an assistant professor in the Department of Psychology at The University of Alabama in Huntsville (UAH).

"We’re at a unique point with the development of the technology where automated systems or platforms are no longer a tool but a teammate that is incorporated into our day-to-day experiences," he says. "So we’re looking at commercial platforms that offer the same systems but in different forms to see whether a certain appearance or characteristic affects the user and in what way."

A student in the ATTAC Lab takes part in a flight simulation. Michael Mercier | UAH 

Take for example, the recent push by the U.S. Department of Defense to incorporate automation into warfighting. As a concept, it makes sense: the more robots we have fighting wars, the less cost there is to human life. But in practice, it’s a little more complex. What should a warfighting robot look like? A person? A machine?

To answer these questions, Dr. Tenhundfeld has partnered with Dr. Beth Phillips who works at the U.S. Air Force Academy, where he conducted research as a postdoctoral fellow, to use "a massive database of robots" so that they determine how various components of things like motion might affect the perception of a robot. "We want to know things like, does a robot with wheels or a track fit better with our expectation of what we should be sending to war versus a humanoid robot?" he says. "And, does having a face on the robot affect whether we want to put it in harm’s way?"

Even if there were easy answers – which there aren’t – there’s another equally important factor to consider beyond the robot’s user interface: trust. For a robot to be effective, the user must trust the information that it is providing. To explain, Dr. Tenhundfeld points to research he conducted on the Tesla Model X while at the Academy. Looking at the car’s autoparking feature specifically, he and his team wanted to determine the user’s willingness to let the car complete its task as a function of their risk-taking preference or confidence in their abilities.

"The data suggest automated vehicles tend to be safer than humans, but humans don’t like to relinquish control," he says with a laugh. "So we had this pattern where there were high intervention rates at first, but as they developed trust in the system – after it wasn’t so novel and it started to meet their expectations – they began to trust it more and the intervention rates went down."

A student conducts research in the ATTAC Lab with a robot named Cozmo.  Michael Mercier | UAH 

The flip side of that coin, however, is the potential for empathy in, or attachment to, a particular automated system users may have developed trust in. To illustrate this concept, he recounts a case study of explosive-ordinance disposal teams who employ robots to safely blow up bombs. "When they have to send the robots back to get repaired, they have an issue when they’re given a different robot," he says. "So they’ve placed this trust in a specific robot even though the intelligence/capability is the same across all of the robots."

And lest it start to sound like there is already more than enough for Dr. Tenhundfeld to factor in, there is also situational trust, which sits somewhere between trust and overtrust. In this scenario, a user may develop a certain level of trust as a whole over time, but then realize they don’t trust some aspects as much as much as others. "Say I have an automated system, or robot, providing intelligence in a mission-planning environment, and it screws that up," he says. "I might not trust it in a different environment, such as on the battlefield, even though it has a different physical embodiment for use in that environment, and may be distinctly capable on the battlefield."

In short, the increasingly digital nature of our world introduces a seemingly endless list of considerations when it comes to ensuring automated systems can successfully meet human needs – all of which Dr. Tenhundfeld must take into account with the research he is doing in his Advanced Teaming, Technology, Automation, and Computing Lab, or ATTAC Lab. But given UAH’s role as an academic partner to this emerging industry, it’s a challenge that he and his fellow researchers have embraced. "Businesses are focused on being first to market with a product," he says. "We help them improve the product so that it works well for the user."

More about Diana LaChance

Diana LaChance earned a bachelor’s degree from the College of William & Mary in Williamsburg, Virginia, and a master’s degree from the University of Florida in Gainesville, Florida. She began her professional career in Northern Virginia as an editor in the book publishing industry before transitioni...

Wevolver 2022