advertisement
Science News
from research organizations

Human-like robots may be perceived as having mental states

Some people perceive robots that display emotions as intentional agents, study finds

Date:
July 7, 2022
Source:
American Psychological Association
Summary:
When robots appear to engage with people and display human-like emotions, people may perceive them as capable of 'thinking,' or acting on their own beliefs and desires rather than their programs, according to new research.
Share:
advertisement

FULL STORY

When robots appear to engage with people and display human-like emotions, people may perceive them as capable of "thinking," or acting on their own beliefs and desires rather than their programs, according to research published by the American Psychological Association.

"The relationship between anthropomorphic shape, human-like behavior and the tendency to attribute independent thought and intentional behavior to robots is yet to be understood," said study author Agnieszka Wykowska, PhD, a principal investigator at the Italian Institute of Technology. "As artificial intelligence increasingly becomes a part of our lives, it is important to understand how interacting with a robot that displays human-like behaviors might induce higher likelihood of attribution of intentional agency to the robot."

The research was published in the journalTechnology, Mind, and Behavior.

Across three experiments involving 119 participants, researchers examined how individuals would perceive a human-like robot, the iCub, after socializing with it and watching videos together. Before and after interacting with the robot, participants completed a questionnaire that showed them pictures of the robot in different situations and asked them to choose whether the robot's motivation in each situation was mechanical or intentional. For example, participants viewed three photos depicting the robot selecting a tool and then chose whether the robot "grasped the closest object" or "was fascinated by tool use."

在前两个实验中,研究人员雷莫tely controlled iCub's actions so it would behave gregariously, greeting participants, introducing itself and asking for the participants' names. Cameras in the robot's eyes were also able to recognize participants' faces and maintain eye contact. The participants then watched three short documentary videos with the robot, which was programmed to respond to the videos with sounds and facial expressions of sadness, awe or happiness.

In the third experiment, the researchers programmed iCub to behave more like a machine while it watched videos with the participants. The cameras in the robot's eyes were deactivated so it could not maintain eye contact and it only spoke recorded sentences to the participants about the calibration process it was undergoing. All emotional reactions to the videos were replaced with a "beep" and repetitive movements of its torso, head and neck.

The researchers found that participants who watched videos with the human-like robot were more likely to rate the robot's actions as intentional, rather than programmed, while those who only interacted with the machine-like robot were not. This shows that mere exposure to a human-like robot is not enough to make people believe it is capable of thoughts and emotions. It is human-like behavior that might be crucial for being perceived as an intentional agent.

According to Wykowska, these findings show that people might be more likely to believe artificial intelligence is capable of independent thought when it creates the impression that it can behave just like humans. This could inform the design of social robots of the future, she said.

“社会联系与机器人可能是有益的some contexts, like with socially assistive robots. For example, in elderly care, social bonding with robots might induce a higher degree of compliance with respect to following recommendations regarding taking medication," Wykowska said. "Determining contexts in which social bonding and attribution of intentionality is beneficial for the well-being of humans is the next step of research in this area."

advertisement

Story Source:

Materialsprovided byAmerican Psychological Association.注意:内容可能被编辑风格d length.


Journal Reference:

  1. Serena Marchesi, Davide De Tommaso, Jairo Perez-Osorio and Agnieszka Wykowska.Belief in Sharing the Same Phenomenological Experience Increases the Likelihood of Adopting the Intentional Stance Toward a Humanoid Robot.Technology, Mind, and Behavior, 2022 DOI:10.1037/tmb0000072

Cite This Page:

American Psychological Association. "Human-like robots may be perceived as having mental states: Some people perceive robots that display emotions as intentional agents, study finds." ScienceDaily. ScienceDaily, 7 July 2022. .
American Psychological Association. (2022, July 7). Human-like robots may be perceived as having mental states: Some people perceive robots that display emotions as intentional agents, study finds.ScienceDaily. Retrieved July 1, 2023 from www.koonmotors.com/releases/2022/07/220707100907.htm
American Psychological Association. "Human-like robots may be perceived as having mental states: Some people perceive robots that display emotions as intentional agents, study finds." ScienceDaily. www.koonmotors.com/releases/2022/07/220707100907.htm (accessed July 1, 2023).

Explore More
from ScienceDaily

RELATED STORIES

advertisement