People’s trust of ‘fake faces’ could make them more susceptible to fake news and have dire consequences with trust
Participants in new research conducted by Royal Holloway and the Warburg Institute, both University of London, were likely to trust artificially generated faces, known as generative adversarial network (GAN) faces, than real ones, making them more susceptible to misinformation and creating a cultural visual landscape towards a more deceptive society.
The study, published in the journal iScience and titled “On the Realness of People Who Do Not Exist: The Social Processing of Artificial Faces,” lab-led by Professor Manos Tsakiris from the Department of Psychology at Royal Holloway, and study-led by Dr. Raffaele Tucciarelli at the Warburg Institute, School of Advanced Study, showed participants a series of faces that either belonged to real people or were artificially generated (AI).
Participants were asked to judge whether the faces were real or not and, intriguingly, GAN faces were more likely to be perceived as real over those of actual people.
The findings highlight the recent advances in the technology used to generate artificial images, with an impressive acceleration over the last decade of GAN technology—resulting in faces looking incredibly realistic today.
GAN faces are realistic-looking images of non-existing people that are computer generated. The people depicted in these images do not exist and never existed. They are increasingly used in marketing, journalism and social media, but also for malicious purposes, such as political propaganda and espionage.
The findings are important when society considers the social implications: from mainstream news to social media, edited photos to deep fake videos, humans to bots, and from alternative facts to fake news—people have to judge the reliability of the information they see.
The ever-increasing use of fake images and videos is shifting the cultural visual landscape from being primarily truthful to being potentially deceptive. The question of who to trust becomes particularly relevant.
For that reason, the same team of researchers went on to show, in a separate study, that participants are more likely to trust the information conveyed by faces they had previously judged to be real, even if they were artificially generated.
Professor Manos Tsakiris, professor of psychology at Royal Holloway—and also director of the Centre for the Politics of Feelings at the School of Advanced Study lab-led the study, said, “Many have argued that one of biggest casualties to artificial intelligence will be the erosion of trust in what we see and hear.
“As we show in our study, the realness that people project onto artificial faces makes them more likely to be trusted as informational sources, but later, when people realize there are AI images out there, their whole trust of any information given to them is drastically reduced.
“This could lead to people disengaging with messages given to them in the future as they do not know who or what to trust. Educating audiences about such technologies and advancing their digital literacy may make us less gullible but at the same time it may make us, in general, less trusting.”
Dr. Raffaele Tucciarelli, who has recently completed his research at the Warburg Institute, added, “Our results show that people’s biases in perceiving artificial faces as real can then make them more gullible to information conveyed by such artificial agents.
“As we live in times when more than ever before in human history, we are asked to judge the realness, truthfulness and trustworthiness of our social world, and what information we decide to trust can have far-reaching consequences.”
To further understand the role that knowledge of the existence of such technology may play in people’s behavior, researchers then conducted a third study where half of the participants were informed about the existence of GAN faces while the other half were kept in the dark.
The results showed that informed participants showed overall lower levels of trust in their virtual interaction with all faces presented to them, independently of whether these were real or not.
The subversive and ubiquitous use of technologies that can generate realistic-yet-fake photos and videos can erode trust in the main informational sources that we use, such as social media, with far-reaching consequences for our societies.
Raffaele Tucciarelli et al, On the realness of people who do not exist: The social processing of artificial faces, iScience (2022). DOI: 10.1016/j.isci.2022.105441
Royal Holloway, University of London
People’s trust of ‘fake faces’ could make them more susceptible to fake news and have dire consequences with trust (2022, December 8)