英语轻松读发新版了,欢迎下载、更新

We gave an AI a Rorschach test. What it saw in the inkblots offers a window into the human mind

2025-02-24 14:00:00 英文原文

By Chris Stokel-Walker profile image

Estudio Santa Rita Illustration of a silhouetted person looking at a digital Rorschach inkblot (Credit: Estudio Santa Rita)Estudio Santa Rita

Rorschach tests play with the human imagination and our mind's ability to impart meaning onto the world around us – but what does AI see in them?

For more than a century, the Rorschach inkblot test has been widely used as a window into people's personality.

Even if you haven't done one yourself, you'll recognise the mirrored ink smudges that form ambiguous shapes on cards. Developed by Swiss psychiatrist Hermann Rorschach in 1921, the test involves showing a number of inkblots to subjects, then asking them to say what they see. The images are deliberately inscrutable and open to interpretation.

For decades they were popular among psychologists as a way of understanding a person's psyche from the creatures, objects or scenes they perceive in the shapes. It relies upon a phenomenon known as pareidolia, which is a tendency to find meaningful interpretations in things where there is none. It is the same reason why people see faces or animals in cloud formations or on the surface of the Moon.

You might see a butterfly in the image while someone else might see a skull. According to proponents of the test, both interpretations shed light on how you think.

"When a person interprets a Rorschach image, they unconsciously project elements of their psyche such as fears, desires, and cognitive biases," says Barbara Santini, a London-based psychologist, who uses the Rorschach test with her clients. "The test works because human vision isn't passive, but a meaning-making process shaped by personal experience."

But what happens if you have no personal experience, or you offer it to a "brain" that works in an entirely different way? What might an artificial intelligence (AI) algorithm recognise in these inkblots? And what would their answers tell us about the human mind?

Estudio Santa Rita Rorschach inkblots take advantage of a phenomenon known as pareidolia (Credit: Estudio Santa Rita)Estudio Santa Rita

Rorschach inkblots take advantage of a phenomenon known as pareidolia (Credit: Estudio Santa Rita)

Bat or moth?

Recent developments in the capabilities of AI models like OpenAI's ChatGPT means it's now possible to put that to the test. So-called "multimodal models" can accept image-based inputs and provide responses as easily as they are to interact with text-based prompts thanks to their machine vision capabilities.

We gave ChatGPT five of the 10 common images used in an inkblot test to see how it would interpret the images. But the AI system's training data – information that helps establish its "knowledge", makes it trickier to do so.

AI v the Mind

This article is part of AI v the Mind, a series that aims to explore the limits of cutting-edge AI and learn a little about how our own brains work along the way. With some expert help, each article pits different AI tools against the human mind, asking probing questions designed to test the limits of intelligence. Can a machine write a better joke than a professional comedian, or unpick a moral conundrum more elegantly than a philosopher? We hope to find out. 

Showing the AI model the first inkblot card from an online database of inkblots, which is commonly interpreted by humans as a bat, butterfly or moth, highlighted just how literal AI vision algorithms can be.

"This image is a Rorschach inkblot, often used in psychological assessments to explore perception and interpretation," the chatbot replied. "It is designed to be ambiguous so that each person might see something different depending on their experiences, emotions, and imagination."

Pretence spoiled, it then went on to explain what it saw in the same answer, without me having to prompt it further. "For me, it resembles something symmetrical, possibly two animals or figures facing each other, or a single entity with wings outstretched," the chatbot responded. "The beauty of these inkblots is that they invite individual interpretations!"

The chatbot then asked me back what I saw in the image – a question I chose to ignore, instead pressing it to see which of the two resemblances it had suggested it thought was most fitting. "Looking closely, I'd say it most resembles a single entity with wings outstretched – perhaps a bat or a moth, with its wings open symmetrically," the chatbot wrote. "The central area could be interpreted as the body, and the details on the sides evoke the texture or shape of wings."

"I think that's just rehearsing what it knows of the specific test," says Coen Dekker, a Dutch software developer who a decade ago asked a neural network, an early version of the technology underpinning AI chatbots today, to complete a Rorschach test. (Dekker's experiment was significantly more simplified because of the technical constraints of the neural networks of the time.) 

If an AI's response resembles a human's, it's not because it sees the same thing but it's because its training data mirrors our collective visual culture – Barbara Santini

"I believe it mainly identifies patterns, shapes, and textures within the blots, and then compares these features to a vast dataset of human responses to generate its interpretation of what it sees in the inkblots," says Ieva Kubiliute, a London-based psychologist.

The essence of humanity

"ChatGPT provides interesting, human-like responses, but it is likely not genuinely thinking – instead, it appears to be scanning an online dataset," says Chandril Ghosh, lecturer in psychology at the University of Kent, in the UK, specialising in mental health, AI and treatment. It's similar to the way a musician who has never experienced heartbreak can still write a song that pulls at your heartstrings, simply by analysing how other similar songs like it are structured. Or as Santini, the psychologist puts it, "If an AI's response resembles a human's, it's not because it sees the same thing but it's because its training data mirrors our collective visual culture."

What we get is a facsimile of reality – and of thought. "ChatGPT can accurately describe emotions without actually experiencing them," Ghosh says. That doesn't make AI useless when it comes to matters of the mind, however. "ChatGPT can understand emotions, describe them, and help others make sense of their feelings," he says.

Ghosh believes that ChatGPT is simply regurgitating random information from its dataset for the benefit of users – which gives the impression of "thinking", but never actually is. Part of the reason why we perceive it possibly as thinking is the way that the tech companies behind these AI chatbots present their products. ChatGPT and tools like it are always helpful, always conscientious, and always conversational. The anthropomorphising of these tools has helped their adoption – but it can muddy the waters around how they – and their outputs – are perceived.

Estudio Santa Rita The images that an AI "sees" in an inkblot is merely a reflection of the data it has been trained upon (Credit: Estudio Santa Rita)Estudio Santa Rita

The images that an AI "sees" in an inkblot is merely a reflection of the data it has been trained upon (Credit: Estudio Santa Rita)

One way of uncovering that unreality is by simply pressing the refresh button, says Ghosh. Present ChatGPT with the same inkblot image to analyse, and it may well come up with a totally different interpretation in the same conversation.

When we gave ChatGPT the same image twice, the AI did indeed give different responses.

"A human would typically stick to their previous answer because personal experiences and emotions influence their responses," says Ghosh. "In contrast, ChatGPT generates responses based on its dataset."

That's part of the problem when trying to infer meaning in what ChatGPT says about the inkblots it sees: the chatbot is simply regurgitating what its training data has encountered.

This is perhaps best demonstrated by an experiment by researchers at the MediaLab at the Massachusetts Institute of Technology in Cambridge, US. They trained an AI algorithm called "Norman" – named after Alfred Hitchcock's character Norman Bates – on images from a Reddit group that shares images of people dying in gruesome circumstances. When they showed Norman a set of Rorschach inkblots, the images it described seeing reflected the darkness it had been trained upon. While an algorithm trained on a more normal set of image data would see a group of birds on a tree branch, Norman described seeing a man being electrocuted.

It is a stark illustration of just how important the data used to train an AI can be. Train a model using bad data, and the resulting AI will reflect the flaws in that data.

Estudio Santa Rita AIs are good at spotting patterns in large sets of data but they can also "hallucinate" – making up information that isn't true (Credit: Estudio Santa Rita)Estudio Santa Rita

AIs are good at spotting patterns in large sets of data but they can also "hallucinate" – making up information that isn't true (Credit: Estudio Santa Rita)

But what an AI chooses to regurgitate in response to ambiguous stimuli is interesting, says Dekker – in part because of what it says, rather than what it "knows" (AI systems know nothing innately). "It has a general understanding of colour theory and what emotional responses these could give when analysed," says Dekker. "I have a feeling it has a general understanding of the images and what people associate it with. This got me thinking: Wouldn't it be cool if we had – or created – a new set of Rorschach-like images which the model does not know about at all, and let it analyse those?"

While this isn't something Dekker is actively pursuing, we know that AI algorithms do have a tendency to "hallucinate" and invent information that isn't true. But it is also possible force them to hallucinate too. In one experiment by computer scientists at Massachusetts Institute of Technology in Cambridge, US, they were able to 3D print a turtle with a particular colour and texture that fooled a machine vision system into thinking it was a rifle. Similarly, a few stickers on a stop sign can render it invisible to the algorithm – a potential risk for a self-driving car. (Read more about the 'weird events' that make machines hallucinate.)

AI algorithms are particularly good at spotting patterns, but those differing responses to a single ambiguous image also highlight an aspect of the human mind that AI can't replicate: the emotions and unconscious meanings we attach to things we encounter in the world around us. There was no subjectivity in what the AI system was saying when presented with the inkblots, Kubiliute says. "It cannot understand the symbolic meaning or emotional resonance that a human might associate with a particular image."

And that in itself tells us a lot about the human mind. "The human psyche is filled with internal conflicts, such as the tension between desires and morals or fears and ambitions," says Ghosh. "In contrast, AI functions on clear logic and does not struggle with inner dilemmas essential to human thought and decision-making."

--

For more technology news and insights, sign up to our Tech Decoded newsletter, while The Essential List delivers a handpicked selection of features and insights to your inbox twice a week.

For more science, technology, environment and health stories from the BBC, follow us on FacebookX and Instagram.

关于《We gave an AI a Rorschach test. What it saw in the inkblots offers a window into the human mind》的评论


暂无评论

发表评论

摘要

The article discusses the limitations of AI systems like ChatGPT when it comes to interpreting ambiguous stimuli such as Rorschach inkblots, which are designed to reveal unconscious aspects of human psychology. Here's a summary of the key points: 1. **Ambiguity and Interpretation**: - When presented with Rorschach inkblots, AI systems like ChatGPT generate interpretations based on patterns found in their training data. - Unlike humans, who interpret such images subjectively based on emotions and unconscious meanings, AI lacks this subjective element. 2. **Training Data Influence**: - The quality of the data used to train an AI model significantly influences its responses. - A famous example is Norman, trained on disturbing imagery from a subreddit about gore, which led it to describe inkblots as violent scenes rather than abstract patterns. 3. **Subjectivity and Emotion in Human Interpretation**: - Humans bring personal experiences, emotions, and symbolic meanings when interpreting ambiguous images. - AI systems lack the emotional and subjective context that humans use to interpret such stimuli. 4. **Inner Conflicts and Dilemmas**: - The human psyche is complex with internal conflicts like desires versus morals or fears versus ambitions. - AI operates on clear logic without experiencing these inner dilemmas, highlighting a fundamental difference in how humans and machines process information. 5. **Potential for Hallucination**: - AI systems can "hallucinate" and invent information not based on reality when presented with ambiguous data. - Researchers have demonstrated this by tricking machine vision systems into recognizing objects that are not there (e.g., a turtle appearing as a rifle). 6. **Implications of the Experiment**: - While AI is good at pattern recognition, its lack of emotional and subjective interpretation reveals crucial aspects of human cognition and psychology. - This limitation underscores the importance of understanding the differences between human thought processes and those of artificial intelligence. The article concludes by emphasizing that despite AI's ability to analyze patterns in large datasets, it fundamentally lacks the subjectivity and emotional depth inherent in human psychological processes.