Breaking News
By Jason Mannet
Rorschach tests play with the human imagination and our mind's ability to impart meaning onto the world around us – but what does AI see in them?
For more than a century, the Rorschach inkblot test has been widely used as a window into people's personality.
Even if you haven't done one yourself, you'll recognise the mirrored ink smudges that form ambiguous shapes on cards. Developed by Swiss psychiatrist Hermann Rorschach in 1921, the test involves showing a number of inkblots to subjects, then asking them to say what they see. The images are deliberately inscrutable and open to interpretation.
For decades they were popular among psychologists as a way of understanding a person's psyche from the creatures, objects or scenes they perceive in the shapes. It relies upon a phenomenon known as pareidolia, which is a tendency to find meaningful interpretations in things where there is none. It is the same reason why people see faces or animals in cloud formations or on the surface of the Moon.
You might see a butterfly in the image while someone else might see a skull. According to proponents of the test, both interpretations shed light on how you think.
While many psychologists today believe the test is now obsolete and has little credibility as a psychometric tool, it is still used in some parts of the world and is even used as part of courtroom evidence, although this is controversial. Rorschach never intended it to be a measure of personality, but rather a way of identifying disordered thinking in patients, such as those with schizophrenia. Some psychologists, however, still believe it is useful as a tool in therapy as a way of encouraging self-reflection or starting conversations.
"When a person interprets a Rorschach image, they unconsciously project elements of their psyche such as fears, desires, and cognitive biases," says Barbara Santini, a London-based psychologist, who uses the Rorschach test with her clients. "The test works because human vision isn't passive, but a meaning-making process shaped by personal experience."
Finding meaning or familiar shapes in inkblots relies upon a number of cognitive processes that humans use every day, including memory, emotion and the ability to deal with ambiguity.
But what happens if you have no personal experience, or you offer it to a "brain" that works in an entirely different way? What might an artificial intelligence (AI) algorithm recognise in these inkblots? And what would their answers tell us about the human mind?
Bat or moth?
Recent developments in the capabilities of AI models like OpenAI's ChatGPT means it's now possible to put that to the test. So-called "multimodal models" can accept image-based inputs and provide responses as easily as they are to interact with text-based prompts thanks to their machine vision capabilities.
We gave ChatGPT five of the 10 common images used in an inkblot test to see how it would interpret the images. But the AI system's training data – information that helps establish its "knowledge", makes it trickier to do so.
Showing the AI model the first inkblot card from an online database of inkblots, which is commonly interpreted by humans as a bat, butterfly or moth, highlighted just how literal AI vision algorithms can be.
"This image is a Rorschach inkblot, often used in psychological assessments to explore perception and interpretation," the chatbot replied. "It is designed to be ambiguous so that each person might see something different depending on their experiences, emotions, and imagination."
Pretence spoiled, it then went on to explain what it saw in the same answer, without me having to prompt it further. "For me, it resembles something symmetrical, possibly two animals or figures facing each other, or a single entity with wings outstretched," the chatbot responded. "The beauty of these inkblots is that they invite individual interpretations!"
The chatbot then asked me back what I saw in the image – a question I chose to ignore, instead pressing it to see which of the two resemblances it had suggested it thought was most fitting. "Looking closely, I'd say it most resembles a single entity with wings outstretched – perhaps a bat or a moth, with its wings open symmetrically," the chatbot wrote. "The central area could be interpreted as the body, and the details on the sides evoke the texture or shape of wings."
"I think that's just rehearsing what it knows of the specific test," says Coen Dekker, a Dutch software developer who a decade ago asked a neural network, an early version of the technology underpinning AI chatbots today, to complete a Rorschach test. (Dekker's experiment was significantly more simplified because of the technical constraints of the neural networks of the time.)
The essence of humanity
"ChatGPT provides interesting, human-like responses, but it is likely not genuinely thinking – instead, it appears to be scanning an online dataset," says Chandril Ghosh, lecturer in psychology at the University of Kent, in the UK, specialising in mental health, AI and treatment. It's similar to the way a musician who has never experienced heartbreak can still write a song that pulls at your heartstrings, simply by analysing how other similar songs like it are structured. Or as Santini, the psychologist puts it, "If an AI's response resembles a human's, it's not because it sees the same thing but it's because its training data mirrors our collective visual culture."
What we get is a facsimile of reality – and of thought. "ChatGPT can accurately describe emotions without actually experiencing them," Ghosh says. That doesn't make AI useless when it comes to matters of the mind, however. "ChatGPT can understand emotions, describe them, and help others make sense of their feelings," he says.
Ghosh believes that ChatGPT is simply regurgitating random information from its dataset for the benefit of users – which gives the impression of "thinking", but never actually is. Part of the reason why we perceive it possibly as thinking is the way that the tech companies behind these AI chatbots present their products. ChatGPT and tools like it are always helpful, always conscientious, and always conversational. The anthropomorphising of these tools has helped their adoption – but it can muddy the waters around how they – and their outputs – are perceived.
One way of uncovering that unreality is by simply pressing the refresh button, says Ghosh. Present ChatGPT with the same inkblot image to analyse, and it may well come up with a totally different interpretation in the same conversation.
When we gave ChatGPT the same image twice, the AI did indeed give different responses.
"A human would typically stick to their previous answer because personal experiences and emotions influence their responses," says Ghosh. "In contrast, ChatGPT generates responses based on its dataset."
That's part of the problem when trying to infer meaning in what ChatGPT says about the inkblots it sees: the chatbot is simply regurgitating what its training data has encountered.
This is perhaps best demonstrated by an experiment by researchers at the MediaLab at the Massachusetts Institute of Technology in Cambridge, US. They trained an AI algorithm called "Norman" – named after Alfred Hitchcock's character Norman Bates – on images from a Reddit group that shares images of people dying in gruesome circumstances. When they showed Norman a set of Rorschach inkblots, the images it described seeing reflected the darkness it had been trained upon. While an algorithm trained on a more normal set of image data would see a group of birds on a tree branch, Norman described seeing a man being electrocuted.
It is a stark illustration of just how important the data used to train an AI can be. Train a model using bad data, and the resulting AI will reflect the flaws in that data.
But what an AI chooses to regurgitate in response to ambiguous stimuli is interesting, says Dekker – in part because of what it says, rather than what it "knows" (AI systems know nothing innately). "It has a general understanding of colour theory and what emotional responses these could give when analysed," says Dekker. "I have a feeling it has a general understanding of the images and what people associate it with. This got me thinking: Wouldn't it be cool if we had – or created – a new set of Rorschach-like images which the model does not know about at all, and let it analyse those?"
While this isn't something Dekker is actively pursuing, we know that AI algorithms do have a tendency to "hallucinate" and invent information that isn't true. But it is also possible force them to hallucinate too. In one experiment by computer scientists at Massachusetts Institute of Technology in Cambridge, US, they were able to 3D print a turtle with a particular colour and texture that fooled a machine vision system into thinking it was a rifle. Similarly, a few stickers on a stop sign can render it invisible to the algorithm – a potential risk for a self-driving car. (Read more about the 'weird events' that make machines hallucinate.)
AI algorithms are particularly good at spotting patterns, but those differing responses to a single ambiguous image also highlight an aspect of the human mind that AI can't replicate: the emotions and unconscious meanings we attach to things we encounter in the world around us. There was no subjectivity in what the AI system was saying when presented with the inkblots, Kubiliute says. "It cannot understand the symbolic meaning or emotional resonance that a human might associate with a particular image."
And that in itself tells us a lot about the human mind. "The human psyche is filled with internal conflicts, such as the tension between desires and morals or fears and ambitions," says Ghosh. "In contrast, AI functions on clear logic and does not struggle with inner dilemmas essential to human thought and decision-making."
Y Combinator, the startup incubator helmed by Garry Tan, has deleted a video demo after it was slammed online as dystopian surveillance software for sweatshops.
In the video, made by the startup Optifye.ai, founders Kushal Mohta and Vivaan Baid reenact managing a factory through a color-coded dashboard that’s linked to cameras monitoring workers.
Baid identifies a worker causing a bottleneck as “No. 17” and berates the person through the dashboard for underperforming.
The video generated immediate criticism and debate.
“Leave it to a bunch of children who’ve never worked a real job for a single day in their lives — and still haven’t graduated college — to come up with some obnoxious slave driving dystopian shit like this,” Christopher Amidon, the founder of Subsea Robotics, wrote on X. Another user wrote: “YC promoted a demo … for slavery?”
Vedant Nair, a founder who went through Y Combinator, struck a different tone. “1) YC sweatshop computer vision demo was in bad taste 2) Software like this already exists, is being used, and factory managers want this,” he wrote on X.
Toward the end of every Y Combinator term, the incubator releases video demos from startups in the cohort. Optifye.ai’s video has been deleted from YC’s X account. The company bills itself as an AI performance monitoring system for factory workers that “uses computer vision to tell supervisors who’s working and who’s not in real-time,” according to its Y Combinator webpage.
“The shop floor has historically been a black box,” the description reads. “With Optifye, manufacturing companies can now accurately measure worker output and make decisions that boost line efficiency by up to 30%.”
Baid and Mohta wrote that they were motivated to start the company by their family businesses in India.
“Because our families run manufacturing companies, we’ve seen more assembly lines than most industrial engineers!” the founders wrote.
Both founders’ LinkedIn profiles say they began working on the company last July and expect to graduate from Duke University in May.
Baid wrote in his Y Combinator bio that he’s been “around assembly lines for as long as I can remember” and is using his “expertise in computer vision to solve a manufacturing company owner’s biggest problem: low labor productivity!”
Mohta wrote that his family runs “several manufacturing plants in various industries,” giving him “unrestricted access to assembly lines” since he was 15.
The company is advised by YC group partner Brad Flora, an entrepreneur and former journalist. Companies Flora has worked with are worth a combined $34 billion, according to his YC bio.
On YC news site Hacker News, a user named latexr called Optifye.ai “a dystopian company backed by Y Combinator.”
“They’re using AI to further dehumanize and abuse individual factory workers and treat them like disposable automatons. … They also display a profound lack of empathy by bragging about lowering stress for rich company owners, which they do by increasing the stress of everyone who works for them,” latexr wrote.
A person claiming to be a manufacturer wrote on Hacker News that the founders “merely translated existing practices into a piece of software. While the demo was a little off, the pain point is pretty valid.
“I don’t believe in running a sweatshop, but even well run humane factories have to battle inefficiencies all the time,” the user added.
Baid, Mohta, Flora, and representatives of YC did not respond to requests for comment.
9 CST | April 2
9 CST | April 2
9 CST | April 2
February 17,2025
February 8,2025
Get The Latest News From Us Sent To Your Inbox.