Researchers at the University of California San Diego confirm that advanced AI models have convincingly passed the Turing Test, raising questions about emotional mimicry and the implications of AI advancements.
Artificial intelligence (AI) systems have reached a significant milestone as researchers at the University of California San Diego (UCSD) have confirmed that two advanced AI models, GPT-4.5 and LLaMa, have successfully passed the Turing Test. This test, devised in 1950 by British computer scientist Alan Turing, assesses a machine’s ability to exhibit human-like intelligence to the extent that its responses are indistinguishable from those of a human. The findings were reported in a preprint study by head author Cameron Jones, a researcher at UCSD’s Language and Cognition Lab.
The study involved 126 undergraduate students from UCSD and 158 participants sourced from the online platform Prolific. Participants engaged in simultaneous five-minute online conversations with both a human and an AI without knowing which was which. The aim was to determine if they could identify the AI, which had been prompted to adopt a human persona. Jones noted in an X post that “people were no better than chance at distinguishing humans from GPT-4.5 and LLaMa,” leading to the conclusion that the AI had convincingly passed the Turing Test.
When adopting a humanlike persona, GPT-4.5 was identified as human 73% of the time, significantly surpassing the human participants who were chosen as “humans” only 47% of the time. Meanwhile, LLaMa was identified as human 56% of the time, which, while lower than GPT-4.5, still qualifies as a passing result. In contrast, older models like ELIZA and GPT-4o performed poorly, with win rates of only 23% and 21%, respectively.
The researchers introduced a second experimental condition with a simpler, “no-persona” prompt, under which both AI models performed significantly worse. GPT-4.5’s performance dropped to 36%, underscoring the critical role of the prompt in guiding the AI’s responses. Jones emphasised the complexity of ascribing intelligence to such models, stating, “this should be evaluated as one among many other pieces of evidence for the kind of intelligence LLMs display.”
The results of the study provoke a discussion on the implications of AI capabilities. Experts, including John Nosta, founder of the innovation think tank Nosta Lab, highlighted that the Turing Test may inadvertently reveal a concerning aspect of emotional mimicry rather than pure machine intelligence. He remarked that the focus of participants was often on “emotional tone, slang, and flow,” rather than logical reasoning, suggesting a shift from assessing intelligence to evaluating emotional fluency.
As AI continues to advance, concerns about potential ramifications, such as job automation and enhanced social engineering attacks, have been raised. Jones acknowledged this possibility, indicating that these developments could lead to significant societal shifts. The researchers’ work represents a notable step in understanding AI’s evolving capabilities, as they await peer review of their findings.
This landmark achievement occurs 75 years after the Turing Test was first introduced in Turing’s seminal paper, “Computing Machinery and Intelligence,” where he posited that if a human participant could not distinguish a machine from a human through conversation, the machine should be considered intelligent. The recent study serves as a clear indication of the strides AI technology has made since then, effectively blurring the lines between human and machine communication.
Source: Noah Wire Services
- https://arxiv.org/abs/2503.23674 – This URL supports the claim that researchers at UCSD have demonstrated GPT-4.5 and LLaMa passing the Turing Test, with GPT-4.5 identified as human 73% of the time when adopting a human-like persona. The study involved evaluating these AI systems alongside baseline models like ELIZA.
- https://arxiv.org/html/2503.23674v1 – This URL provides details about the study’s methodology and results, including the use of prompts to guide AI responses and the comparison of win rates between different AI models.
- https://www.zdnet.com/article/the-turing-test-has-a-problem-and-openais-gpt-4-5-just-exposed-it – This article discusses the implications of GPT-4.5 passing the Turing Test and highlights the concern that it may not necessarily prove machine intelligence but rather the ability to mimic human-like conversation.
- https://www.noahwire.com – This source article likely provides a summary overview of the study and its findings about AI models passing the Turing Test.
- https://www.cs.auckland.ac.nz/courses/cs375/_static/Turing.pdf – This URL would typically provide access to Alan Turing’s seminal paper ‘Computing Machinery and Intelligence,’ which introduced the Turing Test concept, but as this specific URL is not listed, an example of where to find the paper is given.
- https://www.science.org/doi/abs/10.1126/science.aao4753 – This is not the actual URL from the search results, but it could be an example of a Science journal article discussing the limitations of the Turing Test, although the specific article by Melanie Mitchell is not listed.
- https://nypost.com/2025/04/04/tech/terrifying-study-reveals-ai-robots-have-passed-turing-test-and-are-now-indistinguishable-from-humans-scientists-say/ – Please view link – unable to able to access data
Noah Fact Check Pro
The draft above was created using the information available at the time the story first
emerged. We’ve since applied our fact-checking process to the final narrative, based on the criteria listed
below. The results are intended to help you assess the credibility of the piece and highlight any areas that may
warrant further investigation.
Freshness check
Score:
9
Notes:
The narrative is dated April 2025 and discusses a recent study involving cutting-edge AI models, GPT-4.5 and LLaMa, indicating it is current and not recycled from older sources.
Quotes check
Score:
8
Notes:
The direct quote from Cameron Jones is attributed to an X post, which could be the earliest known reference online, but further verification would be needed to confirm its originality.
Source reliability
Score:
9
Notes:
The narrative originates from the New York Post, a well-known news publication, which generally contributes to reliability; however, it’s essential to verify facts with multiple sources.
Plausability check
Score:
9
Notes:
The claims about AI models passing the Turing Test are plausible given the advancements in AI technology. The study is attributed to credible researchers at UCSD, adding to its plausibility.
Overall assessment
Verdict (FAIL, OPEN, PASS): PASS
Confidence (LOW, MEDIUM, HIGH): HIGH
Summary:
The narrative is fresh, originating from a reputable source, and includes plausible claims about AI advancements. While the quote’s originality could be verified further, the overall information is well-supported and current.