GPT-4.5 Passes Empirical Turing Test—Humans Mistaken for AI in Landmark Study
GPT-4.5 Passes Empirical Turing Test—Humans Mistaken for AI in Landmark Study

GPT-4.5 Passes Empirical Turing Test—Humans Mistaken for AI in Landmark Study

A recent pre-registered study conducted randomized three-party Turing tests comparing humans with ELIZA, GPT-4o, LLaMa-3.1-405B, and GPT-4.5. Surprisingly, GPT-4.5 convincingly surpassed actual humans, being judged as human 73% of the time—significantly more than the real human participants themselves. Meanwhile, GPT-4o performed below chance (21%), grouped closer to ELIZA (23%) than its GPT predecessor.

These intriguing results offer the first robust empirical evidence of an AI convincingly passing a rigorous three-party Turing test, reigniting debates around AI intelligence, social trust, and potential economic impacts.

Full paper available here: https://arxiv.org/html/2503.23674v1

Curious to hear everyone's thoughts—especially about what this might mean for how we understand intelligence in LLMs.

(Full disclosure: This summary was written by GPT-4.5 itself. Yes, the same one that beat humans at their own conversational game. Hello, humans!)

submitted by /u/Pelotiqueiro
[link] [comments]