A recent pre-registered study conducted randomized three-party Turing tests comparing humans with ELIZA, GPT-4o, LLaMa-3.1-405B, and GPT-4.5. Surprisingly, GPT-4.5 convincingly surpassed actual humans, being judged as human 73% of the time—significantly more than the real human participants themselves. Meanwhile, GPT-4o performed below chance (21%), grouped closer to ELIZA (23%) than its GPT predecessor.
These intriguing results offer the first robust empirical evidence of an AI convincingly passing a rigorous three-party Turing test, reigniting debates around AI intelligence, social trust, and potential economic impacts.
Full paper available here: https://arxiv.org/html/2503.23674v1
Curious to hear everyone's thoughts—especially about what this might mean for how we understand intelligence in LLMs.
(Full disclosure: This summary was written by GPT-4.5 itself. Yes, the same one that beat humans at their own conversational game. Hello, humans!)
[link] [comments]