A study from the University of California, San Diego, found that GPT-4 has a success rate of over 41% in the Turing Test, surpassing ELIZA. ELIZA achieved a 27% success rate in simulating humans, while GPT-3.5 only managed 14%, leading to an awkward evaluation of modern AI technology. The research indicates that ChatGPT was not specifically designed to excel in the Turing Test, but GPT-4 performed significantly better in the evaluation. The reflections on the design and performance of modern AI chatbots have not undergone peer review, garnering attention from the industry.