The translated data: C-Eval is a comprehensive benchmark designed to assess the advanced knowledge and reasoning abilities of Chinese foundational models. It encompasses multiple-choice questions across four difficulty levels, covering 52 distinct subject areas. The question bank is sourced from simulated exams available on the internet. The C-Eval leaderboard showcases the performance of open-source models in this evaluation. This benchmark aids in selecting large models suitable for the field of natural language processing, thereby promoting the development of AI applications.