ScholarQABench is a comprehensive evaluation platform designed to assess large language models (LLMs) in assisting researchers with the synthesis of scientific literature. Originating from the OpenScholar project, it offers a comprehensive evaluation framework comprising various datasets and evaluation scripts to measure models' performances across different scientific domains. The platform's significance lies in its ability to aid researchers and developers in understanding and enhancing the practicality and accuracy of language models in scientific literature research.