Magic搭 Community collaborates with vLLM and FastChat to provide efficient LLM inference and deployment services. Developers can use vLLM as the inference engine in FastChat, offering high-throughput model inference. FastChat is an open platform for training, serving, and evaluating LLM-based ChatBots. vLLM is an LLM service developed by researchers from the University of California, Berkeley, Stanford University, and the University of California, San Diego.