This paper proposes a diffusion prior-based dynamic viewpoint synthesis method for generating novel viewpoints of dynamic scenes from monocular videos. The method achieves both geometric and scene consistency by fine-tuning video frames and knowledge distillation. Qualitative and quantitative experiments demonstrate the effectiveness and robustness of the proposed method, highlighting its advantages in complex scenes.