Sketch2NeRF
Text-to-3D Generation Guided by Multi-view Sketches
CommonProductImageText-to-3D GenerationSketch Control
Sketch2NeRF is a text-to-3D generation framework guided by multi-view sketches. It leverages pre-trained 2D diffusion models (such as Stable Diffusion and ControlNet) to optimize 3D scenes represented by neural radiance fields (NeRF). The method also proposes a novel synchronized generation and reconstruction approach to effectively optimize NeRF. Experiments based on two collected multi-view sketch datasets demonstrate that our method can synthesize consistent 3D content with fine-grained sketch control under high-fidelity text prompts. Extensive results show that our method achieves state-of-the-art performance in terms of sketch similarity and text alignment.
Sketch2NeRF Visit Over Time
Monthly Visits
19075321
Bounce Rate
45.07%
Page per Visit
5.5
Visit Duration
00:05:32