3D-Scenedreamer: Text-Driven 3D-Consistent Scene Generation

Songchun Zhang, Yibo Zhang, Quan Zheng,Rui Ma,Wei Hua,Hujun Bao,Weiwei Xu,Changqing Zou

2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)(2024)

引用 0|浏览49
暂无评分
摘要
Text-driven 3D scene generation techniques have made rapid progress in recentyears. Their success is mainly attributed to using existing generative modelsto iteratively perform image warping and inpainting to generate 3D scenes.However, these methods heavily rely on the outputs of existing models, leadingto error accumulation in geometry and appearance that prevent the models frombeing used in various scenarios (e.g., outdoor and unreal scenarios). Toaddress this limitation, we generatively refine the newly generated local viewsby querying and aggregating global 3D information, and then progressivelygenerate the 3D scene. Specifically, we employ a tri-plane features-based NeRFas a unified representation of the 3D scene to constrain global 3D consistency,and propose a generative refinement network to synthesize new contents withhigher quality by exploiting the natural image prior from 2D diffusion model aswell as the global 3D information of the current scene. Our extensiveexperiments demonstrate that, in comparison to previous methods, our approachsupports wide variety of scene generation and arbitrary camera trajectorieswith improved visual quality and 3D consistency.
更多
查看译文
关键词
Diffusion Model,Visual Quality,3D Information,Unified Representation,3D Scene,Progress In Recent Years,Global Consistency,Visual 3D,Global 3D,Feature Maps,Point Cloud,Error Propagation,Quality Metrics,Depth Map,3D Mesh,Geometric Information,Depth Estimation,3D Representation,Global Representation,Outdoor Scenes,View Synthesis,Multi-view Images,Scene Representation,Input Text,Monocular Depth Estimation,Correction Mechanism
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要