DisCo3D: Distilling Multi-View Consistency for 3D Scene Editing
Abstract
DisCo3D enhances 3D editing by distilling 3D consistency into a 2D editor, improving multi-view consistency and editing quality.
While diffusion models have demonstrated remarkable progress in 2D image generation and editing, extending these capabilities to 3D editing remains challenging, particularly in maintaining multi-view consistency. Classical approaches typically update 3D representations through iterative refinement based on a single editing view. However, these methods often suffer from slow convergence and blurry artifacts caused by cross-view inconsistencies. Recent methods improve efficiency by propagating 2D editing attention features, yet still exhibit fine-grained inconsistencies and failure modes in complex scenes due to insufficient constraints. To address this, we propose DisCo3D, a novel framework that distills 3D consistency priors into a 2D editor. Our method first fine-tunes a 3D generator using multi-view inputs for scene adaptation, then trains a 2D editor through consistency distillation. The edited multi-view outputs are finally optimized into 3D representations via Gaussian Splatting. Experimental results show DisCo3D achieves stable multi-view consistency and outperforms state-of-the-art methods in editing quality.
Community
Hi! I was really wondering how to edit N images on the 2D editing model? If you concat them and use the inner-attention to learn the 3D consistency, the model results will decrease because of the too huge 25 images inputs. If you do it on the batch dimension, that means you must sacrifice the diversity of the editing model as all independent views have only one 3D-consistent expectation?
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper