GEM3D: GEometry-conditioned Multi-View Diffusion for Text-Guided Texturing of 3D Shapes

On going project (Last update 01.Dec. 2023)

Anonymous1

* Equal Contribution

1EPFL  

Abstract

Text-driven texture synthesis benefits applications like 3D asset design and texture editing. Current methods employ pretrained 2D diffusion models for this task, either by defining a canonical view sequence and inpaint view-by-view, or use a shared texture map to synchronize denoising of different views. However, this pipeline has two drawbacks. First, when inpainting one view, the model only considers the colored region visible in that view, resulting in view-inconsistency (a fish having one big and one small eyes) and Janus problem (multiple faces). Second, smooth transition between views usually requires large overlapping, requiring many views, and results in slow generation. Inspired by recent advancement in multi-view diffusion, GEM3D proposes a geometry-conditioned multi-view diffusion. Similar to existing diffusion-based texturing method, we exploit geometry cues, e.g., normal map, as condition. Instead of relying overlapping regions, we use cross-view attention that connects views even without overlapping, effectively reducing cross-view inconsistency. In addition, our method speeds up texture synthesis by utilizing fewer views and getting ride of sequential generation. Moreover, our method is compatible with existing post-processing methods and can further boost texturing quality after generation.

Texturize Results

View-Consistent Texturize

Input Mesh

Source views

More Results

Roll: degrees
Pitch: degrees
Yaw: degrees
Scale: X: , Y: , Z:

Poster

BibTeX

@article{ren2022volrecon,
        author    = {Anonymous},
        title     = {GEM3D: GEometry-conditioned Multi-View Diffusion for Text-Guided Texturing of 3D Shapes},
        journal   = {arXiv},
        year      = {2023},
        }