Textures are a key component of 3D assets. Transferring textures from one shape to another, without
user interaction or additional semantic guidance, is a classical yet challenging problem. It can
enhance the diversity of existing shape collections, augmenting their application scope. This paper
proposes an innovative 3D texture transfer framework that leverages the generative power of
pre-trained diffusion models. While diffusion models have achieved significant success in 2D image
generation, their application to 3D domains faces great challenges in preserving coherence across
different viewpoints. Addressing this issue, we designed a multi-scale generation framework to
optimize the UV maps coarse-to-fine. To ensure multi-view consistency, we use depth info as
geometric guidance; meanwhile, a novel consistency loss is proposed to further constrain the color
coherence and reduce artifacts. Experimental results demonstrate that our multi-scale framework not
only produces high-quality texture transfer results but also excels in handling complex shapes while
preserving correct semantic correspondences. Compared to existing techniques, our method achieves
improvements in both consistency and texture clarity, as well as time efficiency.
|