3DV
A Scalable Attention-Based Approach for Image-to-3D Texture Mapping
High-quality textures are critical for realistic 3D content creation, yet existing generative methods are slow, rely on UV maps, and often fail to remain faithful to a reference image. To address these challenges, we propose a transformer-based framework that predicts a 3D texture field directly from a single image and a mesh, eliminating the need for UV mapping and differentiable rendering, and enabling faster texture generation. Our method integrates a triplane representation with depth-based backprojection losses, enabling efficient training and faster inference. Once trained, it generates high-fidelity textures in a single forward pass, requiring only 0.2s per shape. Extensive qualitative, quantitative, and user preference evaluations demonstrate that our method outperforms state-of-the-art baselines on single-image texture reconstruction in terms of both fidelity to the input image and perceptual quality, highlighting its practicality for scalable, high-quality, and controllable 3D content creation.
Download publicationResearch Authors
AmirHossein Zamani
MILA and Concordia University
Related Publications
2024
Make-A-Shape: a Ten-Million-scale 3D Shape ModelTrained on 10 million 3D shapes, our model exhibits the capability to…
2024
GraphSeam: Supervised Graph Learning Framework for Semantic UV MappingProposing a fully automated UV mapping framework that enables users to…
2023
Neural Shape Diameter Function for Efficient Mesh SegmentationIntroducing a neural approximation of the Shape Diameter Function,…
2024
Wavelet Latent Diffusion: Billion-Parameter 3D Generative Model with Compact Wavelet EncodingsAddressing a common limitation of generative AI models, WaLa encodes…
Get in touch
Something pique your interest? Get in touch if you’d like to learn more about Autodesk Research, our projects, people, and potential collaboration opportunities.
Contact us