Command Palette
Search for a command to run...
Michel Oscar ; Bar-On Roi ; Liu Richard ; Benaim Sagie ; Hanocka Rana

Abstract
In this work, we develop intuitive controls for editing the style of 3Dobjects. Our framework, Text2Mesh, stylizes a 3D mesh by predicting color andlocal geometric details which conform to a target text prompt. We consider adisentangled representation of a 3D object using a fixed mesh input (content)coupled with a learned neural network, which we term neural style fieldnetwork. In order to modify style, we obtain a similarity score between a textprompt (describing style) and a stylized mesh by harnessing therepresentational power of CLIP. Text2Mesh requires neither a pre-trainedgenerative model nor a specialized 3D mesh dataset. It can handle low-qualitymeshes (non-manifold, boundaries, etc.) with arbitrary genus, and does notrequire UV parameterization. We demonstrate the ability of our technique tosynthesize a myriad of styles over a wide variety of 3D meshes.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| neural-stylization-on-meshes | VQGAN | Mean Opinion Score (Q1:Overall): 2.83 ± 0.39 Mean Opinion Score (Q2: Content): 3.6 ± 0.68 Mean Opinion Score (Q3: Style): 2.59 ± 0.44 |
| neural-stylization-on-meshes | Text2Mesh | Mean Opinion Score (Q1:Overall): 3.9 ± 0.37 Mean Opinion Score (Q2: Content): 4.04 ± 0.53 Mean Opinion Score (Q3: Style): 3.91 ± 0.51 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.