HairFree: Compositional 2D Head Prior for Text-Driven 360° Bald Texture Synthesis
Mirela Ostrek, Michael J. Black, and Justus Thies
The Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS) 2025, San Diego, USA
Abstract
Synthesizing high-quality 3D head textures is crucial for gaming, virtual reality, and digital humans. Achieving seamless 360° textures typically requires expensive multi-view datasets with precise tracking. However, traditional methods struggle without back-view data or precise geometry, especially for human heads, where even minor inconsistencies disrupt realism. We introduce HairFree, an unsupervised texturing framework guided by textual descriptions and 2D diffusion priors, producing high-consistency 360° bald head textures—including non-human skin with fine details—without any texture, back-view, bald, non-human, or synthetic training data. We fine-tune a diffusion prior on a dataset of mostly frontal faces, conditioned on predicted 3D head geometry and face parsing. During inference, HairFree uses precise skin masks and 3D FLAME geometry as input conditioning, ensuring high 3D consistency and alignment. We synthesize the full 360° texture by first generating a frontal RGB image aligned to the 3D FLAME pose and mapping it to UV space. As the virtual camera moves, we inpaint and merge missing regions. A built-in semantic prior enables precise region separation—particularly for isolating and removing hair—allowing seamless integration with various assets like customizable 3D hair, eyeglasses, jewelry, etc. We evaluate HairFree quantitatively and qualitatively, demonstrating its superiority over state-of-the-art 3D head avatar generation methods.
Method

System overview: (A) Data Processing: Estimate face parsing, 3D head mesh, and remove the foreground to form compositional inputs. (B.1) Training Prior: Fine-tune an LDM via ControlNet using these inputs and a generic “face” prompt. (B.2) Inference Prior: Swap in a generic LDM-Inpainting prior. (C) Texture Generation: Generate a frontal view, map to UV space, iteratively render "seen" and inpaint "unseen" regions while moving the camera, building a full 360° texture.
Citing the HairFree Paper and Dataset
If you find our paper or dataset useful to your research, please cite our work:
@inproceedings{HAIRFREE:NEURIPS:25, title = {HairFree: Compositional 2D Head Prior for Text-Driven 360° Bald Texture Synthesis}, author = {Ostrek, Mirela and Black, Michael J. and Thies, Justus}, booktitle = {Advances in Neural Information Processing Systems (NeurIPS)}, month = December, year = {2025}, month_numeric = {12}}
