This repository contains the official implementation for our paper
Hash3D: Training-free Acceleration for 3D Generation
🥯[Project Page] 📝[Paper] </>[code]
Xingyi Yang, Xinchao Wang
National University of Singapore
We present Hash3D, a universal solution to acclerate score distillation samplin (SDS) based 3D generation. By effectively hashing and reusing these feature maps across neighboring timesteps and camera angles, Hash3D substantially prevents redundant calculations, thus accelerating the diffusion model's inference in 3D generation tasks.
What we offer:
- ⭐ Compatiable to Any 3D generation method using SDS.
- ⭐ Inplace Accerlation for 1.3X - 4X.
- ⭐ Training-Free.
Prompt | Gaussian-Dreamer | Hash3D + Gaussian-Dreamer |
---|---|---|
A bear dressed as a lumberjack |
a.bear.dressed.as.a.lumberjack.mp4 |
a.bear.dressed.as.a.lumberjack_hash.mp4 |
A train engine made out of clay |
a.train.engine.made.out.of.clay.mp4 |
a.train.engine.made.out.of.clay_hash.mp4 |
The repository is organized into three main directories, each catering to a different repo that Hash3D can be applied on:
threesdtudio-hash3d
: Contains the implementation of Hash3D tailored for use with thethreestudio
.dreamgaussian-hash3d
: Focuses on integrating Hash3D with the DreamGaussian for image-to-3D generation.gaussian-dreamer-hash3d
: Dedicated to applying Hash3D to GaussianDreamer for faster text-to-3D tasks.
The core implementation is in the guidance_loss
for each SDS loss computation. We
See hash3D/threestudio-hash3d/threestudio/models/guidance/zero123_unified_guidance_cache.py
for example. The code for the hash table implementation is in hash3D/threestudio-hash3d/threestudio/utils/hash_table.py
.
Navigate to each of the specific directories for environment-specific installation instructions.
Refer to the README
within each directory for detailed usage instructions tailored to each environment.
For example, to run Zero123+SDS with hash3D
cd threestudio-hash3d
python launch.py --config configs/stable-zero123_hash3d.yaml --train --gpu 0 data.image_path=https://adamdad.github.io/hash3D/load/images/dog1_rgba.png
- Image-to-3D: GSO dataset GT meshes and renderings can be found online. With the rendering of the reconstructed 3D objects at
pred_dir
and the gt rendering atgt_dir
, run
python eval_nvs.py --gt $gt_dir --pr $pred_dir
- Text-to-3D: Run all the prompts in
assets/prompt.txt
. And compute the CLIP score between text and rendered image as
python eval_clip_sim.py "$gt_prompt" $pred_dir --mode text
We borrow part of the code from DeepCache for feature extraction from diffusion models. We also thanks the implementation from threestudio, DreamGaussian, Gaussian-Dreamer, and the valuable disscussion with @FlorinShum and @Horseee.
@misc{yang2024hash3d,
title={Hash3D: Training-free Acceleration for 3D Generation},
author={Xingyi Yang and Xinchao Wang},
year={2024},
eprint={2404.06091},
archivePrefix={arXiv},
primaryClass={cs.CV}
}