Hash3D Save

Hash3D: Training-free Acceleration for 3D Generation

Project README

Training-free Acceleration
for 3D Generation 🏎️💨

License arXiv arXiv

Introduction

This repository contains the official implementation for our paper

Hash3D: Training-free Acceleration for 3D Generation

🥯[Project Page] 📝[Paper] </>[code]

Xingyi Yang, Xinchao Wang

National University of Singapore

pipeline

We present Hash3D, a universal solution to acclerate score distillation samplin (SDS) based 3D generation. By effectively hashing and reusing these feature maps across neighboring timesteps and camera angles, Hash3D substantially prevents redundant calculations, thus accelerating the diffusion model's inference in 3D generation tasks.

What we offer:

  • ⭐ Compatiable to Any 3D generation method using SDS.
  • ⭐ Inplace Accerlation for 1.3X - 4X.
  • ⭐ Training-Free.

Results Visualizations

Project Structure

The repository is organized into three main directories, each catering to a different repo that Hash3D can be applied on:

  1. threesdtudio-hash3d: Contains the implementation of Hash3D tailored for use with the threestudio.
  2. dreamgaussian-hash3d: Focuses on integrating Hash3D with the DreamGaussian for image-to-3D generation.
  3. gaussian-dreamer-hash3d: Dedicated to applying Hash3D to GaussianDreamer for faster text-to-3D tasks.

What we add?

The core implementation is in the guidance_loss for each SDS loss computation. We

See hash3D/threestudio-hash3d/threestudio/models/guidance/zero123_unified_guidance_cache.py for example. The code for the hash table implementation is in hash3D/threestudio-hash3d/threestudio/utils/hash_table.py.

Getting Started

Installation

Navigate to each of the specific directories for environment-specific installation instructions.

Usage

Refer to the README within each directory for detailed usage instructions tailored to each environment.

For example, to run Zero123+SDS with hash3D

cd threestudio-hash3d
python launch.py --config configs/stable-zero123_hash3d.yaml --train --gpu 0 data.image_path=https://adamdad.github.io/hash3D/load/images/dog1_rgba.png

Evaliation

  1. Image-to-3D: GSO dataset GT meshes and renderings can be found online. With the rendering of the reconstructed 3D objects at pred_dir and the gt rendering at gt_dir, run
python eval_nvs.py --gt $gt_dir --pr $pred_dir 
  1. Text-to-3D: Run all the prompts in assets/prompt.txt. And compute the CLIP score between text and rendered image as
python eval_clip_sim.py "$gt_prompt" $pred_dir --mode text

Acknowledgement

We borrow part of the code from DeepCache for feature extraction from diffusion models. We also thanks the implementation from threestudio, DreamGaussian, Gaussian-Dreamer, and the valuable disscussion with @FlorinShum and @Horseee.

Citation

@misc{yang2024hash3d,
      title={Hash3D: Training-free Acceleration for 3D Generation}, 
      author={Xingyi Yang and Xinchao Wang},
      year={2024},
      eprint={2404.06091},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}
Open Source Agenda is not affiliated with "Hash3D" Project. README Source: Adamdad/hash3D
Stars
132
Open Issues
2
Last Commit
1 month ago
Repository

Open Source Agenda Badge

Open Source Agenda Rating