Relighting4D Save

[ECCV 2022] Relighting4D: Neural Relightable Human from Videos

Project README

Relighting4D: Neural Relightable Human from Videos

S-Lab, Nanyang Technological University

ECCV 2022

Project Page | Video | Paper

Updates

[08/2022] Model weights released. Google Drive

[07/2022] Paper uploaded to arXiv. arXiv

[07/2022] Code released.

Citation

If you find our work useful for your research, please consider citing this paper:

@inproceedings{chen2022relighting,
    title={Relighting4D: Neural Relightable Human from Videos},
    author={Zhaoxi Chen and Ziwei Liu},
    booktitle={ECCV},
    year={2022}
}

Installation

We recommend using Anaconda to manage your python environment. You can setup the required environment by the following command:

conda env create -f environment.yml
conda activate relighting4d

Datasets

People-Snapshot

We follow NeuralBody for data preparation.

  1. Download the People-Snapshot dataset here.

  2. Process the People-Snapshot dataset using the script.

  3. Create a soft link:

    cd /path/to/Relighting4D
    mkdir -p data
    cd data
    ln -s /path/to/people_snapshot people_snapshot
    

ZJU-MoCap

Please refer to here for requesting the download link. Once downloaded, don't forget to add a soft link:

cd /path/to/Relighting4D
mkdir -p data
cd data
ln -s /path/to/zju_mocap zju_mocap

Training

We first reconstruct an auxiliary density field in Stage I and then train the whole pipeline in Stage II. All trainings are done on a Tesla V100 GPU with 16GB memory.

Take the training on female-3-casual as an example.

  • Stage I:

    python train_net.py --cfg_file configs/snapshot_exp/snapshot_f3c.yaml exp_name female3c resume False gpus "0,"
    

    The model weights will be saved to /data/trained_model/if_nerf/female3c/latest.pth.

  • Stage II:

    python train_net.py --cfg_file configs/snapshot_exp/snapshot_f3c.yaml task relighte2e exp_name female3c_relight train_relight True resume False train_relight_cfg.smpl_model_ckpt ./data/trained_model/if_nerf/female3c/latest.pth gpus "0,"
    

    The final model will be saved to /data/trained_model/relighte2e/female3c_relight/latest.pth.

  • Tensorboard:

    tensorboard --logdir data/record/if_nerf
    tensorboard --logdir data/record/relighte2e
    

Rendering

To relight a human performer from the trained video, our model requires an HDR environment map as input. We provide 8 HDR maps at light-probes. You can also use your own HDRIs or download some samples from Poly Haven.

You are welcome to download our checkpoints from Google Drive.

Here, we take the rendering on female-3-casual as an example.

  • Relight with novel views of single frame

    python run.py --type relight --cfg_file configs/snapshot_exp/snapshot_f3c.yaml exp_name female3c_relight task relighte2e vis_relight True ratio 0.5 gpus "0,"
    
  • Relight the dynamic humans in video frames

    python run.py --type relight_npose --cfg_file configs/snapshot_exp/snapshot_f3c.yaml exp_name female3c_relight task relighte2e vis_relight True vis_relight_npose True ratio 0.5 pyramid False gpus "0,"
    

The results of rendering are located at /data/render/. For example, rendering results with courtyard HDR environment are shown as follows:

Acknowledgements

This work is supported by the National Research Foundation, Singapore under its AI Singapore Programme, NTU NAP, MOE AcRF Tier 2 (T2EP20221-0033), and under the RIE2020 Industry Alignment Fund - Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner(s).

Relighting4D is implemented on top of the NeuralBody codebase.

Open Source Agenda is not affiliated with "Relighting4D" Project. README Source: FrozenBurning/Relighting4D

Open Source Agenda Badge

Open Source Agenda Rating