PIR: Photometric Inverse Rendering with Shading Cues Modeling and Surface Reflectance Regularization
The official implementation of paper "PIR: Photometric Inverse Rendering with Shading Modeling and Surface Reflectance Regularization". We propose a framework that jointly accounts for self-shadows, indirect illumination and material decomposition. To enhance surface reflectance decomposition, we introduce a new regularization by distilling DINO features to foster accurate and consistent material decomposition.
- Technical Report
- Release Code
- Release Dataset Rendering Code
- This repo is still under construction. The code will be refactored for better use.
You may follow the next steps to use PIR
.
git clone https://github.com/ZqlwMatt/PIR
cd PIR
conda create -n pir python=3.10 && conda activate pir
First install Pytorch by your CUDA version:
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
Then install the other requirements via:
pip install -r requirements.txt
pip install tensorboard kornia
conda install -c conda-forge igl
We have tested the training process in Ubuntu 22.04.5 with a RTX3090. We provide training and evaluation scripts in the folder scripts
. You should modify the data_dir
to your data folder. Please see Dataset
section to download and prepare the dataset.
sh scripts/train_vis.sh
sh scripts/eval_vis.sh
The training contains 3 stages:
- Volume rendering to initialize the geometry (SDF) and diffuse albedo network;
- Initialize DINO features and Optimize the offset (save visibility map);
- Optimize PBR materials.
We have set up a few network warmup tricks to stabilize the training, and you can refer to the code for details.
For mesh evaluation, you can src/eval_mesh.py
to calculate the Chamfer distance.
The default camera is collocated with the point light. You should add an offset to the light source. Check mitsuba/render_rgb_flabh_mat.py
for details, or you can use our rendered dataset via Google Drive.
Note that you should have Mitsuba installed, you can use this Dockerfile to set up the environment.
Check mitsuba/render_rgb_envmap_mat.py
.
We use the OpenCV camera convention just like NeRF++; you might want to use the camera visualization and debugging tools in that codebase to inspect if there's any issue with the camera parameters. Note we also assume the objects are inside the unit sphere.
Please refer to our website for more results and comparison.
results.mp4
@article{bao2024pir,
title={PIR: Photometric Inverse Rendering with Shading Cues Modeling and Surface Reflectance Regularization},
author={Jingzhi Bao and Guanying Chen and Shuguang Cui},
journal={arXiv preprint arXiv:2408.06828},
year={2024}
}
This codebase is built upon the IRON. Thanks to the authors for open-sourcing their projects. The shader is modified to approximately align with Mitsuba point emitter shader.