(Translated by https://www.hiragana.jp/)
GitHub - NVlabs/latentfusion: LatentFusion: End-to-End Differentiable Reconstruction and Rendering for Unseen Object Pose Estimation
Skip to content

LatentFusion: End-to-End Differentiable Reconstruction and Rendering for Unseen Object Pose Estimation

License

Notifications You must be signed in to change notification settings

NVlabs/latentfusion

Repository files navigation

LatentFusion

Citing LatentFusion

If you find the LatentFusion code or data useful, please consider citing:

@inproceedings{park2019latentfusion,
  title={LatentFusion: End-to-End Differentiable Reconstruction and Rendering for Unseen Object Pose Estimation},
  author={Park, Keunhong and Mousavian, Arsalan and Xiang, Yu and Fox, Dieter},
  booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
  year={2020}
}

Setup

Please start by installing Miniconda3 with Python 3.7 or above.

wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh
bash Miniconda3-latest-Linux-x86_64.sh

Then create a Conda environment with our environment file:

conda env create -n latentfusion -f environment.yml
conda activate latentfusion

Please make sure the project root is added to the $PYTHONPATH. We provide a simple script for this:

# Activates the Conda environment and sets PYTHONPATH.
source env.sh

For training, we make use of Automatic Mixed Precision. Until PyTorch 1.6 is released, you must install the nightly version of PyTorch.

conda install pytorch torchvision cudatoolkit=10.2 -c pytorch-nightly

We've exluded PyTorch from the environment.yml file for this reason.

Trained Model

We provide a trained model. You can download the weights from here. The weights are licensed under a Creative Commons license. Please see the weights license for details.

You can use the model like this:

import torch
from latentfusion.recon.inference import LatentFusionModel
checkpoint = torch.load('path-to-checkpoint')
model = LatentFusionModel.from_checkpoint(checkpoint)

Please see the example notebook for details.

Dataset Download

BOP/LINEMOD

We use the BOP version of LINEMOD and other datasets. You can get them at the BOP website.

MOPED

You can download the MOPED dataset at the project page.

Pose Estimation

We provide an example script for pose estimation in the examples directory in the form of a Jupyter notebook. Download the weights from here and open the notebook with Jupyter.

Training

To train LatentFusion, we recommend that you use at least 4 RTX 2080 Ti GPUs. First, modify tools/train/train.sh with the correct paths to ShapeNet and MS-COCO.

You must first pre-process ShapeNet with our preprocessing script in tools/dataset/preprocess_shapenet.py. This requires that you have Blender installed.

sudo apt install blender
blender -P tools/dataset/preprocess_shapenet.py -- "$SHAPENET_PATH" "$OUT_PATH" --strip-materials --out-name ShapeNetCore-nomat

Once you have all of the data in place simply call the training script:

bash tools/train/train.sh

Training will roughly take 2 weeks on 4x RTX 2080 Ti GPUs and 1 week on 4x Tesla V100 GPUs.

About

LatentFusion: End-to-End Differentiable Reconstruction and Rendering for Unseen Object Pose Estimation

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages