Skip to content

Latest commit

 

History

History
121 lines (95 loc) · 10.9 KB

README.md

File metadata and controls

121 lines (95 loc) · 10.9 KB

IMPORTANT An updated dataset is now available which fixes a critical issue with 3D60, the lighting bias introduced by the light source placed at the origin. More information can be found at the Pano3D project page.

OmniDepth Conference Project Page

Spherical View Synthesis Conference Project Page

Surface Regression Conference Project Page

3D60 Toolset

A set of tools for working with the 3D60 dataset:

  • PyTorch data loaders
  • Dataset splits generation scripts

The 3D60 dataset was generated by ray-casting existing 3D datasets, making it a derivative of:

Requirements

This code has been tested with:

Besides PyTorch, the following Python packages are needed:

Data Loading

An example data loading usage can be found in visualize_dataset.py where the dataset is loaded and visualized using visdom.

Given that 3D60 can be used in a variety of machine learning tasks, data loading can be modified w.r.t which datasets, image_types and placements (for stereo viewpoints) will be loaded by constructing customized dataset iterators:

dataset_iterator = ThreeD60.get_datasets(".//splits//3D60_train.txt",\
	    datasets=["suncg", "m3d", "s2d3d"],\
	    placements=[ThreeD60.Placements.CENTER, ThreeD60.Placements.RIGHT, ThreeD60.Placements.UP],\
	    image_types=[ThreeD60.ImageTypes.COLOR, ThreeD60.ImageTypes.DEPTH, ThreeD60.ImageTypes.NORMAL],\
	    longitudinal_rotation=True)

In addition, randomized horizontal rotation augmentations can also be triggered with the longitudinal_rotation flag.

The returned iterator can be used to construct a PyTorch DataLoader:

dataset_loader = torch.utils.data.DataLoader(dataset_iterator,\
  batch_size=32, shuffle=True, pin_memory=False, num_workers=4)

Specific image tensors can be extracted from the returned dictionary with the extract_image function:

for i, b in  enumerate(dataset_loader):
	center_color_image = ThreeD60.extract_image(b, ThreeD60.Placements.CENTER, ThreeD60.ImageTypes.COLOR)
	center_depth_map = ThreeD60.extract_image(b, ThreeD60.Placements.CENTER, ThreeD60.ImageTypes.DEPTH)
	center_normal_map = ThreeD60.extract_image(b, ThreeD60.Placements.CENTER, ThreeD60.ImageTypes.NORMAL)

Splits

Published

A set of published splits are provided which are used in the corresponding works:

  • 360D ECCV18 (used in OmniDepth [4])
  • 3D60 3DV19 (with a smaller synthetic part used in [5] & [6])
  • v1 (with a larger synthetic part)

These splits rely on the official splits of the real datasets -- i.e. Matterport3D and Stanford2D3D (fold#1) -- but use a random selection of scenes from the synthetic SunCG dataset.

Custom

We also offer a set of scripts to generate new splits:

This script calculates a depth value distribution histogram w.r.t. a --max_depth argument value (default: 10.0m) as well as the percentage of values lower than 0.5m and over 5.0m for each of the datasets. The resulting .csv files are saved in the --stats_path argument folder (default: ./splits/), one for each rendered 3D dataset and prefixed with their codename: "m3d", "s2d3d" and "suncg". The paths containing the rendered data for each dataset are provided with the --m3d_path, --s2d3d_path and --suncg_path for Matterport3D, Stanford2D3D and SunCG respectively.

This script has two modes based on the --action argument:

  • 'calc': Finds and saves outlier renders based on a set of heuristics w.r.t. their near (--lower_threshold) and far (--upper_threshold) depth value distributions. When the percentage of total pixels of each examined depth map exceeds specific percentage bounds either under the near value threshold or over the far value threshold, it is considered as an outlier or bad render. The depth maps are located in each dataset's respective folder provided by the codename prefixed arguments --m3d_path, --s2d3d_path and --suncg_path similar to calculate_statistics.py. This can happen due to incomplete scans, scanning artifacts and errors, unfinished 3D modeling or missing assets. Different lower and upper bounds can be set for each dataset as their typical depth distribution values differ. They are fractional percentages, set using codename prefixed lower and upper bound arguments:

    • --m3d_lower_bound and --m3d_upper_bound for Matterport3D,
    • --s2d3d_lower_bound and --s2d3d_upper_bound for Stanford2D3D, and
    • --suncg_lower_bound and --suncg_upper_bound for SunCG.

    The resulting .csv files contain the file names of the outlier renders, prefixed with each dataset's codename and saved in the --outliers_path argument folder (default: ./splits/).

  • 'save': Stores the calculated outliers read from the --outliers_path argument folder in .png images saved in the same path. The images contain multiple tiled outliers for quick visual inspection of the results.

This script creates the split (i.e. train, test and val) files that in turn contain the filenames of each rendered modality and placement for each viewpoint. The split files are prefixed with the --name argument and saved in the --outliers_path folder, where the outlier files are also read from. The script has the train/test/validation set splits from Matterport3D and Stanford2D3D's fold#1 hardcoded in it, but uses a random selection for SunCG. It scans each dataset's folder, as provided by the codename prefixed arguments --m3d_path, --s2d3d_path and --suncg_path similar to calculate_statistics.py and find_outliers.py, and ignores the outliers as read from the available files for each dataset.

The splits available in ./splits/v1 were generated by running the aforementioned scripts in order and using their default parameters. However, they can also be used to create new splits using different parameterisations for outlier rejection, based on custom distance thresholds, or by ignoring specific datasets. If any of the codename prefixed arguments --m3d_path, --s2d3d_path and --suncg_path are not provided, they are skipped from all steps/scripts, and thus, single dataset splits can also be generated (i.e. for leave-one-out experiments).

Important Note: Taking this into account, consistency between experiments may not always be possible without using the aforementioned splits or when involving custom splits that contain synthetic samples. However, the realistic (i.e. Matterport3D and Stanford2D3D) samples test and validation sets should be consistent between published and custom splits.

References

[1] Chang, A., Dai, A., Funkhouser, T., Halber, M., Niessner, M., Savva, M., Song, S., Zeng, A. and Zhang, Y. (2017). Matterport3d: Learning from rgb-d data in indoor environments. In Proceedings of the International Conference on 3D Vision (3DV).

[2] Armeni, I., Sax, S., Zamir, A.R. and Savarese, S., 2017. Joint 2d-3d-semantic data for indoor scene understanding. arXiv preprint arXiv:1702.01105.

[3] Song, S., Yu, F., Zeng, A., Chang, A.X., Savva, M. and Funkhouser, T., 2017. Semantic scene completion from a single depth image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4] Zioulis, N.*, Karakottas, A.*, Zarpalas, D., and Daras, P. (2018). Omnidepth: Dense depth estimation for indoors spherical panoramas. In Proceedings of the European Conference on Computer Vision (ECCV).

[5] Zioulis, N., Karakottas, A., Zarpalas, D., Alvarez, F., and Daras, P. (2019). Spherical View Synthesis for Self-Supervised 360o Depth Estimation. In Proceedings of the International Conference on 3D Vision (3DV).

[6] Karakottas, A., Zioulis, N., Samaras, S., Ataloglou, D., Gkitsas, V., Zarpalas, D., and Daras, P. (2019). 360o Surface Regression with a Hyper-sphere Loss. In Proceedings of the International Conference on 3D Vision (3DV).