You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+3Lines changed: 3 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -29,6 +29,9 @@ The code-base has additional support for:
29
29
* Total Variation Loss for smoother embeddings (use `--tv-loss-weight` to enable)
30
30
* Sparsity-inducing loss on the ray weights (use `--sparse-loss-weight` to enable)
31
31
32
+
## ScanNet dataset support
33
+
The repo now supports training a NeRF model on a scene from the ScanNet dataset. I personally found setting up the ScanNet dataset to be a bit tricky. Please find some instructions/notes in [ScanNet.md](ScanNet.md).
I personally found it a bit tricky to setup the ScanNet dataset the first time I tried it. So, I am compiling some notes/instructions on how to do it in case someone finds it useful.
4
+
5
+
### 1. Dataset download
6
+
7
+
To download ScanNet data and its labels, follow the instructions [here](https://github.com/ScanNet/ScanNet). Basically, fill out the ScanNet Terms of Use agreement and email it to [scannet@googlegroups.com](mailto:scannet@googlegroups.com). You will receive a download link to the dataset. Download the dataset and unzip it.
8
+
9
+
### 2. Use [SensReader](https://github.com/ScanNet/ScanNet/tree/master/SensReader/python) to extract RGB-D and camera data
10
+
Use the `reader.py` script as follows for each scene you want to work with:
11
+
```
12
+
python reader.py --filename [.sens file to export data from] --output_path [output directory to export data to]
13
+
Options:
14
+
--export_depth_images: export all depth frames as 16-bit pngs (depth shift 1000)
15
+
--export_color_images: export all color frames as 8-bit rgb jpgs
16
+
--export_poses: export all camera poses (4x4 matrix, camera to world)
17
+
--export_intrinsics: export camera intrinsics (4x4 matrix)
18
+
```
19
+
20
+
### 3. Then, use this [script](https://github.com/zju3dv/object_nerf/blob/main/data_preparation/scannet_sens_reader/convert_to_nerf_style_data.py) to convert the data to NeRF-style format. For instructions, see Step 1 [here](https://github.com/zju3dv/object_nerf/tree/main/data_preparation).
21
+
1. The generated transforms_xxx.json comes with transformation matrix (from camera coordinate to world coordinate) in SLAM / OpenCV format (xyz -> right down forward). You need to change to NDC format (xyz -> right up back) in the dataloader for training with NeRF convention.
22
+
2. For example, see the conversion done [here](https://github.com/cvg/nice-slam/blob/7af15cc33729aa5a8ca052908d96f495e34ab34c/src/utils/datasets.py#L205).
0 commit comments