Update README
Browse files- README.md +66 -0
- assets/teaser-full.png +3 -0
README.md
CHANGED
@@ -1,3 +1,69 @@
|
|
1 |
---
|
|
|
|
|
2 |
license: cc-by-4.0
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
pretty_name: Deblur e-NeRF
|
3 |
+
paperswithcode_id: deblur-e-nerf-synthetic-event-dataset
|
4 |
license: cc-by-4.0
|
5 |
+
viewer: false
|
6 |
+
size_categories:
|
7 |
+
- n<1K
|
8 |
---
|
9 |
+
|
10 |
+
# Deblur *e*-NeRF Synthetic Event Dataset
|
11 |
+
|
12 |
+
<table style="display: block">
|
13 |
+
<tr>
|
14 |
+
<td><a href="https://wengflow.github.io/deblur-e-nerf"><img src="https://img.shields.io/badge/Project_Page-black?style=for-the-badge" alt="Project Page"></a></td>
|
15 |
+
<td><a href="https://arxiv.org/abs/2409.17988"><img src="https://img.shields.io/badge/arXiv-black?style=for-the-badge" alt="arXiv"></a></td>
|
16 |
+
<td><a href="https://github.com/wengflow/deblur-e-nerf"><img src="https://img.shields.io/badge/Code-black?style=for-the-badge" alt="Code"></a></td>
|
17 |
+
<td><a href="https://github.com/wengflow/rpg_esim"><img src="https://img.shields.io/badge/Simulator-black?style=for-the-badge" alt="Simulator"></a></td>
|
18 |
+
</tr>
|
19 |
+
</table>
|
20 |
+
|
21 |
+
<p align="center">
|
22 |
+
<img src="assets/teaser-full.png" alt="Teaser" width=100%/>
|
23 |
+
</p>
|
24 |
+
|
25 |
+
This repository contains the synthetic event dataset used in [**Deblur *e*-NeRF**](https://wengflow.github.io/deblur-e-nerf) to study the collective effect of camera speed and scene illuminance on the quality of NeRF reconstruction from a moving event camera. It is an extension of the synthetic event dataset used in [**Robust *e*-NeRF**](https://wengflow.github.io/robust-e-nerf). The dataset is simulated using an [improved version of ESIM](https://github.com/wengflow/rpg_esim) with three different camera configurations of increasing difficulty levels (*i.e.* *easy*, *medium* and *hard*) on seven Realistic Synthetic 360 scenes (adopted in the synthetic experiments of NeRF), resulting in a total of 21 sequence recordings. Please refer to the [Deblur *e*-NeRF paper](https://arxiv.org/abs/2409.17988) for more details.
|
26 |
+
|
27 |
+
This synthetic event dataset allows for a retrospective comparison between event-based and image-based NeRF reconstruction methods, as the event sequences were simulated under highly similar conditions as the NeRF synthetic dataset. In particular, we adopt the same camera intrinsics and camera distance to the object at the origin. Furthermore, the event camera travels in a hemi-/spherical spiral motion about the object, thereby having a similar camera pose distribution for training. Apart from that, we also use the same test camera poses/views. Nonetheless, this new synthetic event dataset is not only specific to NeRF reconstruction, but also suitable for novel view synthesis, 3D reconstruction, localization and SLAM in general.
|
28 |
+
|
29 |
+
If you use this synthetic event dataset for your work, please cite:
|
30 |
+
|
31 |
+
```bibtex
|
32 |
+
@inproceedings{low2024_deblur-e-nerf,
|
33 |
+
title = {Deblur e-NeRF: NeRF from Motion-Blurred Events under High-speed or Low-light Conditions},
|
34 |
+
author = {Low, Weng Fei and Lee, Gim Hee},
|
35 |
+
booktitle = {European Conference on Computer Vision (ECCV)},
|
36 |
+
year = {2024}
|
37 |
+
}
|
38 |
+
```
|
39 |
+
|
40 |
+
## Dataset Structure and Contents
|
41 |
+
|
42 |
+
This synthetic event dataset is organized first by scene, then by level of difficulty. Each sequence recording is given in the form of a [ROS bag](http://wiki.ros.org/rosbag) named `esim.bag`, with the following data streams:
|
43 |
+
|
44 |
+
| ROS Topic | Data | Publishing Rate (Hz) |
|
45 |
+
| :--- | :--- | :--- |
|
46 |
+
| `/cam0/events` | Events | - |
|
47 |
+
| `/cam0/pose` | Camera Pose | 1000 |
|
48 |
+
| `/imu` | IMU measurements with simulated noise | 1000 |
|
49 |
+
| `/cam0/image_raw` | RGB image | 250 |
|
50 |
+
| `/cam0/depthmap` | Depth map | 10 |
|
51 |
+
| `/cam0/optic_flow` | Optical flow map | 10 |
|
52 |
+
| `/cam0/camera_info` | Camera intrinsic and lens distortion parameters | 10
|
53 |
+
|
54 |
+
It is obtained by running the improved ESIM with the associated `esim.conf` configuration file, which references camera intrinsics configuration files `pinhole_mono_nodistort_f={1111, 1250}.yaml` and camera trajectory CSV files `{hemisphere, sphere}_spiral-rev=4[...].csv`.
|
55 |
+
|
56 |
+
The validation and test views of each scene are given in the `views/` folder, which is structured according to the NeRF synthetic dataset (except for the depth and normal maps). These views are rendered from the scene Blend-files, given in the `scenes/` folder. Specifically, we create a [Conda](https://docs.conda.io/en/latest/) environment with [Blender as a Python module](https://docs.blender.org/api/current/info_advanced_blender_as_bpy.html) installed, according to [these instructions](https://github.com/wengflow/rpg_esim#blender), to run the `bpy_render_views.py` Python script for rendering the evaluation views.
|
57 |
+
|
58 |
+
## Setup
|
59 |
+
|
60 |
+
1. Install [Git LFS](https://git-lfs.com/) according to the [official instructions](https://github.com/git-lfs/git-lfs?utm_source=gitlfs_site&utm_medium=installation_link&utm_campaign=gitlfs#installing).
|
61 |
+
2. Setup Git LFS for your user account with:
|
62 |
+
```bash
|
63 |
+
git lfs install
|
64 |
+
```
|
65 |
+
3. Clone this dataset repository into the desired destination directory with:
|
66 |
+
```bash
|
67 |
+
git lfs clone https://huggingface.co/datasets/wengflow/deblur-e-nerf
|
68 |
+
```
|
69 |
+
4. To minimize disk usage, remove the `.git/` folder. However, this would complicate the pulling of changes in this upstream dataset repository.
|
assets/teaser-full.png
ADDED
Git LFS Details
|