yifutao's picture
Improve dataset card: Add task categories, sample usage, HF paper link, and citation (#2)
0020bc7 verified
metadata
license: cc-by-nc-sa-4.0
task_categories:
  - robotics
  - image-to-3d
tags:
  - slam
  - lidar
  - 3d-reconstruction
  - nerf
  - 3d-gaussian-splatting
  - localization
  - sfm
  - mvs
  - multimodal
  - oxford

We present the Oxford Spires Dataset, captured in and around well-known landmarks in Oxford using a custom-built multi-sensor perception unit as well as a millimetre-accurate map from a terrestrial LiDAR scanner (TLS). The perception unit includes three global shutter colour cameras, an automotive 3D LiDAR scanner, and an inertial sensor — all precisely calibrated.

Sample Usage

Download the Dataset

You can download the dataset from Hugging Face using the provided script. You can specify which folders to download by changing the example_pattern. Core sequences are also defined in the script.

python scripts/dataset_download.py

Install Python Tools

Install oxspires_tools to access Python utilities for using the dataset:

pip install .

To enable C++/Python bindings (requires PCL and Octomap):

BUILD_CPP=1 pip install .

Alternatively, use the provided Docker container:

docker compose -f .docker/oxspires/docker-compose.yml run --build oxspires_utils

Generate Depth Images

The following script downloads synchronised images and LiDAR data from a sequence on Hugging Face and generates depth images, LiDAR overlaid on camera images, and surface normal images:

python scripts/generate_depth.py

Citation

If you use The Oxford Spires Dataset in your research, please cite the following paper:

@article{tao2025spires,
title={The Oxford Spires Dataset: Benchmarking Large-Scale LiDAR-Visual Localisation, Reconstruction and Radiance Field Methods},
author={Tao, Yifu and Mu{\~n}oz-Ba{\~n}{\'o}n, Miguel {\'A}ngel and Zhang, Lintong and Wang, Jiahao and Fu, Lanke Frank Tarimo and Fallon, Maurice},
journal={International Journal of Robotics Research}, 
year={2025},
}