Awesome
<br/> <div align="center"> <img src="resources/XRSLAM.png" width="600"/> </div> <br/> <div align="left"> <div align="left"> </div>Introduction
OpenXRLab Visual-inertial SLAM Toolbox and Benchmark. It is a part of the OpenXRLab project.
Major Features
- Robust and Lightweight optimization-based Visual Inertial Odometry
- Both desktop and mobile platforms are supported
- An interactive and real-time AR application on iPhone
- Visual localization module for running XRARDemo on prebuilt scenarios
- Compared to other state-of-the-art systems, XRSLAM achieves competitive accuracy
Installation
We provide detailed installation tutorial for XRSLAM, users can install from scratch or use provided Dockerfile.
Getting Started
Please refer to quick start for the basic usage of XRSLAM.
License
The license of our codebase is Apache-2.0. Note that this license only applies to code in our library, the dependencies of which are separate and individually licensed. We would like to pay tribute to open-source implementations to which we rely on. Please be aware that using the content of dependencies may affect the license of our codebase. Some supported methods may carry additional licenses.
FAQ
Please refer to FAQ for frequently asked questions.
Citation
If you use this toolbox or benchmark in your research, please cite this project.
@misc{xrslam,
title={OpenXRLab Visual-inertial SLAM Toolbox and Benchmark},
author={XRSLAM Contributors},
howpublished = {\url{https://github.com/openxrlab/xrslam}},
year={2022}
}
If you use the Robust Visual-Inertial Odometry in your research, please cite:
@article{li2023rd,
title={RD-VIO: Robust Visual-Inertial Odometry for Mobile Augmented Reality in Dynamic Environments},
author={Li, Jinyu and Pan, Xiaokun and Huang, Gan and Zhang, Ziyang and Wang, Nan and Bao, Hujun and Zhang, Guofeng},
journal={arXiv preprint arXiv:2310.15072},
year={2023}
}
Contributing
We appreciate all contributions to improve XRSLAM. Please refer to CONTRIBUTING.md for the contributing guideline.
Acknowledgement
XRSLAM is an open source project that is contributed by researchers and engineers from both the academia and the industry. We appreciate all the contributors who implement their methods or add new features, as well as users who give valuable feedbacks. We wish that the toolbox and benchmark could serve the growing research community by providing a flexible toolkit to reimplement existing methods and develop their own new models.
Projects in OpenXRLab
- XRPrimer: OpenXRLab foundational library for XR-related algorithms.
- XRSLAM: OpenXRLab Visual-inertial SLAM Toolbox and Benchmark.
- XRSfM: OpenXRLab Structure-from-Motion Toolbox and Benchmark.
- XRLocalization: OpenXRLab Visual Localization Toolbox and Server.
- XRMoCap: OpenXRLab Multi-view Motion Capture Toolbox and Benchmark.
- XRMoGen: OpenXRLab Human Motion Generation Toolbox and Benchmark.
- XRNeRF: OpenXRLab Neural Radiance Field (NeRF) Toolbox and Benchmark.