Abstract
Visual navigation systems have increasingly been adopted in many urban safety–critical applications, such as urban air mobility and highly automated vehicle, for which they must continuously provide accurate and safety-assured pose estimates. Extensive studies have focused on improving visual navigation accuracy and robustness in complex environment, while insufficient attention has been paid to ensuring navigation safety in the presence of outliers. From safety perspective, integrity is the most important navigation performance criterion because it measures the trust that can be placed in the correctness of the navigation output. Through leveraging the concept of integrity, this paper develops an integrity monitoring framework to protect visual navigation system against misleading measurements and to quantify the reliability of the navigation output. We firstly present the iterative least squares (LS)-based pose estimation algorithm and derive the associated covariance estimation methodology. Then we develop a two-layer fault detection scheme through combining random sampling consensus (RANSAC) with multiple hypotheses solution separation (MHSS) to achieve high efficiency and high reliability. Finally, the framework determines the probabilistic error bound of the navigation output that rigorously captures the undetected faults and the measurement uncertainty. The proposed algorithms are validated using various simulations, and the results suggest the promising performance.
Similar content being viewed by others
References
Aqel MOA, Marhaban MH, Saripan MI, Ismail NB (2016) Review of visual odometry: types, approaches, challenges, and applications. SpringerPlus 5(1897):1–26
Arana GD, Joerger M, Spenko M (2019) Efficient Integrity Monitoring for KF-based Localization. In: International Conference on Robotics and Automation (ICRA), vol 2019, pp 6374–6380
Bhamidipati, S., & Gao, G. X. (2019). SLAM-based integrity monitoring using GPS and fish-eye camera, pp 1–14. arXiv:1910.02165
Blanch J, Walker T, Enge P et al (2015) Baseline advanced RAIM user algorithm and possible improvements. IEEE Trans Aerosp Electron Syst 51(1):713–732
Brown R (1992) A baseline GPS RAIM scheme and a note on the equivalence of three RAIM methods. Navigation 39(3):301–316
Geiger A, Lenz P, Stiller C, Urtasun R (2013) Vision meets robotics: the KITTI dataset. Int J Robot Res 32(11):1231–1237
Huang, B., Zhao, J., & Liu, J. (2019). A survey of simultaneous localization and mapping with an envision in 6G wireless networks, pp 1–17. arXiv:1909.05214
Joerger M, Pervan B (2019) Quantifying safety of laser-based navigation. IEEE Trans Aerosp Electron Syst 55(1):273–288
Kitt B, Geiger A, Lategahn H (2010) Visual odometry based on stereo image sequences with RANSAC-based outlier rejection scheme. In: Proceedings of IEEE Intelligent Vehicles Symposium, pp 486–492
Li C, Waslander SL (2019) Visual measurement integrity monitoring for UAV localization. In: Proceedings of 2019 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), pp 22–29
Li Z, Wang J (2015) Comparison of multiple fault detection methods for monocular visual navigation with 3D maps. In: 2014 Ubiquitous positioning indoor navigation and location based service, UPINLBS 2014—Conference Proceedings, pp 228–237
Mostafa M, Zahran S, Moussa A, El-Sheimy N, Sesay A (2018) Radar and visual odometry integrated system aided navigation for UAVS in GNSS denied environment. Sensors (Switzerland) 18(9):1–29
Mur-Artal R, Tardos JD (2017) ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans Robot 33(5):1255–1262
Tanıl Ç, Khanafseh S, Joerger M, Kujur B, Kruger B, De Groot L, Pervan B (2019) Optimal INS/GNSS coupling for autonomous car positioning integrity. In: Proceedings of the 32nd International Technical Meeting of the Satellite Division of the Institute of Navigation, ION GNSS+ vol 2019, pp 3123–3140
Zhu N, Marais J, Betaille D, Berbineau M (2018) GNSS position integrity in urban environments: a review of literature. IEEE Trans Intell Transp Syst 19(9):2762–2778
Acknowledgements
This study was supported by Shanghai Jiao Tong University (SJTU) Global Strategic Partnership Fund (2019 SJTU–UoT).
Author information
Authors and Affiliations
Corresponding author
Appendix: Proof of Eq. (24)
Appendix: Proof of Eq. (24)
\(\delta \boldsymbol{\varphi }\) is usually called misalignment angles, while \({{\varvec{\varepsilon}}}_{\varphi }\) denotes the errors in attitude angles. The relationship between them is given by the following:
where \({{\varvec{C}}}_{b}^{n}={{\varvec{R}}}^{\mathrm{T}}\) and \({{\varvec{C}}}_{A}^{\omega }\) is shown as:
This equation is obtained from the attitude kinematic equation. Substituting (40) to (39),
Therefore, we have:
Rights and permissions
About this article
Cite this article
Wang, S., Zhan, X., Fu, Y. et al. Feature-based visual navigation integrity monitoring for urban autonomous platforms. AS 3, 167–179 (2020). https://doi.org/10.1007/s42401-020-00057-8
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s42401-020-00057-8