Skip to main content
Log in

Feature-based visual navigation integrity monitoring for urban autonomous platforms

  • Original Paper
  • Published:
Aerospace Systems Aims and scope Submit manuscript

Abstract

Visual navigation systems have increasingly been adopted in many urban safety–critical applications, such as urban air mobility and highly automated vehicle, for which they must continuously provide accurate and safety-assured pose estimates. Extensive studies have focused on improving visual navigation accuracy and robustness in complex environment, while insufficient attention has been paid to ensuring navigation safety in the presence of outliers. From safety perspective, integrity is the most important navigation performance criterion because it measures the trust that can be placed in the correctness of the navigation output. Through leveraging the concept of integrity, this paper develops an integrity monitoring framework to protect visual navigation system against misleading measurements and to quantify the reliability of the navigation output. We firstly present the iterative least squares (LS)-based pose estimation algorithm and derive the associated covariance estimation methodology. Then we develop a two-layer fault detection scheme through combining random sampling consensus (RANSAC) with multiple hypotheses solution separation (MHSS) to achieve high efficiency and high reliability. Finally, the framework determines the probabilistic error bound of the navigation output that rigorously captures the undetected faults and the measurement uncertainty. The proposed algorithms are validated using various simulations, and the results suggest the promising performance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

References

  1. Aqel MOA, Marhaban MH, Saripan MI, Ismail NB (2016) Review of visual odometry: types, approaches, challenges, and applications. SpringerPlus 5(1897):1–26

    Google Scholar 

  2. Arana GD, Joerger M, Spenko M (2019) Efficient Integrity Monitoring for KF-based Localization. In: International Conference on Robotics and Automation (ICRA), vol 2019, pp 6374–6380

  3. Bhamidipati, S., & Gao, G. X. (2019). SLAM-based integrity monitoring using GPS and fish-eye camera, pp 1–14. arXiv:1910.02165

  4. Blanch J, Walker T, Enge P et al (2015) Baseline advanced RAIM user algorithm and possible improvements. IEEE Trans Aerosp Electron Syst 51(1):713–732

    Article  Google Scholar 

  5. Brown R (1992) A baseline GPS RAIM scheme and a note on the equivalence of three RAIM methods. Navigation 39(3):301–316

    Article  Google Scholar 

  6. Geiger A, Lenz P, Stiller C, Urtasun R (2013) Vision meets robotics: the KITTI dataset. Int J Robot Res 32(11):1231–1237

    Article  Google Scholar 

  7. Huang, B., Zhao, J., & Liu, J. (2019). A survey of simultaneous localization and mapping with an envision in 6G wireless networks, pp 1–17. arXiv:1909.05214

  8. Joerger M, Pervan B (2019) Quantifying safety of laser-based navigation. IEEE Trans Aerosp Electron Syst 55(1):273–288

    Article  Google Scholar 

  9. Kitt B, Geiger A, Lategahn H (2010) Visual odometry based on stereo image sequences with RANSAC-based outlier rejection scheme. In: Proceedings of IEEE Intelligent Vehicles Symposium, pp 486–492

  10. Li C, Waslander SL (2019) Visual measurement integrity monitoring for UAV localization. In: Proceedings of 2019 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), pp 22–29

  11. Li Z, Wang J (2015) Comparison of multiple fault detection methods for monocular visual navigation with 3D maps. In: 2014 Ubiquitous positioning indoor navigation and location based service, UPINLBS 2014—Conference Proceedings, pp 228–237

  12. Mostafa M, Zahran S, Moussa A, El-Sheimy N, Sesay A (2018) Radar and visual odometry integrated system aided navigation for UAVS in GNSS denied environment. Sensors (Switzerland) 18(9):1–29

    Article  Google Scholar 

  13. Mur-Artal R, Tardos JD (2017) ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans Robot 33(5):1255–1262

    Article  Google Scholar 

  14. Tanıl Ç, Khanafseh S, Joerger M, Kujur B, Kruger B, De Groot L, Pervan B (2019) Optimal INS/GNSS coupling for autonomous car positioning integrity. In: Proceedings of the 32nd International Technical Meeting of the Satellite Division of the Institute of Navigation, ION GNSS+ vol 2019, pp 3123–3140

  15. Zhu N, Marais J, Betaille D, Berbineau M (2018) GNSS position integrity in urban environments: a review of literature. IEEE Trans Intell Transp Syst 19(9):2762–2778

    Article  Google Scholar 

Download references

Acknowledgements

This study was supported by Shanghai Jiao Tong University (SJTU) Global Strategic Partnership Fund (2019 SJTU–UoT).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xingqun Zhan.

Appendix: Proof of Eq. (24)

Appendix: Proof of Eq. (24)

\(\delta \boldsymbol{\varphi }\) is usually called misalignment angles, while \({{\varvec{\varepsilon}}}_{\varphi }\) denotes the errors in attitude angles. The relationship between them is given by the following:

$$ \delta {\varvec{\varphi }} = {\varvec{C}}_{b}^{n} \, \cdot \,{\varvec{C}}_{A}^{\omega } \, \cdot \,{\varvec{\varepsilon}}_{\varphi } , $$
(39)

where \({{\varvec{C}}}_{b}^{n}={{\varvec{R}}}^{\mathrm{T}}\) and \({{\varvec{C}}}_{A}^{\omega }\) is shown as:

$$ {\varvec{C}}_{A}^{\omega } = \left[ {\begin{array}{*{20}c} 1 & 0 & { - s\beta } \\ 0 & {c\alpha } & {s\alpha c\beta } \\ 0 & { - s\alpha } & {c\alpha c\beta } \\ \end{array} } \right]. $$
(40)

This equation is obtained from the attitude kinematic equation. Substituting (40) to (39),

$$ \delta {\varvec{\varphi }} = \left[ {\begin{array}{*{20}c} {c\beta c\gamma } & { - s\gamma } & 0 \\ {c\beta s\gamma } & {c\gamma } & 0 \\ { - s\beta } & 0 & 1 \\ \end{array} } \right]\, \cdot \,{\varvec{\varepsilon}}_{\varphi } . $$
(41)

Therefore, we have:

$$ {\varvec{\varepsilon}}_{\varphi } = \left[ {\begin{array}{*{20}c} {c\gamma /c\beta } & {s\gamma /c\beta } & 0 \\ { - s\gamma } & {c\gamma } & 0 \\ {c\gamma s\beta /c\beta } & {s\gamma s\beta /c\beta } & 1 \\ \end{array} } \right]\, \cdot \,\delta {\varvec{\varphi }} $$
(42)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, S., Zhan, X., Fu, Y. et al. Feature-based visual navigation integrity monitoring for urban autonomous platforms. AS 3, 167–179 (2020). https://doi.org/10.1007/s42401-020-00057-8

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s42401-020-00057-8

Keywords

Navigation