Abstract
Purpose
Laparoscopic interventions require the precise navigation of medical instruments through the patient’s body, while taking critical structures into account. Although numerous concepts have been proposed for displaying subsurface anatomical detail using augmented reality, clinical translation of these methods has suffered from a lack of robustness as well as from cumbersome integration into the clinical workflow. The purpose of this study was to investigate the feasibility of a new approach to intra-operative registration based on fluorescent markers.
Methods
The proposed approach to augmented reality visualization relies on metabolizable fluorescent markers that are attached to the target organ to guide a 2D/3D intra-operative registration algorithm. In an ex vivo porcine study, marker tracking performance is evaluated in the presence of smoke, blood, and tissue in the field of view of the endoscope.
Results
In contrast to state-of-the-art needle-shaped fiducial markers, the fluorescent markers can be reliably tracked when occluded by smoke, blood or tissue. This makes the new 2D/3D intra-operative registration approach considerably more robust than state-of-the-art marker-based methods.
Conclusion
As the concept can be smoothly integrated into the clinical workflow, its potential for application in clinical laparoscopy is high.
Similar content being viewed by others
Notes
This concept is part of patent application no. 15 193 169.8 at the European Patent Office and preliminary results have been presented at the German conference [25].
cv::findContours.
cv::minEnclosingCircle.
References
Alander JT, Kaartinen I, Laakso A, Pätilä T, Spillmann T, Tuchin VV, Venermo M, Välisuo P (2012) A review of indocyanine green fluorescent imaging in surgery. J Biomed Imaging 2012:7
Baumhauer M, Feuerstein M, Meinzer HP, Rassweiler J (2008) Navigation in endoscopic soft tissue surgery: perspectives and limitations. J Endourol 22(4):751–766
Baumhauer M, Simpfendoerfer T, Schwarz R, Seitel M, Müller-Stich B, Gutt C, Rassweiler J, Meinzer HP, Wolf I (2007) Soft tissue navigation for laparoscopic prostatectomy: Evaluation of camera pose estimation for enhanced visualization. Proc. SPIE 6509, Medical Imaging 2007:650911. doi:10.1117/12.709655
van den Berg NS, van Leeuwen FW, van der Poel HG (2012) Fluorescence guidance in urologic surgery. Curr Opin Urol 22(2):109–120
Cleary K, Peters TM (2010) Image-guided interventions: technology review and clinical applications. Annu Rev Biomed Eng 12:119–142
Dalvi A, Faria M, Pinto A (1986) Non-suture closure of wound using cyanoacrylate. J Postgrad Med 32(2):97
Dementhon DF, Davis LS (1995) Model-based object pose in 25 lines of code. Int J Comput Vis 15(1–2):123–141
Eder M, Schafer M, Bauder-Wust U, Hull WE, Wangler C, Mier W, Haberkorn U, Eisenhut M (2012) 68ga-complex lipophilicity and the targeting property of a urea-based PSMA inhibitor for pet imaging. Bioconjug Chem 23(4):688–697
Gioux S, Choi HS, Frangioni JV (2010) Image-guided surgery using invisible near-infrared light: fundamentals of clinical translation. Mol Imaging 9(5):237
Glocker B, Sotiras A, Komodakis N, Paragios N (2011) Deformable medical image registration: setting the state of the art with discrete methods*. Annu Rev Biomed Eng 13:219–244
Itseez: Open source computer vision (2015). http://opencv.org
Lepetit V, Moreno-Noguer F, Fua P (2009) Epnp: an accurate o (n) solution to the pnp problem. Int J Comput Vis 81(2):155–166
Lu CP, Hager GD, Mjolsness E (2000) Fast and globally convergent pose estimation from video images. IEEE Trans Pattern Anal 22(6):610–622
Maier-Hein L, Müller SA, Pianka F, Wörz S, Müller-Stich BP, Seitel A, Rohr K, Meinzer HP, Schmied BM, Wolf I (2008) Respiratory motion compensation for ct-guided interventions in the liver. Comput Aided Surg 13(3):125–138
Markelj P, Tomaževič D, Likar B, Pernuš F (2012) A review of 3d/2d registration methods for image-guided interventions. Med Image Anal 16(3):642–661
Müller M (2013) Augmented reality navigation for percutaneous nephrolithotomy on mobile devices. Ph.D. thesis, Universität Heidelberg. http://www.ub.uni-heidelberg.de/archiv/16102
Nolden M, Zelzer S, Seitel A, Wald D, Müller M, Franz AM, Maleike D, Fangerau M, Baumhauer M, Maier-Hein L, Maier-Hein K, Meinzer HP, Wolf I (2013) The medical imaging interaction toolkit: challenges and advances. Int J Comput Assist Radiol Surg 8(4):607–620
Park H, Park JI (2005) Invisible marker based augmented reality system. Proc. SPIE 5960, Visual Communications and Image Processing 2005, 59601I. doi:10.1117/12.631416
Pluim JP, Maintz JA, Viergever M (2003) Mutual-information-based registration of medical images: a survey. IEEE Trans Med Imaging 22(8):986–1004
Prahl S (2015) Optical absorption of hemoglobin. http://omlc.org/spectra/hemoglobin/
Rucker DC, Wu Y, Clements LW, Ondrake JE, Pheiffer TS, Simpson AL, Jarnagin WR, Miga M (2014) A mechanics-based nonrigid registration method for liver surgery using sparse intraoperative data. IEEE Trans Med Imaging 33(1):147–158
Schweighofer G, Pinz A (2006) Robust pose estimation from a planar target. IEEE Trans Pattern Anal 28(12):2024–2030
Simpfendörfer T, Baumhauer M, Müller M, Gutt CN, Meinzer HP, Rassweiler JJ, Guven S, Teber D (2011) Augmented reality visualization during laparoscopic radical prostatectomy. J Endourol 25(12):1841–1845
Suwelack S, Röhl S, Bodenstedt S, Reichard D, Dillmann R, dos Santos T, Maier-Hein L, Wagner M, Wünscher J, Kenngott H, Müller BP, Speidel S (2014) Physics-based shape matching for intraoperative image guidance. Med Phys 41(11):111,901
Wild E, Teber D, Schmid D, Simpfendörfer T, Müller M, Kenngott H, Maier-Hein L (2016) Robuste intraoperative Registrierung mit fluoreszierenden Markern für die computergestützte Laparoskopie. Bildverarbeitung für die Medizin 2016:74–79. doi:10.1007/978-3-662-49465-3_15
Zhang Z (2000) A flexible new technique for camera calibration. IEEE Trans Pattern Anal 22(11):1330–1334
Zheng Y, Kuang Y, Sugimoto S, Astrom K, Okutomi M (2013) Revisiting the pnp problem: A fast, general and optimal solution. In: 2013 IEEE international conference on computer vision (ICCV), pp 2344–2351
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Human and animal rights
This article does not contain any studies with human participants or animals performed by any of the authors.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
About this article
Cite this article
Wild, E., Teber, D., Schmid, D. et al. Robust augmented reality guidance with fluorescent markers in laparoscopic surgery. Int J CARS 11, 899–907 (2016). https://doi.org/10.1007/s11548-016-1385-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11548-016-1385-4