Abstract
Range imaging devices have already shown their value for patient setup and motion management in external beam radiation therapy. However current systems need several range imaging devices recording the patient’s surface from different viewpoints to achieve the required stability and accuracy. Since range imaging devices come as add-ons to regular linear accelerators, they have to share the limited space with other sensors in the treatment room to get a line of sight to the patient’s isocenter. The objective of this work is to describe a new registration framework which enables stable tracking using only one range imager. We unveil the design of our solution to the problem of tracking a patient over long trajectories and large viewpoint changes including surface acquisition, pose estimation and simultaneous surface reconstruction. We evaluate the performance of the system using three clinically motivated experiments: (i) motion management, (ii) non-coplanar patient setup and (iii) tracking over very large angles. We compare our framework to the state-of-art ICP algorithm and to a ground-truth stereoscopic X-ray system from BrainLab. Results demonstrate that we could track subtle movements up to 2.5 cm with a mean target registration error of 0.44 mm and 0.02°. Subsequent non-coplanar field setup on 30° and 2 cm motion yielded a target registration error of 2.88 mm which is within clinical tolerances. Our sensor design demonstrates the potential of simultaneous reconstruction and tracking algorithms and its use for patient setup and motion management in radiation therapy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bauer, S., et al.: Multi-modal surface registration for markerless initial patient setup in radiation therapy using microsoft’s Kinect sensor. In: 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops). IEEE (2011)
Bert, C., et al.: Clinical experience with a 3D surface patient setup system for alignment of partial-breast irradiation patients. Int. J. Radiat. Oncol. Biol. Phys. 64(4), 1265–1274 (2006)
Schaller, C., et al.: Time-of-flight sensor for patient positioning. In: SPIE Medical Imaging, pp. 726110–726110 (2009)
Placht, S., et al.: Fast time-of-flight camera based surface registration for radiotherapy patient positioning. Med. Phys. 39(1), 4–17 (2012)
Brahme, A., et al.: 4D laser camera for accurate patient positioning, collision avoidance, image fusion and adaptive approaches during diagnostic and therapeutic procedures. Med. Phys. 35(5), 1670–1681 (2008)
Lindl, B.L., et al.: TOPOS: A new topometric patient positioning and tracking system for radiation therapy based on structured white light. Med. Phys. 40(4), 042701 (2013)
Peng, J.L., et al.: Characterization of a real-time surface image-guided stereotactic positioning system. Med. Phys. 37(10), 5421–5433 (2010)
Stieler, F., et al.: A novel surface imaging system for patient positioning and surveillance during radiotherapy: a phantom study and clinical evaluation. Strahlenther. Onkol. 189(11), 938–944 (2013)
Humenberger, M., et al.: A fast stereo matching algorithm suitable for embedded real-time systems. Comput. Vis. Image Underst. 114(11), 1180–1202 (2010)
Segal, A., et al.: Generalized-ICP. Robot. Sci. Syst. 2(4) (2009)
Gelfand, N., et al.: Geometrically stable sampling for the ICP algorithm. In: Proceedings of the Fourth International Conference on 3-D Digital Imaging and Modeling. IEEE (2003)
Newcombe, R.A., et al.: KinectFusion: real-time dense surface mapping and tracking. In: Mixed and Augmented Reality (ISMAR), pp. 127–136 (2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Kaiser, H., Fallavollita, P., Navab, N. (2015). ‘On the Fly’ Reconstruction and Tracking System for Patient Setup in Radiation Therapy. In: Linte, C., Yaniv, Z., Fallavollita, P. (eds) Augmented Environments for Computer-Assisted Interventions. AE-CAI 2015. Lecture Notes in Computer Science(), vol 9365. Springer, Cham. https://doi.org/10.1007/978-3-319-24601-7_4
Download citation
DOI: https://doi.org/10.1007/978-3-319-24601-7_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-24600-0
Online ISBN: 978-3-319-24601-7
eBook Packages: Computer ScienceComputer Science (R0)