| Literature DB >> 23286094 |
Raphael Sznitman1, Karim Ali, Rogério Richa, Russell H Taylor, Gregory D Hager, Pascal Fual.
Abstract
In the context of retinal microsurgery, visual tracking of instruments is a key component of robotics assistance. The difficulty of the task and major reason why most existing strategies fail on in-vivo image sequences lies in the fact that complex and severe changes in instrument appearance are challenging to model. This paper introduces a novel approach, that is both data-driven and complementary to existing tracking techniques. In particular, we show how to learn and integrate an accurate detector with a simple gradient-based tracker within a robust pipeline which runs at framerate. In addition, we present a fully annotated dataset of retinal instruments in in-vivo surgeries, which we use to quantitatively validate our approach. We also demonstrate an application of our method in a laparascopy image sequence.Mesh:
Year: 2012 PMID: 23286094 DOI: 10.1007/978-3-642-33418-4_70
Source DB: PubMed Journal: Med Image Comput Comput Assist Interv