A polynomial model of surgical gestures for real-time retrieval of surgery videos

  • Authors:
  • Gwénolé Quellec;Mathieu Lamard;Zakarya Droueche;Béatrice Cochener;Christian Roux;Guy Cazuguel

  • Affiliations:
  • Inserm, UMR 1101, Brest, France;Inserm, UMR 1101, Brest, France, Univ Bretagne Occidentale, Brest, France;Inserm, UMR 1101, Brest, France, Dpt. ITI, INSTITUT. TELECOM, TELECOM Bretagne, UEB, Brest, France;Inserm, UMR 1101, Brest, France, Univ Bretagne Occidentale, Brest, France, CHU Brest, Service d'Ophtalmologie, Brest, France;Inserm, UMR 1101, Brest, France, Dpt. ITI, INSTITUT. TELECOM, TELECOM Bretagne, UEB, Brest, France;Inserm, UMR 1101, Brest, France, Dpt. ITI, INSTITUT. TELECOM, TELECOM Bretagne, UEB, Brest, France

  • Venue:
  • MCBR-CDS'12 Proceedings of the Third MICCAI international conference on Medical Content-Based Retrieval for Clinical Decision Support
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper introduces a novel retrieval framework for surgery videos. Given a query video, the goal is to retrieve videos in which similar surgical gestures appear. In this framework, the motion content of short video subsequences is modeled, in real-time, using spatiotemporal polynomials. The retrieval engine needs to be trained: key spatiotemporal polynomials, characterizing semantically-relevant surgical gestures, are identified through multiple-instance learning. Then, videos are compared in a high-level space spanned by these key spatiotemporal polynomials. The framework was applied to a dataset of 900 manually-delimited clips from 100 cataract surgery videos. High classification performance (Az=0.816±0.118) and retrieval performance (MAP=0.358) were observed.