Abstract
Purpose
Surgical processes are complex entities characterized by expressive models and data. Recognizable activities define each surgical process. The principal limitation of current vision-based recognition methods is inefficiency due to the large amount of information captured during a surgical procedure. To overcome this technical challenge, we introduce a surgical gesture recognition system using temperature-based recognition.
Methods
An infrared thermal camera was combined with a hierarchical temporal memory and was used during surgical procedures. The recordings were analyzed for recognition of surgical activities. The image sequence information acquired included hand temperatures. This datum was analyzed to perform gesture extraction and recognition based on heat differences between the surgeon’s warm hands and the colder background of the environment.
Results
The system was validated by simulating a functional endoscopic sinus surgery, a common type of otolaryngologic surgery. The thermal camera was directed toward the hands of the surgeon while handling different instruments. The system achieved an online recognition accuracy of 96 % with high precision and recall rates of approximately 60 %.
Conclusion
Vision-based recognition methods are the current best practice approaches for monitoring surgical processes. Problems of information overflow and extended recognition times in vision-based approaches were overcome by changing the spectral range to infrared. This change enables the real-time recognition of surgical activities and provides online monitoring information to surgical assistance systems and workflow management systems.
Similar content being viewed by others
References
Archer T, Macario A (2006) The drive for operating room efficiency will increase quality of patient care. Curr Opin Anaesthesiol 19:171–176. doi:10.1097/01.aco.0000192796.02797.82
Sutherland J, van den Heuvel W-J (2006) Towards an intelligent hospital environment: adaptive workflow in the OR of the future. Proceedings of the 39th annual Hawaii international conference on system sciences, volume 05. IEEE Computer Society, Washington, DC, USA, p 100b
Cleary K, Kinsella A, Mun SK (2005) OR 2020 workshop report: operating room of the future. Int Congr Ser 1281:832–838. doi:10.1016/j.ics.2005.03.279
Neumuth T, Jannin P, Schlomberg J et al (2011) Analysis of surgical intervention populations using generic surgical process models. Int J Comput Assist Radiol Surg 6:59–71. doi:10.1007/s11548-010-0475-y
Neumuth T (2013) Surgical process modeling: theory, methods, and applications
Neumuth T, Liebmann P, Wiedemann P, Meixensberger J (2012) Surgical workflow management schemata for cataract procedures. Process model-based design and validation of workflow schemata. Methods Inf Med 51:371–382. doi:10.3414/ME11-01-0093
Franke S, Meixensberger J, Neumuth T (2013) Intervention time prediction from surgical low-level tasks. J Biomed Inform 46:152–159. doi:10.1016/j.jbi.2012.10.002
Tiwari V, Dexter F, Rothman BS et al (2013) Explanation for the near-constant mean time remaining in surgical cases exceeding their estimated duration, necessary for appropriate display on electronic white boards. Anesth Analg 117:487–493. doi:10.1213/ANE.0b013e31829772e9
Epstein RH, Dexter F (2012) Mediated interruptions of anaesthesia providers using predictions of workload from anaesthesia information management system data. Anaesth Intensive Care 40:803–812
Lin HC, Shafran I, Yuh D, Hager GD (2006) Towards automatic skill evaluation: detection and segmentation of robot-assisted surgical motions. Comput Aided Surg 11:220–230. doi:10.3109/10929080600989189
Judkins TN, Oleynikov D, Stergiou N (2009) Objective evaluation of expert and novice performance during robotic surgical training tasks. Surg Endosc 23:590–597. doi:10.1007/s00464-008-9933-9
Reiley CE, Lin HC, Varadarajan B et al (2008) Automatic recognition of surgical motions using statistical modeling for capturing variability. Stud Health Technol Inform 132:396
Vankipuram M, Kahol K, Cohen T, Patel VL (2009) Visualization and analysis of activities in critical care environments. AMIA Annu Symp Proc 2009:662–666
Bouarfa L, Jonker PP, Dankelman J (2009) Surgical context discovery by monitoring low-level activities in the OR
Neumuth T, Meissner C (2012) Online recognition of surgical instruments by information fusion. Int J Comput Assist Radiol Surg 7:297–304. doi:10.1007/s11548-011-0662-5
Meißner C, Neumuth T (2012) RFID-based surgical instrument detection using Hidden Markov models. Biomed Tech (Berl). doi:10.1515/bmt-2012-4047
Kranzfelder M, Zywitza D, Jell T et al (2012) Real-time monitoring for detection of retained surgical sponges and team motion in the surgical operation room using radio-frequency-identification (RFID) technology: a preclinical evaluation. J Surg Res 175:191–198. doi:10.1016/j.jss.2011.03.029
Ahmadi S-A, Padoy N, Heining SM et al (2008) Introducing wearable accelerometers in the surgery room for activity detection. 7. Jahrestagung der Deutschen Gesellschaft fuer Computer-und Roboter-Assistierte Chirurgie (CURAC 2008)
Lester J, Choudhury T, Kern N et al (2005) A hybrid discriminative/generative approach for modeling human activities. In: Proceedings of the international joint conference on artificial intelligence (IJCAI), pp 766–772
Klank U, Padoy N, Feussner H, Navab N (2008) Automatic feature generation in endoscopic images. Int J CARS 3:331–339. doi:10.1007/s11548-008-0223-8
Blum T, Feußner H, Navab N (2010) Modeling and segmentation of surgical workflow from laparoscopic video. In: Jiang T, Navab N, Pluim J, Viergever M (eds) Medical image computing and computer-assisted intervention—MICCAI 2010. Springer, Berlin, pp 400–407
Lalys F, Riffaud L, Bouget D, Jannin P (2011) An application-dependent framework for the recognition of high-level surgical tasks in the OR. In: Fichtinger G, Martel A, Peters T (eds) Medical image computing and computer-assisted intervention—MICCAI 2011. Springer, Berlin, pp 331–338
Lalys F, Bouget D, Riffaud L, Jannin P (2013) Automatic knowledge-based recognition of low-level tasks in ophthalmological procedures. Int J Comput Assist Radiol Surg 8:39–49. doi:10.1007/s11548-012-0685-6
Haro BB, Zappella L, Vidal R (2012) Surgical gesture classification from video data. In: Ayache N, Delingette H, Golland P, Mori K (eds) Medical image computing and computer-assisted intervention—MICCAI 2012. Springer, Berlin, pp 34–41
Appenrodt J, Al-Hamadi A, Michaelis B (2010) Data gathering for gesture recognition systems based on single color-, stereo color-and thermal cameras. Int J Signal Process Image Process Pattern Recognit 3:37–50
Hawkins J, George D (2006) Hierarchical temporal memory: concepts, theory, and terminology
Hawkins J, Blakeslee S (2007) On intelligence. Macmillan, New York
Kapuscinski T (2010) Using hierarchical temporal memory for vision-based hand shape recognition under large variations in hand’s rotation. In: Rutkowski L, Scherer R, Tadeusiewicz R et al (eds) Artificial intelligence and soft computing. Springer, Berlin, pp 272–279
Kapuscinski T, Wysocki M (2009) Using hierarchical temporal memory for recognition of signed polish words. In: Kurzynski M, Wozniak M (eds) Computer recognition systems 3. Springer, Berlin, pp 355–362
Bengio Y (2009) Learning deep architectures for AI. Found Trends Mach Learn 2:1–127. doi:10.1561/2200000006
Bouvrie J, Rosasco L, Poggio T (2009) On invariance in hierarchical models. Adv Neural Inf Process Syst 22:162–170
Acknowledgments
ICCAS is funded by the German Federal Ministry of Education and Research (BMBF) in the scope of the Unternehmen Region (Grant Number 03Z1LN12) and by the German Ministry of Economics (BMWi) in the scope of the Zentrales Innovationsprogramm Mittelstand (ZIM) (Grant Number KF2036709FO0).
Conflict of interest
The authors declare that they have no conflict of interest.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Unger, M., Chalopin, C. & Neumuth, T. Vision-based online recognition of surgical activities. Int J CARS 9, 979–986 (2014). https://doi.org/10.1007/s11548-014-0994-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11548-014-0994-z