Within the BiOPAss (Bild, Ontologie und Prozessgestützte Assistenz für die minimal-invasive endoskopische Chirurgie) project, a novel marker-free image-based navigation approach will be developed, which would support the endoscopist in identifying the anatomical position of the endoscope within the lumina of the human body. Together with clinical domain experts, research groups at Leizpig University and TU Munich, and companies; concepts, data structures, and efficient algorithms will be developed that allow a match between image features of newly acquired with previously stored endoscopic image sequences. In combination with process information and dedicated ontologies, positional hints will be visualized graphically within an anatomical reference model to provide a navigational aid to the surgeon.

Introduction

Currently, endoscopy is the most widely used technique for diagnosis and treatment of gastro-intestinal tract and sinusitis related diseases. However, surgeon face challenges in endoscope navigation during the procedure due to limited field of view (FOV) of the operating area, rotating images and impaired hand-eye coordination. Experienced endoscopic surgeons are able to mentally assign an anatomical position of the endoscope to the endoscopic image, however, novice surgeons often face challenges in determining the anatomical position and orientation of the endoscope from the endoscopic images displayed on the monitor.

Endonasal endoscopic surgery

 Fig. 1 Endonasal endoscopic surgery

 

To overcome these issue, optical tracking (OPT) or electro-magniectic tracking (EMT) devices have been introduced to the operating room (OR), however they either demenad direct line of sight requirements or susecptible to errors due to ferro-magnetic material apart from increased technical complexity in the usually crowded OR.

 Endoscope tracking using OT

Fig. 2 Endoscope tracking using OT (in red)

With the aim of reducing technical complexity of the OR, we support the endoscopist in identifying the anatomical position of the endoscope within the lumina of the human body by annotating real-time image frames with a label in the training database which is having similar appearance and annotated by the expert surgeons previously.

Real-time query frame is mapped to the corresponding label in

Fig. 3 Real-time query frame is mapped to the corresponding label in the database, and visualized w.r.t CT image of the landmark.

 

Scene Recognition Pipeline

To achieve the goal, we select a limited set of representative images from expert annotated endoscopic videos; extract informative image features; and create a classification model by learning these image features, which in turn generate corresponding annotation for real-time query image frames.

Scene Recognition Pipeline

Fig. 4 Training and testing stages of the scene recognition pipeline

 

Visualization

The operator is provided with a roadmap, consisting of images of the nasal divisions along the route. When the operator approaches a division in the airways, the system shows which branch to follow.

Current position and navigation graphical hints are overlaid on the endoscopic video and visualized within an anatomical reference model and CT.

Fig. 4 Current position and navigation graphical hints are overlaid on the endoscopic video and visualized within an anatomical reference model and CT.

Publications

2022
Vision-based Context-awareness in Minimally Invasive Surgical Video Streams Doctoral thesis, Technische Universität Berlin, Stefan Zachow, Anirban Mukhopadhyay (Advisors), 2022 Manish Sahu BibTeX
BiOPAss
2019
CATARACTS: Challenge on Automatic Tool Annotation for cataRACT Surgery Medical Image Analysis, 52(2), pp. 24-41, 2019 Hassan Al Hajj, Manish Sahu, Mathieu Lamard, Pierre-Henri Conze, Soumali Roychowdhury, Xiaowei Hu, Gabija Marsalkaite, Odysseas Zisimopoulos, Muneer Ahmad Dedmari, Fenqiang Zhao, Jonas Prellberg, Adrian Galdran, Teresa Araujo, Duc My Vo, Chandan Panda, Navdeep Dahiya, Satoshi Kondo, Zhengbing Bian, Jonas Bialopetravicius, Chenghui Qiu, Sabrina Dill, Anirban Mukhopadyay, Pedro Costa, Guilherme Aresta, Senthil Ramamurthy, Sang-Woong Lee, Aurelio Campilho, Stefan Zachow, Shunren Xia, Sailesh Conjeti, Jogundas Armaitis, Pheng-Ann Heng, Arash Vahdat, Beatrice Cochener, Gwenole Quellec BibTeX
DOI
BiOPAss
2018
Joint Feature Learning and Classification - Deep Learning for Surgical Phase Detection Master's thesis, Technische Universität Berlin, Manish Sahu (Advisor), 2018 Sabrina Dill BibTeX
URN
BiOPAss
Joint Feature Learning and Classification - Deep Learning for Surgical Phase Detection Master's thesis, Technische Universität Berlin, Manish Sahu, Stefan Zachow (Advisors), 2018 Sabrina Patricia Dill BibTeX
BiOPAss
2017
Addressing multi-label imbalance problem of Surgical Tool Detection using CNN International Journal of Computer Assisted Radiology and Surgery, 12(6), pp. 1013-1020, 2017 Manish Sahu, Anirban Mukhopadhyay, Angelika Szengel, Stefan Zachow PDF
BibTeX
DOI
BiOPAss
Surgical Tool Presence Detection for Cataract Procedures ZIB-Report 18-28 Manish Sahu, Sabrina Dill, Anirban Mukhopadyay, Stefan Zachow PDF
BibTeX
URN
BiOPAss
2016
Instrument State Recognition and Tracking for Effective Control of Robotized Laparoscopic Systems International Journal of Mechanical Engineering and Robotics Research, 5(1), pp. 33-38, 2016 Manish Sahu, Daniil Moerman, Philip Mewes, Peter Mountney, Georg Rose BibTeX
DOI
BiOPAss
2015
A Multimodal Nonverbal Human-robot Communication System VI International Conference on Computational Bioengineering, 2015 Salah Saleh, Manish Sahu, Zuhair Zafar, Karsten Berns BibTeX
BiOPAss