Skip to Main content Skip to Navigation

Video analysis for augmented cataract surgery

Abstract : The digital era is increasingly changing the world due to the sheer volume of data produced every day. The medical domain is highly affected by this revolution, because analysing this data can be a source of education/support for the clinicians. In this thesis, we propose to reuse the surgery videos recorded in the operating rooms for computer-assisted surgery system. We are chiefly interested in recognizing the surgical gesture being performed at each instant in order to provide relevant information. To achieve this goal, this thesis addresses the surgical tool recognition problem, with applications in cataract surgery. The main objective of this thesis is to address the surgical tool recognition problem in cataract surgery videos.In the surgical field, those tools are partially visible in videos and highly similar to one another. To address the visual challenges in the cataract surgical field, we propose to add an additional camera filming the surgical tray. Our goal is to detect the tool presence in the two complementary types of videos: tool-tissue interaction and surgical tray videos. The former records the patient's eye and the latter records the surgical tray activities.Two tasks are proposed to perform the task on the surgical tray videos: tools change detection and tool presence detection.First, we establish a similar pipeline for both tasks. It is based on standard classification methods on top of visual learning features. It yields satisfactory results for the tools change task, howev-lateer, it badly performs the surgical tool presence task on the tray. Second, we design deep learning architectures for the surgical tool detection on both video types in order to address the difficulties in manually designing the visual features.To alleviate the inherent challenges on the surgical tray videos, we propose to generate simulated surgical tray scenes along with a patch-based convolutional neural network (CNN).Ultimately, we study the temporal information using RNN processing the CNN results. Contrary to our primary hypothesis, the experimental results show deficient results for surgical tool presence on the tray but very good results on the tool-tissue interaction videos. We achieve even better results in the surgical field after fusing the tool change information coming from the tray and tool presence signals on the tool-tissue interaction videos.
Document type :
Complete list of metadatas

Cited literature [248 references]  Display  Hide  Download
Contributor : Abes Star :  Contact
Submitted on : Wednesday, October 10, 2018 - 11:39:06 AM
Last modification on : Friday, October 23, 2020 - 4:59:18 PM
Long-term archiving on: : Friday, January 11, 2019 - 1:40:44 PM


Version validated by the jury (STAR)


  • HAL Id : tel-01892032, version 1


Hassan Al Hajj. Video analysis for augmented cataract surgery. Human health and pathology. Université de Bretagne occidentale - Brest, 2018. English. ⟨NNT : 2018BRES0041⟩. ⟨tel-01892032⟩



Record views


Files downloads