Last updated: May 10, 2018
Feedback is a crucial component needed for one to improve. However, manually providing feedback on a procedure done by a surgeon is tedious, and requires the expertise of a colleague within the same field. Because of this, there has been a push to automate the process of generating feedback for a surgeon given information about the surgery itself. If the ability to find videos in a database that are similar in actions to the query works at a high resolution, it could be possible to construct novel feedback from any existing commentary of the database videos for the query. Similarly, the skill of surgery video clip can be inferred as well. Such information would decrease the manpower necessary to train a novice surgeon, as well as advance the ability for surgeons to quickly find areas to improve. This project is part of a larger overarching project studying clips of Cataract Surgery.
We wish to create a method that will aid in automating the process in which surgeons can receive critique given videos of their work. Specifically, given a surgery clip query, we aim to develop a pipeline of neural networks to (1) search our database for clips of similar activity. Furthermore, this pipeline will be adapted to (2) query for clips of a similar skill level in a single activity. For the purpose of this project, we will define an activity as a distinct phase in a surgery. In addition to that, we will attempt to develop a ranking system for comparative skill level analysis within the context of this database.
Currently, this will be applied in the context of cataract surgical data. To get an idea of the interclass and intraclass variability of the data, some examples are attached below.
These three images come from the same activity, performed in different surgeries. Our goal is to classify these images to be the same.
The next two images come from different activities, but performed in the same surgery. Our goal is to classify these images to be different.
The figure above describes the current proposed structure of how we will query by video. As can be seen, the three most prominent components of the diagram are the following:
This splits our project into the following steps.
Dependency | Planned Solution | Solved by | Contingency plan |
---|---|---|---|
GPU processing | Obtain access to MARCC cluster under Dr. Hager’s group for both team members | 2/28 (SOLVED) | At least one team member already has access, so if necessary GPU jobs can be submitted on his account exclusively. |
Machine Learning, statistics and linear algebra packages | There is plenty of open source packages available in Python. We are using PyTorch and Numpy, both available on the MARCC cluster. | 2/10 (SOLVED) | N/A |
Annotated Training dataset | Dr. Vedula has provided over 60 videos of entire cataract surgeries, as well as annotations for which frames correspond to which activities, as well as skill levels of the surgery. | 2/21 (SOLVED) | Although we hope to obtain more data, this amount should be adequate for our needs. |
Gianluca’s inclusion to iRB | Dr. Vedula listed required online courses as well as link for inclusion request | 2/15 (SOLVED) | N/A |
Here give list of other project files (e.g., source code) associated with the project. If these are online give a link to an appropriate external repository or to uploaded media files under this name space.2018-02