TY - JOUR
T1 - Hand motion-aware surgical tool localization and classification from an egocentric camera
AU - Shimizu, Tomohiro
AU - Hachiuma, Ryo
AU - Kajita, Hiroki
AU - Takatsume, Yoshifumi
AU - Saito, Hideo
N1 - Funding Information:
Funding: This research was funded by JST-Mirai Program Grant Number JPMJMI19B2, ROIS NII Open Collaborative Research 2020-20S0404, the MIC/SCOPE #201603003, and MHLW Program Grant Number JPMH20AC1004.
Publisher Copyright:
© 2021 by the authors. Licensee MDPI, Basel, Switzerland.
PY - 2021/2
Y1 - 2021/2
N2 - Detecting surgical tools is an essential task for the analysis and evaluation of surgical videos. However, in open surgery such as plastic surgery, it is difficult to detect them because there are surgical tools with similar shapes, such as scissors and needle holders. Unlike endoscopic surgery, the tips of the tools are often hidden in the operating field and are not captured clearly due to low camera resolution, whereas the movements of the tools and hands can be captured. As a result that the different uses of each tool require different hand movements, it is possible to use hand movement data to classify the two types of tools. We combined three modules for localization, selection, and classification, for the detection of the two tools. In the localization module, we employed the Faster R-CNN to detect surgical tools and target hands, and in the classification module, we extracted hand movement information by combining ResNet-18 and LSTM to classify two tools. We created a dataset in which seven different types of open surgery were recorded, and we provided the annotation of surgical tool detection. Our experiments show that our approach successfully detected the two different tools and outperformed the two baseline methods.
AB - Detecting surgical tools is an essential task for the analysis and evaluation of surgical videos. However, in open surgery such as plastic surgery, it is difficult to detect them because there are surgical tools with similar shapes, such as scissors and needle holders. Unlike endoscopic surgery, the tips of the tools are often hidden in the operating field and are not captured clearly due to low camera resolution, whereas the movements of the tools and hands can be captured. As a result that the different uses of each tool require different hand movements, it is possible to use hand movement data to classify the two types of tools. We combined three modules for localization, selection, and classification, for the detection of the two tools. In the localization module, we employed the Faster R-CNN to detect surgical tools and target hands, and in the classification module, we extracted hand movement information by combining ResNet-18 and LSTM to classify two tools. We created a dataset in which seven different types of open surgery were recorded, and we provided the annotation of surgical tool detection. Our experiments show that our approach successfully detected the two different tools and outperformed the two baseline methods.
KW - Egocentric camera
KW - Object detection
KW - Open surgery
KW - Surgical tools
UR - http://www.scopus.com/inward/record.url?scp=85106954739&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85106954739&partnerID=8YFLogxK
U2 - 10.3390/jimaging7020015
DO - 10.3390/jimaging7020015
M3 - Article
AN - SCOPUS:85106954739
SN - 2313-433X
VL - 7
JO - Journal of Imaging
JF - Journal of Imaging
IS - 2
M1 - 15
ER -