3260 papers • 126 benchmarks • 313 datasets
Presence detection of various classes of surgical instruments in endoscopy videos.
(Image credit: Papersgraph)
These leaderboards are used to track progress in surgical-tool-detection
Use these libraries to find surgical-tool-detection models and implementations
No subtasks available.
The approach, named SSD, discretizes the output space of bounding boxes into a set of default boxes over different aspect ratios and scales per feature map location, which makes SSD easy to train and straightforward to integrate into systems that require a detection component.
This paper proposes a novel method for phase recognition that uses a convolutional neural network (CNN) to automatically learn features from cholecystectomy videos and that relies uniquely on visual information.
A deep architecture, trained solely on image level annotations, that can be used for both tool presence detection and localization in surgical videos, and relies on a fully convolutional neural network, trained end-to-end.
It is demonstrated that binary presence labels are sufficient for training a deep learning tracking model using the proposed method, and it is shown that the ConvLSTM can leverage the spatiotemporal coherence of consecutive image frames across a surgical video to improve tool presence detection, spatial localization, and motion tracking.
A novel method by developing a multi-task recurrent convolutional network with correlation loss (MTRCNet-CL) to exploit their relatedness to simultaneously boost the performance of both surgical tool presence detection and phase recognition.
Correct transfer of these methods to surgery leads to substantial performance gains over generic uses of SSL - up to 7.4% on phase recognition and 20% on tool presence detection - as well as state-of-the-art semi-supervised phase recognition approaches by up to 14%.
The accuracy of surgical instrument segmentation is improved, surpassing most methods of instance segmentation via weakly supervised bounding boxes and when applied to the public HOSPI-Tools dataset.
This paper introduces an SSL framework in the surgical tool detection paradigm, which aims to mitigate training data scarcity and data imbalance problems through a knowledge distillation approach.
The CholecTriplet2022 challenge is presented, which extends surgical action triplet modeling from recognition to detection, and includes weakly-supervised bounding box localization of every visible surgical instrument (or tool), as the key actors, and the modeling of each tool-activity in the form of ‹instrument, verb, target› triplet.
EgoSurgery-Tool is introduced, an extension of the existing EgoSurgery-Phase dataset, which contains real open surgery videos captured using an egocentric camera attached to the surgeon's head, along with phase annotations, and is superior to existing datasets due to its larger scale, greater variety of surgical tools, more annotations, and denser scenes.
Adding a benchmark result helps the community track progress.