Surgical phase recognition
27 papers with code • 4 benchmarks • 5 datasets
The first 40 videos are used for training, the last 40 videos are used for testing.
Most implemented papers
Pixel-Wise Recognition for Holistic Surgical Scene Understanding
To exploit our proposed benchmark, we introduce the Transformers for Actions, Phases, Steps, and Instrument Segmentation (TAPIS) model, a general architecture that combines a global video feature extractor with localized region proposals from an instrument segmentation model to tackle the multi-granularity of our benchmark.
TeCNO: Surgical Phase Recognition with Multi-Stage Temporal Convolutional Networks
Automatic surgical phase recognition is a challenging and crucial task with the potential to improve patient safety and become an integral part of intra-operative decision-support systems.
Not End-to-End: Explore Multi-Stage Architecture for Online Surgical Phase Recognition
To address the problem, we propose a new non end-to-end training strategy and explore different designs of multi-stage architecture for surgical phase recognition task.
Temporal coherence-based self-supervised learning for laparoscopic workflow analysis
To achieve this, methods for surgical workflow analysis are crucial.
Learning from a tiny dataset of manual annotations: a teacher/student approach for surgical phase recognition
Vision algorithms capable of interpreting scenes from a real-time video stream are necessary for computer-assisted surgery systems to achieve context-aware behavior.
Multi-Task Recurrent Convolutional Network with Correlation Loss for Surgical Video Analysis
Mutually leveraging both low-level feature sharing and high-level prediction correlating, our MTRCNet-CL method can encourage the interactions between the two tasks to a large extent, and hence can bring about benefits to each other.
Trans-SVNet: Accurate Phase Recognition from Surgical Videos via Hybrid Embedding Aggregation Transformer
In this paper, we introduce, for the first time in surgical workflow analysis, Transformer to reconsider the ignored complementary effects of spatial and temporal features for accurate surgical phase recognition.
LensID: A CNN-RNN-Based Framework Towards Lens Irregularity Detection in Cataract Surgery Videos
In particular, we propose (I) an end-to-end recurrent neural network to recognize the lens-implantation phase and (II) a novel semantic segmentation network to segment the lens and pupil after the implantation phase.
Exploring Segment-level Semantics for Online Phase Recognition from Surgical Videos
Automatic surgical phase recognition plays a vital role in robot-assisted surgeries.
Less is More: Surgical Phase Recognition from Timestamp Supervision
Our study uncovers unique insights of surgical phase recognition with timestamp supervisions: 1) timestamp annotation can reduce 74% annotation time compared with the full annotation, and surgeons tend to annotate those timestamps near the middle of phases; 2) extensive experiments demonstrate that our method can achieve competitive results compared with full supervision methods, while reducing manual annotation cost; 3) less is more in surgical phase recognition, i. e., less but discriminative pseudo labels outperform full but containing ambiguous frames; 4) the proposed UATD can be used as a plug and play method to clean ambiguous labels near boundaries between phases, and improve the performance of the current surgical phase recognition methods.