Attention: Restrictions on use of AUA, AUAER, and UCF content in third party applications, including artificial intelligence technologies, such as large language models and generative AI.
You are prohibited from using or uploading content you accessed through this website into external applications, bots, software, or websites, including those using artificial intelligence technologies and infrastructure, including deep learning, machine learning and large language models and generative AI.

ARTIFICIAL INTELLIGENCE Artificial Intelligence–Driven Surgical Phase Recognition for Robotic Urologic Surgical Skill Assessment

By: Pieter De Backer, MD, PhD, MScEng, ORSI Academy, Melle, Belgium, Ghent University, Belgium, Onze-Lieve-Vrouwziekenhuis Hospital, Aalst, Belgium; Marco Mezzina, MScEng, ORSI Academy, Melle, Belgium; Anthony G. Gallagher, PhD, DSc, MAE, ORSI Academy, Melle, Belgium, School of Medicine, Ulster University, United Kingdom, Katholieke Universiteit Leuven, Belgium; Alexandre Mottrie, MD, PhD, ORSI Academy, Melle, Belgium, Onze-Lieve-Vrouwziekenhuis Hospital, Aalst, Belgium | Posted on: 21 Feb 2024

In the past 2 decades, the use of robotic surgery has significantly increased given its minimally invasive nature, offering benefits like enhanced 3D vision, improved precision through the articulated instrument tips, ergonomic advantages for surgeons, and shorter learning curves when compared to classical laparoscopy. The importance of surgical skill in influencing patient outcomes1 highlights the need for validated training curricula in robotic surgery. However, there is currently no universally accepted standard for structured training and assessment. Various manual and automated methods for skill assessment have been proposed.2

Manual qualitative assessment approaches involve global scores using Likert scales, exemplified by the Global Evaluative Assessment of Robotic Skills or Objective Structured Assessment of Technical Skills,3 and have widespread acceptance in the surgical community, although error-based approaches deploying binary scored metrics, such as proficiency-based progression (PBP),4 have proven superior to traditional training for several tasks. For robotic surgery, PBP has recently been demonstrated to have significantly better scoring reliability, sensitivity, and specificity when compared to the Global Evaluative Assessment of Robotic Skills.5 The time-intensive nature of manual expert assessments has led to increased interest in automated approaches using artificial intelligence (AI). Up to present, these methods often rely on kinematic data extracted from virtual reality simulators or robotic systems. Other recent evidence on AI robotic surgical skill assessment suggests that further AI research should most probably be focused on validating and implementing existing tools rather than developing completely new AI-driven assessment methods.6

One specific enabler for automated PBP scoring is the automated indexation of procedure-specific surgical phases and steps as defined in PBP curricula.7 Figure 1 shows the increased complexity in surgical steps and errors when comparing the scoring of a surgical suturing and knot-tying task to a fully characterized procedure such as radical prostatectomy or partial nephrectomy.

image

Figure 1. Proficiency-based progression surgical skill assessment granularity. Shifting from basic robotic surgery skills towards real-life surgery expands the number of steps and errors significantly.7,11,12

Consequently, automated surgical phase or step detection represents an enabling step towards realizing automated PBP surgical skill assessment. From a computational point of view, automated PBP skill assessment systems should most probably first identify the current phase and subsequent step before further specifying the errors intrinsic to that particular step.

image

Figure 2. Automated surgical phase assessment for 2 different robot-assisted radical prostatectomies (Procedure A and Procedure B) comparing the perfect solution (“Manual labels”) to the AI-predicted labels (“Predictions”). Each color block depicts a different surgical phase, which can in turn be further divided into steps.

Apart from PBP, automated surgical phase indexation enables consistent and standardized categorization across procedures, facilitating the processes of video review, analysis, and sharing. Advancements in video analysis, computer vision, AI, and data analytics, sometimes collectively referred to as “surgical data science,” are slowly integrating into surgical practice. With it, the significance of accurately and uniformly defining surgical phases equally increases.8 Previous research has highlighted the lack of uniform definitions for phase delineations.9 PBP can help in acquiring clinically relevant surgical phase definitions; surgical phase durations as defined from PBP curricula have recently shown to correlate to certain patient specific parameters for robot-assisted partial nephrectomy.10

Figure 2 provides an example of automated surgical phase recognition for robot-assisted radical prostatectomies. It depicts the manually annotated, ground truth surgical phases on the top line as reviewed by a urologist, compared to the AI-assisted prediction for 2 unseen procedures. The AI system (more precisely, a deep learning network) was trained on 50 robot-assisted radical prostatectomies, labeled according to PBP phases. The system was able to predict surgical phases in unseen prostatectomies with an 89.07% sensitivity, 90% accuracy, 90.4% precision, and an 89.32% F1-score, representing the number of video frames that were automatically assigned to the correct phase.

In conclusion, present evidence indicates significant potential for using binary-scored, error-based metrics such as PBP as objective and validated skill assessment tools. The time-intensive nature of these manual assessment methods is expected to be boosted through the use of AI video analysis, while PBP phase assessment in turn is also expected to generate more homogenous and evidence-based definitions for surgical phase definition for the whole surgical AI field.

  1. Curtis NJ, Foster JD, Miskovic D, et al. Association of surgical skill assessment with clinical outcomes in cancer surgery. JAMA Surg. 2020;155(7):590-598.
  2. Chen J, Cheng N, Cacciamani G, et al. Objective assessment of robotic surgical technical skill: a systematic review. J Urol. 2019;201(3):461-469.
  3. Goh AC, Goldfarb DW, Sander JC, Miles BJ, Dunkin BJ. Global evaluative assessment of robotic skills: validation of a clinical assessment tool to measure robotic surgical skills. J Urol. 2012;187(1):247-252.
  4. Gallagher AG, Ritter EM, Champion H, et al. Virtual reality simulation for the operating room: proficiency-based training as a paradigm shift in surgical skills training. Ann Surg. 2005;241(2):364-372.
  5. De Groote R, Puliatti S, Amato M, et al. Discrimination, reliability, sensitivity, and specificity of robotic surgical proficiency assessment with global evaluative assessment of robotic skills and binary scoring metrics: results from a randomized controlled trial. Ann Surg Open. 2023;4(3):e307.
  6. Boal MWE, Anastasiou D, Tesfai F, et al. Evaluation of objective tools and artificial intelligence in robotic surgery technical skills assessment: a systematic review. Br J Surg. 2024;111(1):znad331.
  7. Mottrie A, Mazzone E, Wiklund P, et al. Objective assessment of intraoperative skills for robot-assisted radical prostatectomy (RARP): results from the ERUS Scientific and Educational Working Groups Metrics Initiative. BJU Int. 2021;128(1):103-111.
  8. Eckhoff JA, Rosman G, Altieri MS, et al. SAGES consensus recommendations on surgical video data use, structure, and exploration (for research in artificial intelligence, clinical quality improvement, and surgical education). Surg Endosc. 2023;37(11):8690-8707.
  9. Garrow CR, Kowalewski KF, Li L, et al. Machine learning for surgical phase recognition. Ann Surg. 2021;273(4):684-693.
  10. De Backer P, Peraire Lores M, Demuynck M, et al. Surgical phase duration in robot-assisted partial nephrectomy: a surgical data science exploration for clinical relevance. Diagnostics (Basel). 2023;13(21):3386.
  11. Puliatti S, Mazzone E, Amato M, De Groote R, Mottrie A, Gallagher AG. Development and validation of the objective assessment of robotic suturing and knot tying skills for chicken anastomotic model. Surg Endosc. 2021;35(8):4285-4294.
  12. Farinha R, Breda A, Porter J, et al. International expert consensus on metric-based characterization of robot-assisted partial nephrectomy. Eur Urol Focus. 2023;9(2):388-395.

advertisement

advertisement