Abstract
The surgical operating room (OR) presents many opportunities for automation and optimisation. Videos from various sources in the OR are becoming increasingly available. The medical community seeks to leverage this wealth of data to develop automated methods to advance interventional care, lower costs, and improve overall patient outcomes. Existing datasets from OR room cameras are thus far limited in size or modalities acquired, leaving it unclear which sensor modalities are best suited for tasks such as recognising surgical action from videos. This study demonstrates that the task of surgical workflow classification is highly dependent on the sensor modalities used. We perform a systematic analysis on several commonly available sensor modalities, evaluating two commonly used fusion approaches that can improve classification performance. Our findings are consistent across model architectures as well as separate camera views. The analyses are carried out on a set of multi-view RGB-D video recordings of 16 laparoscopic interventions.
Dokumententyp: | Zeitschriftenartikel |
---|---|
Fakultät: | Medizin |
Themengebiete: | 600 Technik, Medizin, angewandte Wissenschaften > 610 Medizin und Gesundheit |
ISSN: | 2168-1163 |
Sprache: | Englisch |
Dokumenten ID: | 113030 |
Datum der Veröffentlichung auf Open Access LMU: | 02. Apr. 2024, 07:44 |
Letzte Änderungen: | 02. Apr. 2024, 07:44 |