2012
DOI: 10.1007/978-3-642-33415-3_5
|View full text |Cite
|
Sign up to set email alerts
|

Surgical Gesture Classification from Video Data

Abstract: Abstract. Much of the existing work on automatic classification of gestures and skill in robotic surgery is based on kinematic and dynamic cues, such as time to completion, speed, forces, torque, or robot trajectories. In this paper we show that in a typical surgical training setup, video data can be equally discriminative. To that end, we propose and evaluate three approaches to surgical gesture classification from video. In the first one, we model each video clip from each surgical gesture as the output of a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(3 citation statements)
references
References 22 publications
(26 reference statements)
0
1
0
Order By: Relevance
“…This led to further research in improving temporal learning in MIS with the addition of long short-term memory (LSTM) neural networks, which allows for more efficiency in identifying phases of surgery and tracking surgical instruments. Combining a CNN and LSTM gives the advantage of more efficiently identifying phases of surgery and tracking surgical instruments [30,31]. The EndoNet researchers published follow-up research comparing a CNN with HMM versus a CNN with LTSM [31].…”
Section: Figurementioning
confidence: 99%
“…This led to further research in improving temporal learning in MIS with the addition of long short-term memory (LSTM) neural networks, which allows for more efficiency in identifying phases of surgery and tracking surgical instruments. Combining a CNN and LSTM gives the advantage of more efficiently identifying phases of surgery and tracking surgical instruments [30,31]. The EndoNet researchers published follow-up research comparing a CNN with HMM versus a CNN with LTSM [31].…”
Section: Figurementioning
confidence: 99%
“…For example, researchers have used automated performance metrics to predict a patient's postoperative length of stay within a hospital 26 . Another line of research has instead focused on exclusively exploiting live surgical videos from endoscopic cameras to classify surgical activity 4,29 , gestures 5,[30][31][32][33] and skills 6,7,13,34,35 , among other tasks 36,37 . For information on additional studies, we refer readers to a recent review 9 .…”
Section: Previous Workmentioning
confidence: 99%
“…At the latter level, which is typically concerned with robotassisted surgery or training and assessment of surgeons, we see research on phase detection (Stauder, 2014) and detailed models of individual tool usage patterns based on sensor data (Ahmadi, 2009). Individual hand motions from video data are automatically identified in (Lin, 2006) and (Haro, 2012). A number of models based on sensor data collected during Cholecystectomies (a highly standardized procedure), were developed in (Blum, 2008), (Bouarfa and Dankelman, 2012), (Bouarfa, 2011), and (Neumuth, 2011).…”
Section: Related Workmentioning
confidence: 99%