2021 International Symposium on Medical Robotics (ISMR) 2021
DOI: 10.1109/ismr48346.2021.9661514
|View full text |Cite
|
Sign up to set email alerts
|

Learning from Demonstrations for Autonomous Soft-tissue Retraction

Abstract: The current research focus in Robot-Assisted Minimally Invasive Surgery (RAMIS) is directed towards increasing the level of robot autonomy, to place surgeons in a supervisory position. Although Learning from Demonstrations (LfD) approaches are among the preferred ways for an autonomous surgical system to learn expert gestures, they require a high number of demonstrations and show poor generalization to the variable conditions of the surgical environment. In this work, we propose an LfD methodology based on Gen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 22 publications
(6 citation statements)
references
References 20 publications
0
6
0
Order By: Relevance
“…Instead of exhaustively tuning the algorithm, hyperparameters, network architecture, and reward functions, the goal of these experiments is to show the effect of various environment configurations on task complexity, while keeping the learning setup constant. We train using Proximal Policy Optimization (PPO) (Schulman et al, 2017), as it is a popular algorithm that has been applied successfully to diverse problems in the literature (Andrychowicz et al, 2020;Mirhoseini et al, 2021;Pore et al, 2021b), thus making it suitable for use as a baseline. All environments were tested in several configurations, using either state or image observations, and varying up to two other environment-specific parameters.…”
Section: Reinforcement Learning Experimentsmentioning
confidence: 99%
See 1 more Smart Citation
“…Instead of exhaustively tuning the algorithm, hyperparameters, network architecture, and reward functions, the goal of these experiments is to show the effect of various environment configurations on task complexity, while keeping the learning setup constant. We train using Proximal Policy Optimization (PPO) (Schulman et al, 2017), as it is a popular algorithm that has been applied successfully to diverse problems in the literature (Andrychowicz et al, 2020;Mirhoseini et al, 2021;Pore et al, 2021b), thus making it suitable for use as a baseline. All environments were tested in several configurations, using either state or image observations, and varying up to two other environment-specific parameters.…”
Section: Reinforcement Learning Experimentsmentioning
confidence: 99%
“…In addition, images from the endoscopic camera are the primary source of information in surgical settings, but many existing environment suites do not support image observations. Recent works for automation in RALS propose algorithmic advances for solving surgical subtasks, but evaluate these algorithms in a) handcrafted simulation environments that are often not publicly available (Shin et al, 2019;Scheikl et al, 2021;He et al, 2022;Bourdillon et al, 2022) or b) on expensive robotic hardware such as the daVinci Research Kit (dVRK) (Tagliabue et al, 2020;Pore et al, 2021b;Chiu et al, 2021;Varier et al, 2020). This results in a high barrier for researchers to contribute to the field, and ultimately impedes vital algorithmic development.…”
Section: Introductionmentioning
confidence: 99%
“…Tissue manipulation is a common surgical task that encompasses both direct and indirect manipulation of deformable objects. Examples for direct manipulation are tissue retraction to visualize occluded structures [4]- [6] or organ manipulation to bring a deformable object into a desired shape [7], [8]. Indirect manipulation involves moving specific landmarks on tissue by altering the tissue's shape through manipulation of other points on the tissue.…”
Section: Related Workmentioning
confidence: 99%
“…Training in simulation enables end-to-end Reinforcement Learning (RL) of complex tasks in a safe and controlled environment, without requiring direct access to the real surgical robotic system. Training on real surgical robotic systems is impractical as RL algorithms often require millions of environment interactions to train a policy and unsafe behavior during training may damage the system [3].…”
Section: Introductionmentioning
confidence: 99%