2019
DOI: 10.1007/978-3-030-32254-0_14
|View full text |Cite
|
Sign up to set email alerts
|

Generating Large Labeled Data Sets for Laparoscopic Image Processing Tasks Using Unpaired Image-to-Image Translation

Abstract: In the medical domain, the lack of large training data sets and benchmarks is often a limiting factor for training deep neural networks. In contrast to expensive manual labeling, computer simulations can generate large and fully labeled data sets with a minimum of manual effort. However, models that are trained on simulated data usually do not translate well to real scenarios. To bridge the domain gap between simulated and real laparoscopic images, we exploit recent advances in unpaired image-to-image translat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
53
0
1

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 67 publications
(64 citation statements)
references
References 9 publications
0
53
0
1
Order By: Relevance
“…An alternative direction to mitigate the dependency on annotated video sequences is to utilize synthetic data for the training of DNNs. Recent advances in graphics and simulation infrastructures have paved the way to automatically create a large number of photo-realistic simulated images with accurate pixel-level labels [14,23]. However, the DNNs trained purely on simulated images do not generalize well on real endoscopic videos due to the domain shift/bias issue [30,32].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…An alternative direction to mitigate the dependency on annotated video sequences is to utilize synthetic data for the training of DNNs. Recent advances in graphics and simulation infrastructures have paved the way to automatically create a large number of photo-realistic simulated images with accurate pixel-level labels [14,23]. However, the DNNs trained purely on simulated images do not generalize well on real endoscopic videos due to the domain shift/bias issue [30,32].…”
Section: Introductionmentioning
confidence: 99%
“…Here, the primary goal is to learn domaininvariant feature representations for addressing the domain shift/bias [14,35]. For instance, Pfeiffer et al [23] utilized an image-to-image translation approach, where the simulated images are translated into realistic looking ones by mapping image styles (texture, lighting) of the real data using a Cycle-GAN. In contrast, we argued for a shape-focused joint learning from simulated and real data in an end-to-end fashion and introduced a consistency-learning-based approach Endo-Sim2Real [27] to align the DNNs on both domains.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, generative methods [12] , [25] have shown potential to address the data scarcity problem in endoscopic vision. Alternatively to just reducing the number of labels or time required for dataset curation, dataset synthesis has emerged as an inexpensive approach to generate annotated images automatically.…”
Section: Related Workmentioning
confidence: 99%
“…Several approaches to reducing the annotation effort have been proposed, such as active learning [8][9][10], where only the most informative data points are selected and then are annotated, as well as crowdsourcing, where the "wisdom of the crowd" can be utilized for certain clinical tasks [11,12]. A promising pathway for overcoming the lack of annotated data is to generate realistic synthetic images based on a simple simulation by using generative adversarial networks [13] (Fig. 2).…”
Section: Data Annotationmentioning
confidence: 99%