Web-Books
in the Austria-Forum
Austria-Forum
Web-Books
International
Proceedings of the OAGM&ARW Joint Workshop - Vision, Automation and Robotics
Page - 140 -
  • User
  • Version
    • full version
    • text only version
  • Language
    • Deutsch - German
    • English

Page - 140 - in Proceedings of the OAGM&ARW Joint Workshop - Vision, Automation and Robotics

Image of the Page - 140 -

Image of the Page - 140 - in Proceedings of the OAGM&ARW Joint Workshop - Vision, Automation and Robotics

Text of the Page - 140 -

Generative Adversarial Network based Synthesis for Supervised Medical Image Segmentation* Thomas Neff1, Christian Payer1, Darko Sˇtern2, Martin Urschler2 Abstract—Modern deep learning methods achieve state-of- the-art results in many computer vision tasks. While these methods perform well when trained on large datasets, deep learning methods suffer from overfitting and lack of gener- alization given smaller datasets. Especially in medical image analysis, acquisition of both imaging data and corresponding ground-truth annotations (e.g. pixel-wise segmentation masks) as required for supervised tasks, is time consuming and costly, sinceexpertsareneededtomanuallyannotatedata. Inthiswork westudythisproblembyproposinganewvariantofGenerative AdversarialNetworks (GANs),which, inadditiontosynthesized medical images, also generates segmentation masks for the use in supervised medical image analysis applications. We evaluate our approach on a lung segmentation task involving thorax X-ray images, and show that GANs have the potential to be used for synthesizing training data in this specific application. I. INTRODUCTION Modern machine learning methods based on deep neural network architectures require large amounts of training data to achieve the best possible results. For standard com- puter vision problems, large datasets, such as MNIST [12], CIFAR10 [10], or ImageNet [23], containing millions of images, are publicly available. In the medical field, datasets are typically smaller by several orders of magnitude, as the acquisition process of medical images is costly and time consuming. Furthermore, ethical concerns make it harder to publicly release and share datasets. Finding methods to improve performance when training deep learning methods on small datasets is an area of active research. Recent work in the medical imaging domain has shown that it is possible to improve performance with small datasets by putting application specific prior knowledge into a deep neural network [17]. Another approach has been made popular by the U-Net [21] architecture for biomedical image segmentation, which demonstrated how strong data augmentation can be used to deal with low amounts of training data in deep network architectures. Even though data augmentation is simple to implement and achieves good results, it isonlyable toproducefixedvariationsofanygiven dataset, requiring the augmentation to fit the given dataset. Transfer learning approaches such as [19] show that training on large datasets (e.g. ImageNet) followed by fine- tuning on a small dataset achieves state-of-the-art results for *Thisworkwassupportedby theAustrianScienceFund(FWF):P28078- N33. 1Thomas Neff and Christian Payer are with the Institute for Com- puter Graphics and Vision, Graz University of Technology, Austria thomas.neff@student.tugraz.at 2Darko Sˇtern and Martin Urschler are with Ludwig Boltz- mann Institute for Clinical Forensic Imaging, Graz, Austria martin.urschler@cfi.lbg.ac.at datasets consisting of natural images. For medical imaging, the learned features from large natural image datasets may not be suitable, as the image features are very different com- pared to natural images. Furthermore, there is no straight- forward way of transferring 2D features to 3D features, which poses a limitation when working with 3D medical images. Due to the difference in features between medical and natural images, another approach is to use unsupervised feature extractors (e.g. Autoencoders [27]) which are trained on medical images only. Nevertheless, transferring weights learned by these unsupervised methods requires the target network architecture to be close to the source architecture, which is rarely the case. The requirement for large amounts of training data also popularized image generation methods in deep learning con- texts. Recently, research has shown that Generative Adver- sarial Networks (GANs) [4] can be used for a large variety of applications such as image-to-image translation [6] or unsu- pervised representation learning [18]. GANs have also been successfully used for unsupervised domain adaptation [8] of multi-modal medical imaging data, demonstrating their potential for use with small medical imaging datasets. Our goal was to use GANs in a completely different way, by using the high quality of the generated images to augment our small set of training data. We propose a novel modifica- tion to GANs, which generates new, synthetic images as well as thecorrespondingsegmentationmasks fromrandomnoise. This allows us to use the synthetic data as training data for a supervised segmentation task. We show that this architecture manages to produce convincing segmentation masks for the generated images. We evaluate the generated images in two different scenarios on an image segmentation task and show that training on purely generated images achieves results comparable to trainingonreal images forverysmalldatasets. II. RELATED WORK A. Training Data Augmentation Training data augmentation is a commonly used method to reduce the effects of overfitting with small training datasets as well as improve the generalization of the trained network. Most machine learning frameworks allow for simple aug- mentation such as rotation, translation or intensity shifts of training data. AlexNet [11] was one of the first convolutional neural network (CNN) architectures to implement online data augmentation with successful results. However, data augmentation only achieves good results if the augmentation can actually occur in the data, and is relevant to the required application. For medical imaging, elastic deformations [21] 140
back to the  book Proceedings of the OAGM&ARW Joint Workshop - Vision, Automation and Robotics"
Proceedings of the OAGM&ARW Joint Workshop Vision, Automation and Robotics
Title
Proceedings of the OAGM&ARW Joint Workshop
Subtitle
Vision, Automation and Robotics
Authors
Peter M. Roth
Markus Vincze
Wilfried Kubinger
Andreas Müller
Bernhard Blaschitz
Svorad Stolc
Publisher
Verlag der Technischen Universität Graz
Location
Wien
Date
2017
Language
English
License
CC BY 4.0
ISBN
978-3-85125-524-9
Size
21.0 x 29.7 cm
Pages
188
Keywords
Tagungsband
Categories
International
Tagungsbände

Table of contents

  1. Preface v
  2. Workshop Organization vi
  3. Program Committee OAGM vii
  4. Program Committee ARW viii
  5. Awards 2016 ix
  6. Index of Authors x
  7. Keynote Talks
  8. Austrian Robotics Workshop 4
  9. OAGM Workshop 86
Web-Books
Library
Privacy
Imprint
Austria-Forum
Austria-Forum
Web-Books
Proceedings of the OAGM&ARW Joint Workshop