Diverse data augmentation for learning image segmentation with cross-modality annotations |
| |
Affiliation: | 1. Department of Radiology and Biomedical Research Imaging Center (BRIC), University of North Carolina, Chapel Hill, NC, USA;2. Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, TX, USA;3. Department of Radiology, Houston Methodist Hospital, TX, USA;1. BCN MedTech, Universitat Pompeu Fabra, Barcelona, Spain;2. BASIRA lab, Faculty of Computer and Informatics, Istanbul Technical University, Istanbul, Turkey;3. School of Science and Engineering, Computing, University of Dundee, UK;4. Fetal i+D Fetal Medicine Research Center, BCNatal - Barcelona Center for Maternal-Fetal and Neonatal Medicine (Hospital Clínic and Hospital Sant Joan de Déu), Institut Clínic de Ginecologia, Obstetricia i Neonatologia, IDIBAPS, Universitat de Barcelona, Barcelona, Spain;5. Centre for Biomedical Research on Rare Diseases (CIBER-ER), Barcelona, Spain;6. Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, USA;7. Department of Brain and Cognitive Engineering, Korea University, Seoul, Republic of Korea;8. German Center for Neurodegenerative Diseases, Bonn, Germany;9. ICREA, Barcelona, Spain;1. Department of Computer Science and Technology, Shandong University, Jinan 250100, China;2. Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC 27599, USA;3. Digital Media Technology Key Lab of Shandong Province, Jinan 250061, China;4. Department of Software, Shandong University, Jinan 250100, China;5. Department of Brain and Cognitive Engineering, Korea University, Seoul 02841, Republic of Korea |
| |
Abstract: | The dearth of annotated data is a major hurdle in building reliable image segmentation models. Manual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural features related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualitatively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT. |
| |
Keywords: | |
本文献已被 ScienceDirect 等数据库收录! |
|