Medical image translation using an edge-guided generative adversarial network with global-to-local feature fusion |
| |
Authors: | Hamed Amini Amirkolaee Hamid Amini Amirkolaee |
| |
Affiliation: | 1.School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran, Tehran 1417935840, Iran2.Civil and Geomatics Engineering Faculty, Tafresh State University, Tafresh 7961139518, Iran |
| |
Abstract: | In this paper, we propose a framework based deep learning for medical image translation using paired and unpaired training data. Initially, a deep neural network with an encoder-decoder structure is proposed for image-to-image translation using paired training data. A multi-scale context aggregation approach is then used to extract various features from different levels of encoding, which are used during the corresponding network decoding stage. At this point, we further propose an edge-guided generative adversarial network for image-to-image translation based on unpaired training data. An edge constraint loss function is used to improve network performance in tissue boundaries. To analyze framework performance, we conducted five different medical image translation tasks. The assessment demonstrates that the proposed deep learning framework brings significant improvement beyond state-of-the-arts. |
| |
Keywords: | edge-guided generative adversarial network global to local medical image translation magnetic resonance imaging computed tomography |
|
| 点击此处可从《生物医学研究杂志》浏览原始摘要信息 |
|
点击此处可从《生物医学研究杂志》下载免费的PDF全文 |
|