Ms RED: A novel multi-scale residual encoding and decoding network for skin lesion segmentation
Affiliation:
1. Institute of Medical Artificial Intelligence, the Second Affiliated Hospital of Xi’an Jiaotong University, Xi’an, 710061, China;2. Australian Institute for Machine Learning, University of Adelaide, Adelaide, 5005, Australia;3. Affiliated Hospital of Jining Medical University, Jining, 272000, China;1. College of Information Science and Engineering, Xinjiang University, Urumqi 830000, China;2. College of Computer Science, Sichuan University, Chengdu 610065, China;3. College of Software Engineering, Xin Jiang University, Urumqi 830000, China;4. Key Laboratory of Software Engineering Technology, Xinjiang University, China;5. Xinjiang Key Laboratory of Dermatology Research, People''s Hospital of Xinjiang Uygur Autonomous Region, China;6. The First Affiliated Hospital of Xinjiang Medical University, Urumqi 830000, China;1. School of Electrical and Electronic Engineering, Nanyang Technological University, 639798, Singapore;2. School of Automation, Central South University, Hunan 410083, China;3. Information Systems Technology and Design Pillar, Singapore University of Technology and Design, 487372, Singapore;1. College of Mechanical Engineering, Quzhou University, Quzhou, Zhejiang, 324000, China;2. School of Mechanical Engineering, Hangzhou Dianzi University, Hangzhou, Zhejiang, 310018, China
Abstract:
Computer-Aided Diagnosis (CAD) for dermatological diseases offers one of the most notable showcases where deep learning technologies display their impressive performance in acquiring and surpassing human experts. In such the CAD process, a critical step is concerned with segmenting skin lesions from dermoscopic images. Despite remarkable successes attained by recent deep learning efforts, much improvement is still anticipated to tackle challenging cases, e.g., segmenting lesions that are irregularly shaped, bearing low contrast, or possessing blurry boundaries. To address such inadequacies, this study proposes a novel Multi-scale Residual Encoding and Decoding network (Ms RED) for skin lesion segmentation, which is able to accurately and reliably segment a variety of lesions with efficiency. Specifically, a multi-scale residual encoding fusion module (MsR-EFM) is employed in an encoder, and a multi-scale residual decoding fusion module (MsR-DFM) is applied in a decoder to fuse multi-scale features adaptively. In addition, to enhance the representation learning capability of the newly proposed pipeline, we propose a novel multi-resolution, multi-channel feature fusion module (M2F2), which replaces conventional convolutional layers in encoder and decoder networks. Furthermore, we introduce a novel pooling module (Soft-pool) to medical image segmentation for the first time, retaining more helpful information when down-sampling and getting better segmentation performance. To validate the effectiveness and advantages of the proposed network, we compare it with several state-of-the-art methods on ISIC 2016, 2017, 2018, and PH2. Experimental results consistently demonstrate that the proposed Ms RED attains significantly superior segmentation performance across five popularly used evaluation criteria. Last but not least, the new model utilizes much fewer model parameters than its peer approaches, leading to a greatly reduced number of labeled samples required for model training, which in turn produces a substantially faster converging training process than its peers. The source code is available at https://github.com/duweidai/Ms-RED.