MDA-Unet: A Multi-Scale Dilated Attention U-Net For Medical Image Segmentation

Amer, Alyaa, Lambrou, Tryphon and Ye, Xujiong (2022) MDA-Unet: A Multi-Scale Dilated Attention U-Net For Medical Image Segmentation. Applied Sciences, 12 (7). ISSN 2076-3417

Full content URL:

MDA-Unet: A Multi-Scale Dilated Attention U-Net For Medical Image Segmentation
MDA-Unet-applsci.pdf - Whole Document
Available under License Creative Commons Attribution 4.0 International.

Item Type:Article
Item Status:Live Archive


The advanced development of deep learning methods has recently made significant improvements in medical image segmentation. Encoder–decoder networks, such as U-Net, have addressed some of the challenges in medical image segmentation with an outstanding performance, which has promoted them to be the most dominating deep learning architecture in this domain. Despite their outstanding performance, we argue that they still lack some aspects. First, there is incompatibility in U-Net’s skip connection between the encoder and decoder features due to the semantic gap between low-processed encoder features and highly processed decoder features, which adversely affects the final prediction. Second, it lacks capturing multi-scale context information and ignores the contribution of all semantic information through the segmentation process. Therefore, we propose a model named MDA-Unet, a novel multi-scale deep learning segmentation model. MDA-Unet improves upon U-Net and enhances its performance in segmenting medical images with variability in the shape and size of the region of interest. The model is integrated with a multi-scale spatial attention module, where spatial attention maps are derived from a hybrid hierarchical dilated convolution module that captures multi-scale context information. To ease the training process and reduce the gradient vanishing problem, residual blocks are deployed instead of the basic U-net blocks. Through a channel attention mechanism, the high-level decoder features are used to guide the low-level encoder features to promote the selection of meaningful context information, thus ensuring effective fusion. We evaluated our model on 2 different datasets: a lung dataset of 2628 axial CT images and an echocardiographic dataset of 2000 images, each with its own challenges. Our model has achieved a significant gain in performance with a slight increase in the number of trainable parameters in comparison with the basic U-Net model, providing a dice score of 98.3% on the lung dataset and 96.7% on the echocardiographic dataset, where the basic U-Net has achieved 94.2% on the lung dataset and 93.9% on the echocardiographic dataset.

Keywords:deep learning, U-Net, medical images, segmentation, computed tomography, echocardiography
Subjects:G Mathematical and Computer Sciences > G400 Computer Science
Divisions:College of Science
ID Code:48943
Deposited On:13 Apr 2022 13:46

Repository Staff Only: item control page