HALF-MAFUNET: A Lightweight Architecture Based on Multi-Scale Adaptive Fusion for Medical Image Segmentation

Keywords: Medical images segmentation, Deep learning, U-Net, Efficient Model

Abstract

Medical image segmentation is a critical component in computer-aided diagnosis systems but many deep learning models still require large numbers of parameters and heavy computation. Classical CNN-based architectures such as U-Net and its variants achieve good accuracy, but are often too heavy for real deployment. Meanwhile, modern Transformer-based or Mamba-based models capture long-range information but typically increase model complexity. Because of these limitations, there is still a need for a lightweight segmentation model that can provide a good balance between accuracy and efficiency across different types of medical images. This paper proposes Half-MAFUNet, a lightweight architecture based on multi-scale adaptive fusion and designed as a simplified version of MAFUNet. The main contribution of this work is combining the efficient encoder structure of Half-UNet with advanced fusion and attention mechanisms. Half-MAFUNet integrates Hierarchy Aware Mamba (HAM) for global feature modelling, Multi-Scale Adaptive Fusion (MAF) to combine global and local information, and two attention modules, Adaptive Channel Attention (ACA) and Adaptive Spatial Attention (ASA), to refine skip connections. In addition, this model incorporates Channel Atrous Spatial Pyramid Pooling (CASPP) to capture multi-scale receptive fields efficiently without increasing computational cost. Together, these components create a compact architecture that maintains strong representational power. The model is trained and evaluated on three public datasets: CVC-ClinicDB for colorectal polyp segmentation, BUSI for breast tumor segmentation, and ISIC-2018 for skin lesion segmentation. All images are resized to 256×256 pixels and processed using geometric and intensity-based augmentations. Half-MAFUNet achieves competitive performance, obtaining mean IoU around 84 85% and Dice/F1-Score around 90 92% across datasets, while using significantly fewer parameters and GFLOPs compared to U-Net, Att-UNet, UNeXt, MALUNet, LightM-UNet, VM-UNet, and UD-Mamba. These results show that Half-MAFUNet provides accurate and efficient medical image segmentation, making it suitable for real-world deployment on devices with limited computational resources.

Downloads

Download data is not yet available.

References

M. G. Linguraru et al., “Clinical, Cultural, Computational, and Regulatory Considerations to Deploy AI in Radiology: Perspectives of RSNA and MICCAI Experts,” Radiol Artif Intell, vol. 6, no. 4, Jul. 2024, doi: 10.1148/ryai.240225.

J. Zhang et al., “Advances in attention mechanisms for medical image segmentation,” May 01, 2025, Elsevier Ireland Ltd. doi: 10.1016/j.cosrev.2024.100721.

Olivier Colliot, “Machine Learning for Brain Disorders,” 2023. doi: https://doi.org/10.1007/978-1-0716-3195-9.

S. Kolhar and J. Jagtap, “Convolutional neural network based encoder-decoder architectures for semantic segmentation of plants,” Ecol Inform, vol. 64, Sep. 2021, doi: 10.1016/j.ecoinf.2021.101373.

O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Springer Verlag, 2015, pp. 234-241. doi: 10.1007/978-3-319-24574-4_28.

Y. Gao, M. Zhou, D. Liu, Z. Yan, S. Zhang, and D. N. Metaxas, “A Data-scalable Transformer for Medical Image Segmentation: Architecture, Model Efficiency, and Benchmark,” Apr. 2023, [Online]. Available: http://arxiv.org/abs/2203.00131

L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking Atrous Convolution for Semantic Image Segmentation,” Dec. 2017, [Online]. Available: http://arxiv.org/abs/1706.05587

A. Dosovitskiy et al., “An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale,” Jun. 2021, [Online]. Available: http://arxiv.org/abs/2010.11929

R. Wu, Y. Liu, P. Liang, and Q. Chang, “H-vmunet: High-order Vision Mamba UNet for medical image segmentation,” Neurocomputing, vol. 624, Apr. 2025, doi: 10.1016/j.neucom.2025.129447.

A. Gu and T. Dao, “Mamba: Linear-Time Sequence Modeling with Selective State Spaces,” May 2024, [Online]. Available: http://arxiv.org/abs/2312.00752

Y. Liu et al., “VMamba: Visual State Space Model.” [Online]. Available: https://github.com/MzeroMiko/VMamba

J. Ruan, J. Li, and S. Xiang, “VM-UNet: Vision Mamba UNet for Medical Image Segmentation,” Nov. 2024, [Online]. Available: http://arxiv.org/abs/2402.02491

J. Xu, “HC-Mamba: Vision MAMBA with Hybrid Convolutional Techniques for Medical Image Segmentation,” Oct. 2024, [Online]. Available: http://arxiv.org/abs/2405.05007

W. Liao, Y. Zhu, X. Wang, C. Pan, Y. Wang, and L. Ma, “LightM-UNet: Mamba Assists in Lightweight UNet for Medical Image Segmentation,” Mar. 2024, [Online]. Available: http://arxiv.org/abs/2403.05246

R. Wu, Y. Liu, P. Liang, and Q. Chang, “UltraLight VM-UNet: Parallel Vision Mamba Significantly Reduces Parameters for Skin Lesion Segmentation,” Apr. 2024, doi: 10.1016/j.patter.2025.101298.

X. Zhu, W. Wang, C. Zhang, and H. Wang, “Polyp-Mamba: A Hybrid Multi-Frequency Perception Gated Selection Network for polyp segmentation,” Information Fusion, vol. 115, Mar. 2025, doi: 10.1016/j.inffus.2024.102759.

M. Yang, Z. Yang, and N. I. R. Ruhaiyem, “MAFUNet: Mamba with adaptive fusion UNet for medical image segmentation,” Image Vis Comput, vol. 162, Oct. 2025, doi: 10.1016/j.imavis.2025.105655.

O. Oktay et al., “Attention U-Net: Learning Where to Look for the Pancreas,” May 2018, [Online]. Available: http://arxiv.org/abs/1804.03999

H. Lu, Y. She, J. Tie, and S. Xu, “Half-UNet: A Simplified U-Net Architecture for Medical Image Segmentation,” Front Neuroinform, vol. 16, Jun. 2022, doi: 10.3389/fninf.2022.911679.

J. Bernal, F. J. Sánchez, G. Fernández-Esparrach, D. Gil, C. Rodríguez, and F. Vilariño, “WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians,” Computerized Medical Imaging and Graphics, vol. 43, pp. 99-111, Jul. 2015, doi: 10.1016/j.compmedimag.2015.02.007.

W. Al-Dhabyani, M. Gomaa, H. Khaled, and A. Fahmy, “Dataset of breast ultrasound images,” Data Brief, vol. 28, Feb. 2020, doi: 10.1016/j.dib.2019.104863.

F. C. N. Codella et al., SKIN LESION ANALYSIS TOWARD MELANOMA DETECTION: A CHALLENGE AT THE 2017 INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING(ISBI), HOSTED BY THE INTERNATIONAL SKIN IMAGING COLLABORATION (ISIC). IEEE, 2018.

R. Alkhatib, “Artificial Neural Network Activation Functions in Exact Analytical Form (Heaviside, ReLU, PReLU, ELU, SELU, ELiSH).”

B. D. Sarira and H. Prasetyo, “Dual Attention and Channel Atrous Spatial Pyramid Pooling Half-UNet for Polyp Segmentation,” Journal of Electronics, Electromedical Engineering, and Medical Informatics, vol. 7, no. 3, pp. 680-691, Jul. 2025, doi: 10.35882/jeeemi.v7i3.893.

J. Wang, P. Lv, H. Wang, and C. Shi, “SAR-U-Net: Squeeze-and-excitation block and atrous spatial pyramid pooling based residual U-Net for automatic liver segmentation in Computed Tomography,” Comput Methods Programs Biomed, vol. 208, Sep. 2021, doi: 10.1016/j.cmpb.2021.106268.

S. R. Dubey, S. K. Singh, and B. B. Chaudhuri, “Activation functions in deep learning: A comprehensive survey and benchmark,” Sep. 07, 2022, Elsevier B.V. doi: 10.1016/j.neucom.2022.06.111.

M. Lee, “Mathematical Analysis and Performance Evaluation of the GELU Activation Function in Deep Learning,” Journal of Mathematics, vol. 2023, 2023, doi: 10.1155/2023/4229924.

J. M. J. Valanarasu and V. M. Patel, “UNeXt: MLP-Based Rapid Medical Image Segmentation Network,” 2022. doi: https://doi.org/10.1007/978-3-031-16443-9_3.

J. Ruan, S. Xiang, M. Xie, T. Liu, and Y. Fu, “MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation,” in Proceedings - 2022 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2022, Institute of Electrical and Electronics Engineers Inc., 2022, pp. 1150-1156. doi: 10.1109/BIBM55620.2022.9995040.

C. Li et al., “FocalUNETR: A Focal Transformer for Boundary-aware Segmentation of CT Images,” Jul. 2023, [Online]. Available: http://arxiv.org/abs/2210.03189

C. Li et al., “U-KAN Makes Strong Backbone for Medical Image Segmentation and Generation,” 2025. [Online]. Available: https://arxiv.org/pdf/2406.02918

W. Zhao, F. Wang, Y. Wang, Y. Xie, Q. Wu, and Y. Zhou, “UD-Mamba: A pixel-level uncertainty-driven Mamba model for medical image segmentation,” Feb. 2025, [Online]. Available: http://arxiv.org/abs/2502.02024

Published
2026-01-12
How to Cite
[1]
A. F. Maula Sandy and H. Prasetyo, “HALF-MAFUNET: A Lightweight Architecture Based on Multi-Scale Adaptive Fusion for Medical Image Segmentation”, j.electron.electromedical.eng.med.inform, vol. 8, no. 1, pp. 222-239, Jan. 2026.
Section
Medical Engineering