COLONOSCOPIC POLYP SEGMENTATION USING SEGFORMER-B0 WITH A DICE-BCE HYBRID LOSS

Authors

  • Ahmad Yani Universitas Teknologi Mataram
  • San Sudirman Universitas Teknologi Mataram
  • M. Zulpahmi Universitas Teknologi Mataram
  • Emi Suryadi Universitas Teknologi Mataram
  • Bahtiar Imran Universitas Teknologi Mataram

DOI:

https://doi.org/10.69916/jkbti.v5i2.476

Keywords:

computer-aided diagnosis, dice-bce loss, kvasir-seg, medical image segmentation, segformer-b0

Abstract

Colorectal cancer is one of the leading causes of cancer-related deaths worldwide, with most cases originating from early lesions such as colon polyps. Early detection through colonoscopy is essential to reduce mortality rates; however, accurate polyp identification remains challenging due to variations in shape, size, texture, and illumination conditions. This study aims to implement and evaluate the SegFormer-B0 architecture combined with a Dice-BCE hybrid loss function for polyp segmentation in colonoscopy images. The study utilized the public Kvasir-SEG dataset consisting of 1,000 colonoscopy images with pixel-level annotations. The dataset was divided into 80% training data and 20% validation data. Image preprocessing included resizing to 256×256 pixels and normalization using ImageNet statistics. The model was trained for 25 epochs using the AdamW optimizer with a learning rate of 1×10⁻⁴. Performance evaluation was conducted using Dice Coefficient, Intersection over Union (IoU), Sensitivity, and Specificity metrics. The experimental results demonstrated that the proposed model achieved a Dice Coefficient of 89.92%, Mean IoU of 81.90%, Sensitivity of 89.12%, and Specificity of 98.51%. The training process also showed stable convergence, supported by a training loss of 7.53% and validation loss of 23.30%. The findings indicate that the integration of SegFormer-B0 with the Dice-BCE hybrid loss effectively improves segmentation accuracy and stability while addressing class imbalance issues in colonoscopy images. Therefore, the proposed approach has strong potential to support computer-aided diagnosis systems for colorectal cancer screening.

Downloads

Download data is not yet available.

References

X. Liu and others, “Colorectal Polyp Segmentation Based on Deep Learning Methods: A Systematic Review,” J. Imaging, vol. 11, no. 9, p. 293, 2025, doi: 10.3390/jimaging11090293.

E. Goceri, “Polyp Segmentation Using a Hybrid Vision Transformer and a Hybrid Loss Function,” Journal of Imaging Informatics in Medicine, 2024, doi: 10.1007/s10278-024-00915-7.

K. He and others, “Deep Residual Learning for Image Recognition,” CVPR, 2016, doi: 10.1109/CVPR.2016.90.

O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” May 2015, [Online]. Available: http://arxiv.org/abs/1505.04597

O. Oktay and others, “Attention U-Net: Learning Where to Look for the Pancreas,” arXiv preprint arXiv:1804.03999, 2018.

S. Kumar and others, “Dilated U-Net-Seg for Polyp Segmentation,” Biomed. Signal Process. Control, vol. 92, p. 105975, 2024, doi: 10.1016/j.bspc.2024.105975.

J. Mei and others, “A Survey on Deep Learning for Polyp Segmentation,” Visual Intelligence, 2025, doi: 10.1007/s44267-024-00071-w.

Z. Zhou and others, “UNet++: A Nested U-Net Architecture for Medical Image Segmentation,” Deep Learning in Medical Image Analysis, 2018, doi: 10.1007/978-3-030-00889-5_1.

A. Dosovitskiy and others, “An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale,” ICLR, 2021.

Z. Liu and others, “Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows,” ICCV, 2021, doi: 10.1109/ICCV48922.2021.00986.

H. Cao and others, “Transformer-Based Medical Image Segmentation: A Review,” Med. Image Anal., vol. 85, p. 102762, 2023, doi: 10.1016/j.media.2023.102762.

D. Liu and others, “NA-SegFormer: A Multi-Level Transformer Model for Medical Image Segmentation,” Sci. Rep., vol. 14, p. 21567, 2024, doi: 10.1038/s41598-024-71523-4.

E. Xie and others, “SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers,” NeurIPS, 2021.

H. Cao et al., “Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation,” May 2021, [Online]. Available: http://arxiv.org/abs/2105.05537

S. Taghanaki and others, “Combo Loss: Handling Input and Output Imbalance in Multi-Organ Segmentation,” Computerized Medical Imaging and Graphics, vol. 85, p. 101744, 2021, doi: 10.1016/j.compmedimag.2020.101744

Downloads

Published

2026-05-10

PlumX Metrics

Scite Metrics

Altmetric

How to Cite

[1]
A. Yani, S. Sudirman, M. Zulpahmi, E. Suryadi, and B. Imran, “COLONOSCOPIC POLYP SEGMENTATION USING SEGFORMER-B0 WITH A DICE-BCE HYBRID LOSS”, JKBTI, vol. 5, no. 2, pp. 235–242, May 2026.

Issue

Section

Articles

Most read articles by the same author(s)