Command Palette
Search for a command to run...
{Joohan Yoo Hyunnam Lee}
Abstract
Transformer-based methods have become dominant in the medical image research field since the Vision Transformer achieved superior performance. Although transformer-based approaches have resolved long-range dependency problems inherent in Convolutional Neural Network (CNN) methods, they struggle to capture local detail information. Recent research focuses on the robust combination of local detail and semantic information. To address this problem, we propose a novel transformer-CNN hybrid network named RAPUNet. The proposed approach employs MetaFormer as the transformer backbone and introduces a custom convolutional block, RAPU (Residual and Atrous Convolution in Parallel Unit), to enhance local features and alleviate the combination problem of local and global features. We evaluate the segmentation performance of RAPUNet on popular benchmarking datasets for polyp segmentation, including Kvasir-SEG, CVC-ClinicDB, CVC-ColonDB, EndoScene-CVC300, and ETIS-LaribPolypDB. Experimental results show that our model achieves competitive performance in terms of mean Dice and mean IoU. Particularly, RAPUNet outperforms state-of-the-art methods on the CVC-ClinicDB dataset. Code available: https://github.com/hyunnamlee/RAPUNet .
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| medical-image-segmentation-on-cvc-clinicdb | RAPUNet | mIoU: 0.926 mean Dice: 0.961 |
| medical-image-segmentation-on-cvc-colondb | RAPUNet | mIoU: 0.9096 mean Dice: 0.9526 |
| medical-image-segmentation-on-etis | RAPUNet | mIoU: 0.9179 mean Dice: 0.9572 |
| medical-image-segmentation-on-kvasir-seg | RAPUNet | mIoU: 0.885 mean Dice: 0.939 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.