» Articles » PMID: 38067211

Empowering Vision Transformer by Network Hyper-Parameter Selection for Whole Pelvis Prostate Planning Target Volume Auto-Segmentation

Overview
Journal Cancers (Basel)
Publisher MDPI
Specialty Oncology
Date 2023 Dec 9
PMID 38067211
Authors
Affiliations
Soon will be listed here.
Abstract

U-Net, based on a deep convolutional network (CNN), has been clinically used to auto-segment normal organs, while still being limited to the planning target volume (PTV) segmentation. This work aims to address the problems in two aspects: 1) apply one of the newest network architectures such as vision transformers other than the CNN-based networks, and 2) find an appropriate combination of network hyper-parameters with reference to recently proposed nnU-Net ("no-new-Net"). VT U-Net was adopted for auto-segmenting the whole pelvis prostate PTV as it consisted of fully transformer architecture. The upgraded version (v.2) applied the nnU-Net-like hyper-parameter optimizations, which did not fully cover the transformer-oriented hyper-parameters. Thus, we tried to find a suitable combination of two key hyper-parameters (patch size and embedded dimension) for 140 CT scans throughout 4-fold cross validation. The VT U-Net v.2 with hyper-parameter tuning yielded the highest dice similarity coefficient (DSC) of 82.5 and the lowest 95% Haussdorff distance (HD95) of 3.5 on average among the seven recently proposed deep learning networks. Importantly, the nnU-Net with hyper-parameter optimization achieved competitive performance, although this was based on the convolution layers. The network hyper-parameter tuning was demonstrated to be necessary even for the newly developed architecture of vision transformers.

Citing Articles

The InterVision Framework: An Enhanced Fine-Tuning Deep Learning Strategy for Auto-Segmentation in Head and Neck.

Choi B, Beltran C, Yoo S, Kwon N, Kim J, Park J J Pers Med. 2024; 14(9).

PMID: 39338233 PMC: 11432789. DOI: 10.3390/jpm14090979.


RadDeploy: A framework for integrating in-house developed software and artificial intelligence models seamlessly into radiotherapy workflows.

Rasmussen M, Dueholm Vestergaard C, Folsted Kallehauge J, Ren J, Haislund Guldberg M, Norrevang O Phys Imaging Radiat Oncol. 2024; 31:100607.

PMID: 39071159 PMC: 11283118. DOI: 10.1016/j.phro.2024.100607.

References
1.
Wang S, He K, Nie D, Zhou S, Gao Y, Shen D . CT male pelvic organ segmentation using fully convolutional networks with boundary sensitive representation. Med Image Anal. 2019; 54:168-178. PMC: 6506162. DOI: 10.1016/j.media.2019.03.003. View

2.
Shahedi M, Ma L, Halicek M, Guo R, Zhang G, Schuster D . A semiautomatic algorithm for three-dimensional segmentation of the prostate on CT images using shape and local texture characteristics. Proc SPIE Int Soc Opt Eng. 2018; 10576. PMC: 6149524. DOI: 10.1117/12.2293195. View

3.
Ma J, Chen J, Ng M, Huang R, Li Y, Li C . Loss odyssey in medical image segmentation. Med Image Anal. 2021; 71:102035. DOI: 10.1016/j.media.2021.102035. View

4.
Isensee F, Jaeger P, Kohl S, Petersen J, Maier-Hein K . nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods. 2020; 18(2):203-211. DOI: 10.1038/s41592-020-01008-z. View

5.
Schlemper J, Oktay O, Schaap M, Heinrich M, Kainz B, Glocker B . Attention gated networks: Learning to leverage salient regions in medical images. Med Image Anal. 2019; 53:197-207. PMC: 7610718. DOI: 10.1016/j.media.2019.01.012. View