TY - GEN
T1 - BiTr-Unet
T2 - 7th International Brain Lesion Workshop, BrainLes 2021, held in conjunction with the Medical Image Computing and Computer Assisted Intervention, MICCAI 2021
AU - Jia, Qiran
AU - Shu, Hai
N1 - Funding Information:
Acknowledgements. This research was partially supported by the grant R21AG0 70303 from the National Institutes of Health and a startup fund from New York University. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health or New York University.
Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2022
Y1 - 2022
N2 - Convolutional neural networks (CNNs) have achieved remarkable success in automatically segmenting organs or lesions on 3D medical images. Recently, vision transformer networks have exhibited exceptional performance in 2D image classification tasks. Compared with CNNs, transformer networks have an appealing advantage of extracting long-range features due to their self-attention algorithm. Therefore, we propose a CNN-Transformer combined model, called BiTr-Unet, with specific modifications for brain tumor segmentation on multi-modal MRI scans. Our BiTr-Unet achieves good performance on the BraTS2021 validation dataset with median Dice score 0.9335, 0.9304 and 0.8899, and median Hausdorff distance 2.8284, 2.2361 and 1.4142 for the whole tumor, tumor core, and enhancing tumor, respectively. On the BraTS2021 testing dataset, the corresponding results are 0.9257, 0.9350 and 0.8874 for Dice score, and 3, 2.2361 and 1.4142 for Hausdorff distance. The code is publicly available at https://github.com/JustaTinyDot/BiTr-Unet.
AB - Convolutional neural networks (CNNs) have achieved remarkable success in automatically segmenting organs or lesions on 3D medical images. Recently, vision transformer networks have exhibited exceptional performance in 2D image classification tasks. Compared with CNNs, transformer networks have an appealing advantage of extracting long-range features due to their self-attention algorithm. Therefore, we propose a CNN-Transformer combined model, called BiTr-Unet, with specific modifications for brain tumor segmentation on multi-modal MRI scans. Our BiTr-Unet achieves good performance on the BraTS2021 validation dataset with median Dice score 0.9335, 0.9304 and 0.8899, and median Hausdorff distance 2.8284, 2.2361 and 1.4142 for the whole tumor, tumor core, and enhancing tumor, respectively. On the BraTS2021 testing dataset, the corresponding results are 0.9257, 0.9350 and 0.8874 for Dice score, and 3, 2.2361 and 1.4142 for Hausdorff distance. The code is publicly available at https://github.com/JustaTinyDot/BiTr-Unet.
KW - Brain tumor
KW - Deep learning
KW - Multi-modal image segmentation
KW - Vision transformer
UR - http://www.scopus.com/inward/record.url?scp=85135160546&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85135160546&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-09002-8_1
DO - 10.1007/978-3-031-09002-8_1
M3 - Conference contribution
AN - SCOPUS:85135160546
SN - 9783031090011
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 3
EP - 14
BT - Brainlesion
A2 - Crimi, Alessandro
A2 - Bakas, Spyridon
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 27 September 2021 through 27 September 2021
ER -