TY - JOUR
T1 - CrossMoDA 2021 challenge: Benchmark of cross-modality domain adaptation techniques for vestibular schwannoma and cochlea segmentation
AU - Dorent, Reuben
AU - Kujawa, Aaron
AU - Ivory, Marina
AU - Bakas, Spyridon
AU - Rieke, Nicola
AU - Joutard, Samuel
AU - Glocker, Ben
AU - Cardoso, Jorge
AU - Modat, Marc
AU - Batmanghelich, Kayhan
AU - Belkov, Arseniy
AU - Calisto, Maria Baldeon
AU - Choi, Jae Won
AU - Dawant, Benoit M.
AU - Dong, Hexin
AU - Escalera, Sergio
AU - Fan, Yubo
AU - Hansen, Lasse
AU - Heinrich, Mattias P.
AU - Joshi, Smriti
AU - Kashtanova, Victoriya
AU - Kim, Hyeon Gyu
AU - Kondo, Satoshi
AU - Kruse, Christian N.
AU - Lai-Yuen, Susana K.
AU - Li, Hao
AU - Liu, Han
AU - Ly, Buntheng
AU - Oguz, Ipek
AU - Shin, Hyungseob
AU - Shirokikh, Boris
AU - Su, Zixian
AU - Wang, Guotai
AU - Wu, Jianghao
AU - Xu, Yanwu
AU - Yao, Kai
AU - Zhang, Li
AU - Ourselin, Sebastien
AU - Shapey, Jonathan
AU - Vercauteren, Tom
N1 - Funding Information:
We would like to thank all the other team members that helped during the challenge: Sewon Kim, Yohan Jun, Taejoon Eo, Dosik Hwang (Samoyed); Fei Yu, Jie Zhao, Bin Dong (PKU_BIALAB); Can Cui, Dingjie Su, Andrew Mcneil (MIP); Xi Yang, Kaizhu Huang, Jie Sun (PremiLab); Yingyu Yang, Aurelien Maillot, Marta Nunez-Garcia, Maxime Sermesant (Epione-Liryc); Dewei Hu, Qibang Zhu, Kathleen E Larson, Huahong Zhang (MedICL); Mingming Gong (DBMI_pitt); Ran Gu, Shuwei Zhai, Wenhui Lei (Hi-Lib); Richard Osuala, Carlos Martın-Isla, Victor M. Campello, Carla Sendra-Balcells, Karim Lekadir (smriti161096); Mikhail Belyaev (IRA). This work was supported by the Engineering and Physical Sciences Research Council (EPSRC) [NS/A000049/1, NS/A000050/1], MRC (MC/PC/180520) and Wellcome Trust [203145Z/16/Z, 203148/Z/16/Z, WT106882]. TV is supported by a Medtronic/Royal Academy of Engineering Research Chair [RCSRF1819/7/34]. Z.S and K.Y. are supported by the National Natural Science Foundation of China [No. 61876155], the Jiangsu Science and Technology Programme (Natural Science Foundation of Jiangsu Province) [No. BE2020006-4] and the Key Program Special Fund in Xi'an Jiaotong-Liverpool University (XJTLU) [KSF-E-37]. C.K. and M.H. are supported by the Federal Ministry of Education and Research [No. 031L0202B]. H.S. and H.G.K. are supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT [2019R1A2B5B01070488, 2021R1A4A1031437], Brain Research Program through the NRF funded by the Ministry of Science, ICT & Future Planning [2018M3C7A1024734], Y-BASE R&E Institute a Brain Korea 21, Yonsei University, and the Artificial Intelligence Graduate School Program, Yonsei University [No. 2020-0-01361]. H.Liu, Y.F. and B.D. are supported by the National Institute of Health (NIH) [R01 DC014462]. L.H. and M.H. are supported by the German Research Foundation (DFG) under grant number 320997906 [HE 7364/2-1]. S.J. and S.E. are supported by the Spanish project PID2019-105093GB-I00 and by ICREA under the ICREA Academia programme B.L. and V.K. are supported by the French Government, through the National Research Agency (ANR) 3IA Côte d'Azur [ANR-19-P3IA-0002], IHU Liryc [ANR- 10-IAHU-04]. The Epione-Liryc team is grateful to the OPAL infrastructure from Université Côte d'Azur for providing resources and support. H.Li and I.O are supported by the National Institute of Health (NIH) [R01-NS094456]. L.Z. and H.D. are supported by the Natural Science Foundation of China (NSFC) under Grants 81801778, 12090022, 11831002. Y.X. and K.B. are supported by NIH Award Number 1R01HL141813-01, NSF 1839332 Tripod+X, SAP SE, and Pennsylvania's Department of Health and are grateful for the computational resources provided by Pittsburgh Super Computing grant number TG-ASC170024. S.B. is supported by the National Cancer Institute (NCI) and the National Institute of Neurological Disorders and Stroke (NINDS) of the National Institutes of Health (NIH), under award numbers NCI:U01CA242871 and NINDS:R01NS042645. The content of this publication is solely the responsibility of the authors and does not represent the official views of the NIH.
Funding Information:
This work was supported by the Engineering and Physical Sciences Research Council (EPSRC) [ NS/A000049/1 , NS/A000050/1 ], MRC ( MC/PC/180520 ) and Wellcome Trust [ 203145Z/16/Z , 203148/Z/16/Z , WT106882 ]. TV is supported by a Medtronic/Royal Academy of Engineering Research Chair [ RCSRF1819/7/34 ]. Z.S and K.Y. are supported by the National Natural Science Foundation of China [No. 61876155 ], the Jiangsu Science and Technology Programme (Natural Science Foundation of Jiangsu Province) [No. BE2020006-4 ] and the Key Program Special Fund in Xi’an Jiaotong-Liverpool University (XJTLU) [ KSF-E-37 ]. C.K. and M.H. are supported by the Federal Ministry of Education and Research [No. 031L0202B ]. H.S. and H.G.K. are supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT [ 2019R1A2B5B01070488 , 2021R1A4A1031437 ], Brain Research Program through the NRF funded by the Ministry of Science, ICT & Future Planning [ 2018M3C7A1024734 ], Y-BASE R&E Institute a Brain Korea 21, Yonsei University , and the Artificial Intelligence Graduate School Program, Yonsei University [No. 2020-0-01361 ]. H.Liu, Y.F. and B.D. are supported by the National Institute of Health (NIH) [ R01 DC014462 ]. L.H. and M.H. are supported by the German Research Foundation (DFG) under grant number 320997906 [HE 7364/2-1]. S.J. and S.E. are supported by the Spanish project PID2019-105093GB-I00 and by ICREA under the ICREA Academia programme B.L. and V.K. are supported by the French Government, through the National Research Agency (ANR) 3IA Côte d’Azur [ ANR-19-P3IA-0002 ], IHU Liryc [ ANR- 10-IAHU-04 ]. The Epione-Liryc team is grateful to the OPAL infrastructure from Université Côte d’Azur for providing resources and support. H.Li and I.O are supported by the National Institute of Health (NIH) [ R01-NS094456 ]. L.Z. and H.D. are supported by the Natural Science Foundation of China (NSFC) under Grants 81801778 , 12090022 , 11831002 . Y.X. and K.B. are supported by NIH Award Number 1R01HL141813-01 , NSF 1839332 Tripod+X, SAP SE, and Pennsylvania’s Department of Health and are grateful for the computational resources provided by Pittsburgh Super Computing grant number TG-ASC170024 . S.B. is supported by the National Cancer Institute (NCI) and the National Institute of Neurological Disorders and Stroke (NINDS) of the National Institutes of Health (NIH) , under award numbers NCI:U01CA242871 and NINDS:R01NS042645 . The content of this publication is solely the responsibility of the authors and does not represent the official views of the NIH.
Publisher Copyright:
© 2022
PY - 2023/1/1
Y1 - 2023/1/1
N2 - Domain Adaptation (DA) has recently been of strong interest in the medical imaging community. While a large variety of DA techniques have been proposed for image segmentation, most of these techniques have been validated either on private datasets or on small publicly available datasets. Moreover, these datasets mostly addressed single-class problems. To tackle these limitations, the Cross-Modality Domain Adaptation (crossMoDA) challenge was organised in conjunction with the 24th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2021). CrossMoDA is the first large and multi-class benchmark for unsupervised cross-modality Domain Adaptation. The goal of the challenge is to segment two key brain structures involved in the follow-up and treatment planning of vestibular schwannoma (VS): the VS and the cochleas. Currently, the diagnosis and surveillance in patients with VS are commonly performed using contrast-enhanced T1 (ceT1) MR imaging. However, there is growing interest in using non-contrast imaging sequences such as high-resolution T2 (hrT2) imaging. For this reason, we established an unsupervised cross-modality segmentation benchmark. The training dataset provides annotated ceT1 scans (N=105) and unpaired non-annotated hrT2 scans (N=105). The aim was to automatically perform unilateral VS and bilateral cochlea segmentation on hrT2 scans as provided in the testing set (N=137). This problem is particularly challenging given the large intensity distribution gap across the modalities and the small volume of the structures. A total of 55 teams from 16 countries submitted predictions to the validation leaderboard. Among them, 16 teams from 9 different countries submitted their algorithm for the evaluation phase. The level of performance reached by the top-performing teams is strikingly high (best median Dice score — VS: 88.4%; Cochleas: 85.7%) and close to full supervision (median Dice score — VS: 92.5%; Cochleas: 87.7%). All top-performing methods made use of an image-to-image translation approach to transform the source-domain images into pseudo-target-domain images. A segmentation network was then trained using these generated images and the manual annotations provided for the source image.
AB - Domain Adaptation (DA) has recently been of strong interest in the medical imaging community. While a large variety of DA techniques have been proposed for image segmentation, most of these techniques have been validated either on private datasets or on small publicly available datasets. Moreover, these datasets mostly addressed single-class problems. To tackle these limitations, the Cross-Modality Domain Adaptation (crossMoDA) challenge was organised in conjunction with the 24th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2021). CrossMoDA is the first large and multi-class benchmark for unsupervised cross-modality Domain Adaptation. The goal of the challenge is to segment two key brain structures involved in the follow-up and treatment planning of vestibular schwannoma (VS): the VS and the cochleas. Currently, the diagnosis and surveillance in patients with VS are commonly performed using contrast-enhanced T1 (ceT1) MR imaging. However, there is growing interest in using non-contrast imaging sequences such as high-resolution T2 (hrT2) imaging. For this reason, we established an unsupervised cross-modality segmentation benchmark. The training dataset provides annotated ceT1 scans (N=105) and unpaired non-annotated hrT2 scans (N=105). The aim was to automatically perform unilateral VS and bilateral cochlea segmentation on hrT2 scans as provided in the testing set (N=137). This problem is particularly challenging given the large intensity distribution gap across the modalities and the small volume of the structures. A total of 55 teams from 16 countries submitted predictions to the validation leaderboard. Among them, 16 teams from 9 different countries submitted their algorithm for the evaluation phase. The level of performance reached by the top-performing teams is strikingly high (best median Dice score — VS: 88.4%; Cochleas: 85.7%) and close to full supervision (median Dice score — VS: 92.5%; Cochleas: 87.7%). All top-performing methods made use of an image-to-image translation approach to transform the source-domain images into pseudo-target-domain images. A segmentation network was then trained using these generated images and the manual annotations provided for the source image.
KW - Domain adaptation
KW - Segmentation
KW - Vestibular schwannoma
UR - http://www.scopus.com/inward/record.url?scp=85140305038&partnerID=8YFLogxK
U2 - 10.1016/j.media.2022.102628
DO - 10.1016/j.media.2022.102628
M3 - Article
SN - 1361-8415
VL - 83
JO - Medical Image Analysis
JF - Medical Image Analysis
M1 - 102628
ER -