Difficulties directly using the pretrained models ¶
By: Anamana on Aug. 5, 2022, 12:23 p.m.
Hi,
I want to experiment with postprocessing and tuning, and would like to save me the time and resources of training by using the pretrained models, like recommended here: https://grand-challenge.org/forums/forum/pi-cai-607/topic/data-preparation-and-training-resources-844/. Ideally, I would like to be able to use the validation splits again for quick validation during experimentation (so that I can e.g. evaluate changes with the model from fold 0 on the validation set of fold 0).
In order to do this I have prepared the data for the semi-supervised nnUNet using the picai_prep code and planned and preprocessed it. Then I copypasted + renamed the pretrained models into their respective folds, copied the plans into the designated folder and reran the validation per fold. Evaluating 'validation_raw_postprocessed' , further postprocessed with extract_lesion_candidates, with the evaluation code from the nnUNet baseline instructions, the results I got are:
AUROCs: [.72, .71, .71, .68, .78] APs: [.17. .22, .12, .14, .24]
On the leaderboard it says: AUROC: 0.820 and AP: 0.608. As my validation results are much worse, I wonder if something went wrong (I also tried similar steps for fold 0 of nnDetection and got a AUROC of .77, which similarly seems far removed from the 0.885 on the leaderboard).
I assume this gap between the ensemble on the hidden set and the separate models on their validation sets isn't reasonable? And, if not, do you happen to see a mistake in what I did, or would you recommend a different way to be able to use the pretrained models while being able to evaluate them easily?
Many thanks