[Question] Understanding a poor training on the ADAM dataset
See original GitHub issueHi! Awesome work 😃
Recently we have trained the nnDetection on the ADAM challenge, i.e., Task019FG_ADAM. However, the predictions on the test set are pretty bad - a lot of false postives and general sensitivity (approaching) 0. We are trying to understand where it went wrong, maybe you could be of help.
-
In your case, did the network generate a low resolution model for the ADAM challenge? Our network did end up generating a low resolution model, which we did not specifically use further on.
-
Do you have any suggestions on what could be different with your run?
The input data was unchanged apart from the omission of one patient due to having a T1 image, and we did not deviate from the instruction steps. We trained all five folds and performed a sweep for all. After that we ran the consolidation and prediction arguments as instructed.
Thank you for your help!
Best, Aaron
Environment Information
Currently using an NVIDIA GeForce RTX 2080 Ti; PyTorch 1.8.0; CUDA 11.2.
nnDetection was installed from [~docker~ | source].
PyTorch Version: <module 'torch.version' from '/opt/conda/lib/python3.8/site-packages/torch/version.py'>
PyTorch CUDA: 11.2
PyTorch Backend cudnn: 8100
PyTorch CUDA Arch List: ['sm_52', 'sm_60', 'sm_61', 'sm_70', 'sm_75', 'sm_80', 'sm_86', 'compute_86']
PyTorch Current Device Capability: (7, 5)
PyTorch CUDA available: True
System NVCC: nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2020 NVIDIA Corporation
Built on Mon_Nov_30_19:08:53_PST_2020
Cuda compilation tools, release 11.2, V11.2.67
Build cuda_11.2.r11.2/compiler.29373293_0
System Arch List: 7.5
System OMP_NUM_THREADS: 1
System CUDA_HOME is None: False
Python Version: 3.8.5 (default, Sep 4 2020, 07:30:14)
[GCC 7.3.0]
Issue Analytics
- State:
- Created 2 years ago
- Comments:14 (7 by maintainers)

Top Related StackOverflow Question
Since we ran your first setup in our experiments (both our MICCAI and nnDetection submission) I don’t think the setup will change the overall results drastically (nnDetection metrics will be worse since treated aneurysms are pretty easy to detect but I wouldn’t expect huge differences in the performance of untreated aneurysms).
From your original run (random split, treated and untreated aneurysms as foreground):
[model dir]/consolidated/val_results/results_boxes.jsonspecifically the values forAP_IoU_0.10_MaxDet_100orFROC_score_IoU_0.10are of interest. If the file/folder does not exist in consolidated, please runnndet_evalwith the fold set to-1(make sure the task ids are correct though and there is no mixup with the new run)[model dir]/consolidated/val_results/pred_hist_IoU@0_1.pngcontains a probability histogram. The title of the figure shows how many “pos” instances there are. This value corresponds to the total number of instances in the data set and thus is great to check for debugging. (the plot of my run shows 156)Quick update: Following the ADAM evaluation method, the best model checkpoint predictions give a sensitivity of
0.63and a FPcount of0.32. I uploaded the loss curves for the five folds over here, could you check with your log files whether they look somewhat similar?Code to plot the curves: