SLIDE 9 Stage1: attention-guided training
Attention Maps Classification Results Input 2.5D Image Label
Down
Classification Loss (categorical cross entropy) Attention Loss (dice coefficient loss)
+ =
Total Loss
PE or not
384×384×5 24×24 2
- Resample volumetric images (bilinear interpolation): slice thickness [0.5mm, 5mm] → 2.5mm
- 10,388 slabs (5 slices) of annotated pairs from 1,670 positive volumetric images
- Same amount of negative slabs randomly sampled from 593 negative volumetric images
- Image cropped to center 384×384, [-1024HU,500HU] → [0, 255]
- 80% training, 20% validation
- Training epochs: 100 (save the model with the highest val. acc.)
Annotation Mask
ResNet18
3x3 conv, 64 3x3 conv, 64 3x3 conv, 64 3x3 conv, 64 3x3 conv, 64 3x3 conv, 128, /2 3x3 conv, 128 3x3 conv, 128 3x3 conv, 128 3x3 conv, 256, /2 3x3 conv, 256 3x3 conv, 256 3x3 conv, 256 3x3 conv, 512, /2 3x3 conv, 512 3x3 conv, 512 3x3 conv, 512
Avg pool fc 2