-
Notifications
You must be signed in to change notification settings - Fork 1
Open
Description
Hi @ldeecke,
Thanks for sharing the code.
I am trying to use your mode normalization in Wideresnet with depth 28 and widen factor 10. Training on CIFAR datasets with batch size 128 and batch norm, it takes 5983M memory. However, when I replace the batch norm with your mode norm, it is out of memory in a 16G V100 GPU. Could you give some analysis on why it consumes so much memory in training? Any suggestions to save memory?
Thanks
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels