-
Notifications
You must be signed in to change notification settings - Fork 43
Open
Description
I noticed that when training RDM, we need to set args.cosine_lr=True to initialize the scheduler in engine_rdm.py. However, the instructions given in the readme defaults to args.cosine_lr=False. I am new in deep learning. I wonder is it correct to keep the learning rate of adamw as a constant during training? Why can it still converge? Looking forward to your reply!
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels