Replies: 3 comments 1 reply
-
@tnq177, apologies for the delay. I will take a closer look this week. |
Beta Was this translation helpful? Give feedback.
-
Thus, you can pass any of the torch lr scheduler objects or any object that implements the expected API to |
Beta Was this translation helpful? Give feedback.
-
@tjruwase sorry I missed this part in the docs. So apparently I can explicitly manage lr by setting Update: I tested, turn out I can just manually adjust learning rate like usual with Pytorch :) Thanks @tjruwase |
Beta Was this translation helpful? Give feedback.
-
Hi, my code structure is simple like this:
This is working well for me so far using 8 GPUs on a single machine. Right now, I'm trying to do 2 things:
Using Deepspeed's learning rate scheduler
If using deepspeed's learning rate scheduler, I wonder how I can manually change the learning rate please? I understand with pytorch, we can simply go through each in
param_groups
and modify thelr
but since for deepspeed, the learning rate scheduler is wrapped withininitialize
, I'm a bit hesitant to do so.Custom learning rate scheduler
Right now I'm using the Warmup + Decay LR. However, for decay I think it's only going down linearly. I want to try exponential decay as well. And potentially in the future I'd like to experiment with other custom learning rate schedulers. How do I implement a custom one please?
Thanks,
T
Beta Was this translation helpful? Give feedback.
All reactions