mindspore.nn.cosine_decay_lr
- mindspore.nn.cosine_decay_lr(min_lr, max_lr, total_step, step_per_epoch, decay_epoch)[source]
Calculate learning rate base on cosine decay function.
For the i-th step, the formula of computing decayed_learning_rate[i] is:
Where
.- Parameters
- Returns
list[float]. The size of list is total_step.
Examples
>>> min_lr = 0.01 >>> max_lr = 0.1 >>> total_step = 6 >>> step_per_epoch = 2 >>> decay_epoch = 2 >>> output = cosine_decay_lr(min_lr, max_lr, total_step, step_per_epoch, decay_epoch) >>> print(output) [0.1, 0.1, 0.05500000000000001, 0.05500000000000001, 0.01, 0.01]