GradientScalerConfig¶
Module: fast_llm.engine.optimizer.config
Fields¶
constant—feature-
Type:
floatorNoneDefault:NoneConstant multiplier applied to the loss. Setting this disables dynamic scaling.
hysteresis—feature-
Type:
intDefault:2Number of failed updates to tolerate before lowering the learning rate in dynamic scaling (fp16).
initial—feature-
Type:
floatDefault:65536Initial loss scale for dynamic scaling (fp16).
minimum—feature-
Type:
floatDefault:1.0Minimum loss scale for dynamic scaling (fp16).
window—feature-
Type:
intDefault:1000Interval between dynamic scaling growth (fp16).