Skip to content

GatedRMSNormalizationConfig

Module: fast_llm.layers.common.normalization.config

Variant of: NormalizationConfig — select with type: gated_rms_norm

Inherits from: RMSNormalizationConfig, LayerNormalizationBaseConfig, NormalizationConfig

Fields

activationcore

Type: ActivationType    Default: "silu"

The MLP intermediate activation type. Default: SiLU for gated MLP, GeLU otherwise.

epsilonarchitecture

Type: float    Default: 1e-05

Regularizer for the division.

weightarchitecture

Type: ParameterConfig    Default: (sub-fields optional)

Configuration for the weight.

zero_centeredarchitecture

Type: bool    Default: False

Write the normalization weight as w = 1 + w', to improve numerical accuracy when close to one.

lr_scalefeature

Type: float or None    Default: None

Scaling factor for the layer learning rate. Combines multiplicatively with the scale set by the parent layer and individual parameters, if applicable.

implementationperformance

Type: NormalizationImplementation    Default: "auto"

The implementation to use for the normalization layer.