LmEvalEvaluatorConfig¶
Module: fast_llm.engine.evaluation.config
Variant of: EvaluatorConfig — select with type: lm_eval
Inherits from: EvaluatorConfig, IntervalConfig
Fields¶
interval—feature-
Type:
intorNoneDefault:NoneThe number of training iterations between each interval. Setting to None will disable.
offset—feature-
Type:
intDefault:0Offset for the first interval.
add_bos_token-
Type:
boolDefault:FalseWhether to prepend a beginning-of-sequence (BOS) token, required for some models like LLaMA; passed to the Fast-LLM lm_eval model wrapper.
cli_args-
Type: list[
str] Default:<lambda>()lm_eval CLI arguments, excluding those related to model, wandb, batch sizes, and device.
communication_timeout_sec-
Type:
floatDefault:600.0Maximum wait time (in seconds) for tensor-parallel or data-parallel model operations such as forward, generate, or gathering data. Needed because some ranks may have no data or post-processing can be slow, exceeding the default 60s timeout.
logits_cache-
Type:
boolDefault:TrueWhether to enable logits caching for speedup and avoiding recomputation during repeated evaluations; passed to the Fast-LLM lm_eval model wrapper.
max_length-
Type:
intorNoneDefault:NoneMaximum sequence length including both prompt and newly generated tokens. If not set, it is inferred from the Fast-LLM model config or tokenizer.
prefix_token_id-
Type:
intorNoneDefault:NoneToken ID to use as a prefix to the input (e.g., for control codes or prompts); passed to the Fast-LLM lm_eval model wrapper.
tokenizer-
Type: TokenizerConfig Default: (sub-fields optional)
Configuration for the tokenizer.
truncation-
Type:
boolDefault:FalseWhether to use truncation during tokenization (useful when inputs exceed model's max length); passed to the Fast-LLM lm_eval model wrapper.