Transformers documentation

Callbacks

You are viewing main version, which requires installation from source. If you'd like regular pip install, checkout the latest stable version (v5.2.0).
Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Callbacks

TrainerCallback hooks into specific training events (epoch start, evaluation, training end) to modify training state or control flow. Use callbacks to log metrics to experiment trackers like Trackio, customize when saving and evaluation happen, or add other custom behavior. Stack multiple callbacks to combine features.

Callbacks can’t modify the training loop itself, like the forward pass. To change what Trainer computes, subclass its methods instead.

The diagram below shows every event a callback can hook into.

on_train_begin
  └─ for each epoch:
      on_epoch_begin
      └─ for each step:
          on_step_begin
          ├─ for each gradient accumulation substep:
          │   on_substep_end
          on_pre_optimizer_step   ← after gradient clipping, before optimizer.step()
          on_optimizer_step       ← after optimizer.step()
          on_step_end
          └─ (conditionally):
              on_log
              on_evaluate
              on_save
      on_epoch_end
on_train_end
on_predict / on_prediction_step
on_push_begin

Creating a callback

Subclass TrainerCallback and override one or more event methods from the diagram above. The example below demonstrates three hooks:

  • on_epoch_begin prints the current epoch from TrainerState (a live value) and the learning rate from TrainingArguments (a static value).
  • on_step_end reads the current learning rate from the lr_scheduler (passed via **kwargs) and sets should_evaluate on TrainerControl (see TrainerControl for a complete list of control objects) to trigger an evaluation when it drops below the threshold.
  • on_train_end prints the best metric and the step where it occurred.
from transformers import TrainerCallback

class EpochLoggerCallback(TrainerCallback):
    def __init__(self, lr_eval_threshold=1e-5):
        self.lr_eval_threshold = lr_eval_threshold

    def on_epoch_begin(self, args, state, control, **kwargs):
        print(f"Starting epoch {int(state.epoch) + 1}/{state.num_train_epochs} "
              f"(lr={args.learning_rate})")

    def on_step_end(self, args, state, control, **kwargs):
        lr_scheduler = kwargs.get("lr_scheduler")
        if lr_scheduler is not None:
            current_lr = lr_scheduler.get_last_lr()[0]
            if current_lr < self.lr_eval_threshold:
                control.should_evaluate = True

    def on_train_end(self, args, state, control, **kwargs):
        print(f"Training complete! Best metric: {state.best_metric} at step {state.best_global_step}")

Register the callback with Trainer in the callbacks argument. You can also pass multiple callbacks as a list.

trainer = Trainer(
    callbacks=[EpochLoggerCallback(lr_eval_threshold=1e-5)],
    ...,
)

Built-in callbacks

Transformers includes several built-in callbacks that are active by default. Additional integrated callbacks log to platforms like Trackio.

DefaultFlowCallback

DefaultFlowCallback manages the default logging, evaluation, and checkpoint schedule based on the logging_strategy, eval_strategy, and save_strategy values in TrainingArguments. At the right step or epoch, it sets the corresponding control flags (should_log, should_evaluate, should_save). It also sets should_training_stop when global_step reaches max_steps.

Overriding this callback is the main way to customize when logging, evaluation, or saving happens.

ProgressCallback and PrinterCallback

Trainer automatically picks between these two callbacks based on the disable_tqdm field in TrainingArguments.

  • ProgressCallback is used by default. It displays a tqdm progress bar during training and a separate bar during evaluation or prediction, and prints the latest metrics on each on_log event. During distributed training, it only runs on the main process to avoid duplicate output.
  • PrinterCallback is used when disable_tqdm=True. It prints the log dictionary to stdout on every on_log event with no progress bar.

You can also swap them manually with remove_callback() and add_callback().

from transformers import PrinterCallback

trainer = Trainer(...)
trainer.remove_callback(ProgressCallback)
trainer.add_callback(PrinterCallback)

EarlyStoppingCallback

EarlyStoppingCallback stops training when an evaluation metric stops improving. After each evaluation, it checks whether the metric improved by more than early_stopping_threshold. If the metric hasn’t improved for early_stopping_patience consecutive evaluations, training stops.

EarlyStoppingCallback requires two TrainingArguments:

  • metric_for_best_model, the evaluation metric to monitor
  • eval_strategy, whether to evaluate on "steps" or "epoch"
from transformers import EarlyStoppingCallback

trainer = Trainer(
    ...,
    callbacks=[EarlyStoppingCallback(early_stopping_patience=3, early_stopping_threshold=0.001)],
)

Next steps

Update on GitHub