Shortcuts

ModelHooks

class lightning.pytorch.core.hooks.ModelHooks[소스]

기반 클래스: object

Hooks to be used in LightningModule.

configure_sharded_model()[소스]

Hook to create modules in a distributed aware context. This is useful for when using sharded plugins, where we’d like to shard the model instantly, which is useful for extremely large models which can save memory and initialization time.

This hook is called during each of fit/val/test/predict stages in the same process, so ensure that implementation of this hook is idempotent.

반환 형식

None

on_after_backward()[소스]

Called after loss.backward() and before optimizers are stepped.

참고

If using native AMP, the gradients will not be unscaled at this point. Use the on_before_optimizer_step if you need the unscaled gradients.

반환 형식

None

on_before_backward(loss)[소스]

Called before loss.backward().

매개변수

loss (Tensor) – Loss divided by number of batches for gradient accumulation and scaled if using AMP.

반환 형식

None

on_before_optimizer_step(optimizer)[소스]

Called before optimizer.step().

If using gradient accumulation, the hook is called once the gradients have been accumulated. See: accumulate_grad_batches.

If using AMP, the loss will be unscaled before calling this hook. See these docs for more information on the scaling of gradients.

If clipping gradients, the gradients will not have been clipped yet.

매개변수

optimizer (Optimizer) – Current optimizer being used.

Example:

def on_before_optimizer_step(self, optimizer):
    # example to inspect gradient information in tensorboard
    if self.trainer.global_step % 25 == 0:  # don't make the tf file huge
        for k, v in self.named_parameters():
            self.logger.experiment.add_histogram(
                tag=k, values=v.grad, global_step=self.trainer.global_step
            )
반환 형식

None

on_before_zero_grad(optimizer)[소스]

Called after training_step() and before optimizer.zero_grad().

Called in the training loop after taking an optimizer step and before zeroing grads. Good place to inspect weight information with weights updated.

This is where it is called:

for optimizer in optimizers:
    out = training_step(...)

    model.on_before_zero_grad(optimizer) # < ---- called here
    optimizer.zero_grad()

    backward()
매개변수

optimizer (Optimizer) – The optimizer for which grads should be zeroed.

반환 형식

None

on_fit_end()[소스]

Called at the very end of fit.

If on DDP it is called on every process

반환 형식

None

on_fit_start()[소스]

Called at the very beginning of fit.

If on DDP it is called on every process

반환 형식

None

on_predict_batch_end(outputs, batch, batch_idx, dataloader_idx=0)[소스]

Called in the predict loop after the batch.

매개변수
  • outputs (Optional[Any]) – The outputs of predict_step(x)

  • batch (Any) – The batched data as it is returned by the prediction DataLoader.

  • batch_idx (int) – the index of the batch

  • dataloader_idx (int) – the index of the dataloader

반환 형식

None

on_predict_batch_start(batch, batch_idx, dataloader_idx=0)[소스]

Called in the predict loop before anything happens for that batch.

매개변수
  • batch (Any) – The batched data as it is returned by the test DataLoader.

  • batch_idx (int) – the index of the batch

  • dataloader_idx (int) – the index of the dataloader

반환 형식

None

on_predict_end()[소스]

Called at the end of predicting.

반환 형식

None

on_predict_epoch_end()[소스]

Called at the end of predicting.

반환 형식

None

on_predict_epoch_start()[소스]

Called at the beginning of predicting.

반환 형식

None

on_predict_model_eval()[소스]

Sets the model to eval during the predict loop.

반환 형식

None

on_predict_start()[소스]

Called at the beginning of predicting.

반환 형식

None

on_test_batch_end(outputs, batch, batch_idx, dataloader_idx=0)[소스]

Called in the test loop after the batch.

매개변수
  • outputs (Union[Tensor, Dict[str, Any], None]) – The outputs of test_step(x)

  • batch (Any) – The batched data as it is returned by the test DataLoader.

  • batch_idx (int) – the index of the batch

  • dataloader_idx (int) – the index of the dataloader

반환 형식

None

on_test_batch_start(batch, batch_idx, dataloader_idx=0)[소스]

Called in the test loop before anything happens for that batch.

매개변수
  • batch (Any) – The batched data as it is returned by the test DataLoader.

  • batch_idx (int) – the index of the batch

  • dataloader_idx (int) – the index of the dataloader

반환 형식

None

on_test_end()[소스]

Called at the end of testing.

반환 형식

None

on_test_epoch_end()[소스]

Called in the test loop at the very end of the epoch.

반환 형식

None

on_test_epoch_start()[소스]

Called in the test loop at the very beginning of the epoch.

반환 형식

None

on_test_model_eval()[소스]

Sets the model to eval during the test loop.

반환 형식

None

on_test_model_train()[소스]

Sets the model to train during the test loop.

반환 형식

None

on_test_start()[소스]

Called at the beginning of testing.

반환 형식

None

on_train_batch_end(outputs, batch, batch_idx)[소스]

Called in the training loop after the batch.

매개변수
  • outputs (Union[Tensor, Dict[str, Any]]) – The outputs of training_step(x)

  • batch (Any) – The batched data as it is returned by the training DataLoader.

  • batch_idx (int) – the index of the batch

반환 형식

None

on_train_batch_start(batch, batch_idx)[소스]

Called in the training loop before anything happens for that batch.

If you return -1 here, you will skip training for the rest of the current epoch.

매개변수
  • batch (Any) – The batched data as it is returned by the training DataLoader.

  • batch_idx (int) – the index of the batch

반환 형식

Optional[int]

on_train_end()[소스]

Called at the end of training before logger experiment is closed.

반환 형식

None

on_train_epoch_end()[소스]

Called in the training loop at the very end of the epoch.

To access all batch outputs at the end of the epoch, you can cache step outputs as an attribute of the LightningModule and access them in this hook:

class MyLightningModule(L.LightningModule):
    def __init__(self):
        super().__init__()
        self.training_step_outputs = []

    def training_step(self):
        loss = ...
        self.training_step_outputs.append(loss)
        return loss

    def on_train_epoch_end(self):
        # do something with all training_step outputs, for example:
        epoch_mean = torch.stack(self.training_step_outputs).mean()
        self.log("training_epoch_mean", epoch_mean)
        # free up the memory
        self.training_step_outputs.clear()
반환 형식

None

on_train_epoch_start()[소스]

Called in the training loop at the very beginning of the epoch.

반환 형식

None

on_train_start()[소스]

Called at the beginning of training after sanity check.

반환 형식

None

on_validation_batch_end(outputs, batch, batch_idx, dataloader_idx=0)[소스]

Called in the validation loop after the batch.

매개변수
  • outputs (Union[Tensor, Dict[str, Any], None]) – The outputs of validation_step(x)

  • batch (Any) – The batched data as it is returned by the validation DataLoader.

  • batch_idx (int) – the index of the batch

  • dataloader_idx (int) – the index of the dataloader

반환 형식

None

on_validation_batch_start(batch, batch_idx, dataloader_idx=0)[소스]

Called in the validation loop before anything happens for that batch.

매개변수
  • batch (Any) – The batched data as it is returned by the validation DataLoader.

  • batch_idx (int) – the index of the batch

  • dataloader_idx (int) – the index of the dataloader

반환 형식

None

on_validation_end()[소스]

Called at the end of validation.

반환 형식

None

on_validation_epoch_end()[소스]

Called in the validation loop at the very end of the epoch.

반환 형식

None

on_validation_epoch_start()[소스]

Called in the validation loop at the very beginning of the epoch.

반환 형식

None

on_validation_model_eval()[소스]

Sets the model to eval during the val loop.

반환 형식

None

on_validation_model_train()[소스]

Sets the model to train during the val loop.

반환 형식

None

on_validation_start()[소스]

Called at the beginning of validation.

반환 형식

None