Shortcuts

strategy_adapters

Fine-Tuning Scheduler Strategy Adapters

class finetuning_scheduler.strategy_adapters.FSDPStrategyAdapter(awp_overrides=None, *args, **kwargs)[source]

Bases: finetuning_scheduler.strategy_adapters.base.StrategyAdapter

A StrategyAdapter that extends FinetuningScheduler (FTS) to support flexible, multi-phase, scheduled fine-tuning with the Fully Sharded Data Parallel (FSDP) strategy (FSDPStrategy).

As with standard FSDP usage, FSDP wrapping of a LightningModule can be performed either by providing an auto_wrap_policy or (for maximal control) by overriding the configure_sharded_model method of LightningModule and manually wrapping the module.

In order to support multi-phase scheduled fine-tuning with FSDP, FTS’s key precondition is that the defined fine-tuning schedule phases have disjoint sets of FSDP-flattened parameters (i.e. FlatParameter s, which are created when wrapping a set of modules in a FSDP instance/unit). This constraint is derived from the fact that the requires_grad attribute currently must be the same for all parameters flattened into the same FlatParameter.

To facilitate module wrapping in alignment with fine-tuning schedule phases, FTS provides the awp_overrides feature which allows users to provide module name-based complements to a given auto_wrap_policy. See the Example: Multi-Phase Scheduled Fine-Tuning with FSDP tutorial for a concrete example and additional guidance.

FTS will attempt to validate that the module is wrapped in a manner that aligns with the defined fine-tuning schedule phases prior to the start of training and provided detailed feedback for the user if a misalignment is discovered.

Warning

FSDPStrategyAdapter is in BETA and subject to change. The interface can bring breaking changes and new features with the next release of PyTorch.

Note

The no_decay attribute that FTS supports on LightningModule with the base StrategyAdapter is not currently supported in the context of FSDP fine-tuning.

Tip

Because of inter-module dependencies (among other reasons), wrapping every submodule in its own separate FSDP instance is often not a viable approach to ensuring fine-tuning schedule/module wrapping alignment. Starting with a provided auto_wrap_policy (e.g. transformer_auto_wrap_policy) and providing module name-based complements as needed using awp_overrides is often the most expedient approach to auto-wrapping in alignment with a fine-tuning schedule. As always, if needed, one can override configure_sharded_model and manually wrap a given LightningModule to align with a desired fine-tuning schedule.

The only user-facing configuration for FSDPStrategyAdapter is awp_overrides, an optional list of module names that should be wrapped in separate FSDP instances, complementing the modules that would be individually wrapped by auto_wrap_policy provided in the FSDPStrategy strategy configuration.

Parameters

awp_overrides (Optional[List]) – A list of module names to wrap in separate FSDP instances (i.e., auto_wrap_policy overrides). Only applicable when complementing/overriding an auto_wrap_policy provided in the FSDPStrategy strategy configuration. Override lists will be ignored when manually wrapping modules via a configure_sharded_model method. If the named modules cannot be found, an exception will be thrown. Defaults to None.

awp_overrides

A list of module names to wrap in separate FSDP instances.

fsdp_param_transform(orig_thaw_pl, inspect_only)[source]

The parameter transformation function currently used by fts_optim_transform() to transform original parameter lists for optimizer operations.

Parameters
  • orig_thaw_pl (List) – The original parameter name list before FSDP’s transformation of them.

  • inspect_only (bool) – Whether to use the specified transform in read-only (i.e. inspect_only) mode, avoiding any persistent state transformation that may accompany normal usage. Typically useful for state inspection and validation contexts.

Returns

A transformed parameter name list that matches the current Optimizer’s view of them after FSDP’s transformation of the original parameter names.

Return type

List

fts_optim_transform(orig_pl, inspect_only=False)[source]

Because FSDP performs parameter transformations that cause the current.

Optimizer’s view of parameter names to diverge from the original parameter names, this parameter transformation is required for optimizer operations.

Parameters
  • orig_pl (List) – The original parameter name list before FSDP’s transformation of them.

  • inspect_only (bool) – Whether to use the specified transform in read-only (i.e. inspect_only) mode, avoiding any persistent state transformation that may accompany normal usage. Typically useful for state inspection and validation contexts.

Returns

A transformed parameter name list that matches the current Optimizer’s view of them after FSDP’s transformation of the original parameter names.

Return type

List

load_optimizer_state_dict(checkpoint_connector)[source]

Override the default load_optimizer_state_dict method so that we can allow FSDP to manage the movement of restored optimizer states to the relevant devices.

Parameters

checkpoint_connector (_CheckpointConnector) – The _CheckpointConnector associated with the current training session.

Return type

None

logical_param_translation(param_names)[source]

Effectively the reverse transformation of fts_optim_transform().

Parameters

param_names (List) – A parameter name list from the current Optimizer’s view of them after FSDP’s transformation of the original parameter names.

Returns

The original parameter name list before a given FSDP’s transformation.

Return type

List

on_after_init_fts()[source]

To accommodate FSDP, we defer executing the first fine-tuning phase that would otherwise be executed in this hook, which fires in FinetuningScheduler setup immediately after init_fts()

Return type

None

on_before_fts_fit_start()[source]

In this hook executed immediately before the FinetuningScheduler on_fit_start() hook begins, we ensure the provided fine-tuning schedule and FSDP wrapped LightningModule are appropriately aligned and valid. If the fine-tuning schedule and wrapped module are detected to be incompatible, detailed feedback is provided to the user (which is why multiple checks are aggregated before returning any alignment exceptions).

Raises

MisconfigurationException – If any FTS FSDP fine-tuning schedule/module wrapping alignment exceptions are thrown. The provided exceptions provide detailed feedback for the user to address the misalignment.

Return type

None

on_before_init_fts()[source]

In this hook executed immediately before init_fts(), to accommodate FSDP we:

  1. Disable Lightning’s restoration of the optimizer to allow us to implement special handling

  2. Prune no_decay specification since it is not currently supported in the context of FSDP fine-tuning

  3. Validate the awp_overrides configuration

  4. Configure FTS wrapping of the provided LightningModule to either use the provided LightningModule.configure_sharded_model method (if present) or a provided auto_wrap_policy.

Return type

None

on_before_restore_optimizers_and_lrs()[source]

Allow the FSDPStrategyAdapter to override the default load_optimizer_state_dict method.

This is necessary so we can allow FSDP to manage the movement of restored optimizer states to the relevant devices.

Return type

None

property lightning_restore_optimizer: bool

Disable Lightning’s restoration of the optimizer to allow FTS to implement special handling.

Returns

Returns False to allow FTS control over optimizer restoration.

Return type

bool

class finetuning_scheduler.strategy_adapters.StrategyAdapter[source]

Bases: object

Base class for all strategy adapters. Implements the default FinetuningScheduler hooks. Can be subclassed to extend FinetuningScheduler support for a complex or custom Strategy via an associated StrategyAdapter.

Warning

StrategyAdapter is in BETA and subject to change. The interface can bring breaking changes and new features with the next release of FTS.

Tip

If you want to extend FTS to use a custom, currently unsupported strategy or override current FTS behavior in the context of a given training strategy, subclassing StrategyAdapter is a way to do so. See FSDPStrategyAdapter for an example implementation.

The default fine-tuning phase execution function is set on StrategyAdapter initialization. This can be overridden by StrategyAdapter subclasses to adapt fine-tuning phase execution to meet strategy-specific requirements.

static base_ft_phase(module, thaw_pl, translation_func=None, init_thaw=False)[source]

Thaw/unfreeze the provided list of parameters in the provided Module

Parameters
  • module (Module) – The Module that will have parameters selectively unfrozen/thawed.

  • thaw_pl (List) – The list of parameters that should be thawed/unfrozen in the Module

  • init_thaw (bool) – If True, modifies message to user accordingly. Defaults to False.

Returns

A Tuple of two lists.
  1. The list of newly thawed/unfrozen parameters thawed by this function

  2. A list of all currently thawed/unfrozen parameters in the target Module

Return type

Tuple[List, List]

connect(fts_parent)[source]

Create a handle for the associated FinetuningScheduler instance.

Parameters

fts_parent (Callback) – The associated FinetuningScheduler instance

Return type

None

fts_optim_transform(orig_pl, inspect_only=False)[source]

A method that can be overridden by a StrategyAdapter if a Strategy performs parameter transformations that cause the current Optimizer’s view of parameter names to diverge from the original parameter names. By default, no transformation of schedule parameter names is required for optimizer operations.

Parameters
  • orig_pl (List) – The original parameter name list before a given Strategy’s transformation of them.

  • inspect_only (bool) – Whether to use the specified transform in read-only (i.e. inspect_only) mode, avoiding any persistent state transformation that may accompany normal usage. Typically useful for state inspection and validation contexts.

Returns

A transformed parameter name list that matches the current

Optimizer’s view of them after a given Strategy’s transformation of the original parameter names.

Return type

List

logical_param_translation(param_names)[source]

Effectively the reverse transformation of fts_optim_transform(). Can be overridden by a StrategyAdapter if a Strategy performs parameter transformations that cause the original user view of parameter names to diverge from the current Optimizer’s view. By default, no transformation of Optimizer parameter names is required.

Parameters

param_names (List) – A parameter name list from the current Optimizer’s view of them after a Strategy’s transformation of the original parameter names.

Returns

The original parameter name list before a given

Strategy’s transformation.

Return type

List

on_after_init_fts()[source]

Hook executed in FinetuningScheduler setup immediately after init_fts().

Return type

None

on_before_fts_fit_start()[source]

Hook executed immediately before the FinetuningScheduler on_fit_start() hook begins.

Return type

None

on_before_init_fts()[source]

Hook executed in FinetuningScheduler setup immediately before init_fts()

Return type

None

on_before_restore_optimizers_and_lrs()[source]

Hook executed immediately before FinetuningScheduler restores optimizers and schedulers.

Return type

None

phase0_optimizer_override()[source]

Reconfigure the user-configured optimizer (configured via configure_optimizers) to optimize the parameters (and only those parameters) scheduled to be optimized in phase 0 of the current fine-tuning schedule.

Reconfiguration only takes place here if FTS discovers the set of parameters to be initially thawed and present in the optimizer differs from the parameters specified in phase 0. Only the parameters included in the optimizer are affected; the choice of optimizer, lr_scheduler etc. remains unaltered.

Return type

None

property pl_module: lightning.pytorch.core.module.LightningModule

Convenient access to the LightningModule being fine- tuned.

Returns

The user’s LightningModule

Return type

LightningModule

property pls_handle: lightning.pytorch.strategies.strategy.Strategy

Convenient access to the current Strategy in use.

Returns

The Strategy in use.

Return type

Strategy

Read the Docs v: v2.0.0
Versions
latest
stable
v2.0.0
v0.4.1
v0.4.0
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.3
v0.2.2
v0.2.1
v0.2.0
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
Downloads
On Read the Docs
Project Home
Builds

Free document hosting provided by Read the Docs.