Skip to content

Conversation

@jpdunc23
Copy link
Member

@jpdunc23 jpdunc23 commented Feb 10, 2026

Continuation of the separation of training-specific concerns from inference stepper configs, building on #809.

This PR makes backwards-incompatible changes affecting fme.coupled training and fine-tuning configs:

  • Training (init from separate uncoupled component checkpoints): Each component's parameter_init: ParameterInitializationConfig is now configured using its respective ComponentTrainingConfig on the CoupledTrainStepperConfig.
  • Fine-tuning (init from a coupled stepper checkpoint): Moved parameter_init: CoupledParameterInitConfig out of CoupledStepperConfig and into CoupledTrainStepperConfig.

Existing fme.ace training YAML configs will continue to work without changes, with parameter_init now transferred to TrainStepperConfig via StepperConfig.get_train_stepper_config(). This backwards compatibility will be removed in a future PR.

Changes:

  • CoupledTrainStepperConfig now owns both CoupledParameterInitConfig and the per-component ParameterInitializationConfig (via ComponentTrainingConfig).

  • StepperConfig.get_stepper() signature changed to accept an optional ParameterInitializer instead of a boolean flag

  • TrainStepperConfig now owns parameter_init and builds both the initializer and the underlying Stepper

  • Tests added

@jpdunc23 jpdunc23 changed the base branch from main to refactor/coupled-train-stepper February 10, 2026 20:52
@jpdunc23 jpdunc23 mentioned this pull request Feb 10, 2026
1 task
Base automatically changed from refactor/coupled-train-stepper to main February 11, 2026 07:56
@jpdunc23 jpdunc23 marked this pull request as ready for review February 12, 2026 20:36
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant