You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Load forked adapter weights on first training call
After _experimental_fork_checkpoint, store the checkpoint path on
the service. On the first _train_dedicated/_train_shared call, load
the adapter weights via load_lora_adapter before training begins.
This is needed because create_unsloth_train_context may initialize
the LoRA architecture from adapter_config.json without loading the
actual trained weights from adapter_model.safetensors, especially
when the checkpoint was trained at a different precision than the
current load config.
0 commit comments