generated from fastai/nbdev_template
-
Notifications
You must be signed in to change notification settings - Fork 2.3k
Closed
Labels
⚡ PEFTRelated to PEFTRelated to PEFT❓ questionSeeking clarification or more informationSeeking clarification or more information
Description
Greetings,
Quick question about our LoRA setup: are we initializing a separate reference model during training (e.g., for KL/anchor comparisons), or are we reusing the frozen base model as the reference since the base is frozen under PEFT?
If a separate reference is currently spun up, is there a way to configure the pipeline to reuse the base model as the reference to save memory/compute? Pointers to the relevant config flags or code path would be much appreciated.
Thank you very much.
Metadata
Metadata
Assignees
Labels
⚡ PEFTRelated to PEFTRelated to PEFT❓ questionSeeking clarification or more informationSeeking clarification or more information