mirror of
https://github.com/hiyouga/LlamaFactory.git
synced 2026-03-19 23:33:09 +00:00
[deps] adapt to transformers v5 (#10147)
Co-authored-by: frozenleaves <frozen@Mac.local> Co-authored-by: hiyouga <hiyouga@buaa.edu.cn>
This commit is contained in:
@@ -103,7 +103,9 @@ class FixValueHeadModelCallback(TrainerCallback):
|
||||
if args.should_save:
|
||||
output_dir = os.path.join(args.output_dir, f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}")
|
||||
fix_valuehead_checkpoint(
|
||||
model=kwargs.pop("model"), output_dir=output_dir, safe_serialization=args.save_safetensors
|
||||
model=kwargs.pop("model"),
|
||||
output_dir=output_dir,
|
||||
safe_serialization=getattr(args, "save_safetensors", True),
|
||||
)
|
||||
|
||||
|
||||
@@ -137,7 +139,7 @@ class PissaConvertCallback(TrainerCallback):
|
||||
if isinstance(model, PeftModel):
|
||||
init_lora_weights = getattr(model.peft_config["default"], "init_lora_weights")
|
||||
setattr(model.peft_config["default"], "init_lora_weights", True)
|
||||
model.save_pretrained(pissa_init_dir, safe_serialization=args.save_safetensors)
|
||||
model.save_pretrained(pissa_init_dir, safe_serialization=getattr(args, "save_safetensors", True))
|
||||
setattr(model.peft_config["default"], "init_lora_weights", init_lora_weights)
|
||||
|
||||
@override
|
||||
@@ -155,11 +157,11 @@ class PissaConvertCallback(TrainerCallback):
|
||||
if isinstance(model, PeftModel):
|
||||
init_lora_weights = getattr(model.peft_config["default"], "init_lora_weights")
|
||||
setattr(model.peft_config["default"], "init_lora_weights", True)
|
||||
model.save_pretrained(pissa_backup_dir, safe_serialization=args.save_safetensors)
|
||||
model.save_pretrained(pissa_backup_dir, safe_serialization=getattr(args, "save_safetensors", True))
|
||||
setattr(model.peft_config["default"], "init_lora_weights", init_lora_weights)
|
||||
model.save_pretrained(
|
||||
pissa_convert_dir,
|
||||
safe_serialization=args.save_safetensors,
|
||||
safe_serialization=getattr(args, "save_safetensors", True),
|
||||
path_initial_model_for_weight_conversion=pissa_init_dir,
|
||||
)
|
||||
model.load_adapter(pissa_backup_dir, "default", is_trainable=True)
|
||||
|
||||
@@ -72,7 +72,7 @@ def run_ppo(
|
||||
ppo_trainer.ppo_train(resume_from_checkpoint=training_args.resume_from_checkpoint)
|
||||
ppo_trainer.save_model()
|
||||
if training_args.should_save:
|
||||
fix_valuehead_checkpoint(model, training_args.output_dir, training_args.save_safetensors)
|
||||
fix_valuehead_checkpoint(model, training_args.output_dir, getattr(training_args, "save_safetensors", True))
|
||||
|
||||
ppo_trainer.save_state() # must be called after save_model to have a folder
|
||||
if ppo_trainer.is_world_process_zero() and finetuning_args.plot_loss:
|
||||
|
||||
@@ -114,7 +114,7 @@ class PairwiseTrainer(Trainer):
|
||||
if state_dict is None:
|
||||
state_dict = self.model.state_dict()
|
||||
|
||||
if self.args.save_safetensors:
|
||||
if getattr(self.args, "save_safetensors", True):
|
||||
from collections import defaultdict
|
||||
|
||||
ptrs = defaultdict(list)
|
||||
|
||||
@@ -65,7 +65,7 @@ def run_rm(
|
||||
train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint)
|
||||
trainer.save_model()
|
||||
if training_args.should_save:
|
||||
fix_valuehead_checkpoint(model, training_args.output_dir, training_args.save_safetensors)
|
||||
fix_valuehead_checkpoint(model, training_args.output_dir, getattr(training_args, "save_safetensors", True))
|
||||
|
||||
trainer.log_metrics("train", train_result.metrics)
|
||||
trainer.save_metrics("train", train_result.metrics)
|
||||
|
||||
Reference in New Issue
Block a user