[trainer] fix vlm loss for transformers 4.49 (#7448)

This commit is contained in:
hoshi-hiyouga
2025-03-24 10:24:05 +08:00
committed by GitHub
parent 3612946dd9
commit 7203365b80
5 changed files with 21 additions and 4 deletions

View File

@@ -59,6 +59,9 @@ class CustomSeq2SeqTrainer(Seq2SeqTrainer):
self.processing_class: PreTrainedTokenizer = kwargs.get("tokenizer")
super().__init__(**kwargs)
if processor is not None:
self.model_accepts_loss_kwargs = False
self.finetuning_args = finetuning_args
if gen_kwargs is not None:
# https://github.com/huggingface/transformers/blob/v4.45.0/src/transformers/trainer_seq2seq.py#L287
@@ -93,6 +96,10 @@ class CustomSeq2SeqTrainer(Seq2SeqTrainer):
return super()._get_train_sampler()
@override
def compute_loss(self, model, inputs, *args, **kwargs):
return super().compute_loss(model, inputs, *args, **kwargs)
@override
def prediction_step(
self,