[version] support transformers 449 (#6982)

* support transformers 449

* fix mm plugin

Former-commit-id: e9118a9df0839d24f6ddff5a0b55ef101a1d3d22
This commit is contained in:
hoshi-hiyouga
2025-02-18 17:05:40 +08:00
committed by GitHub
parent be33ef67fb
commit 1d675a287d
7 changed files with 16 additions and 25 deletions

View File

@@ -187,8 +187,6 @@ class MultiModalDataCollatorForSeq2Seq(DataCollatorForSeq2Seq):
mm_inputs["cross_attention_mask"] = F.pad(cross_attention_mask, (0, 0, 0, 0, 0, seq_len - orig_len))
features.update(mm_inputs)
if isinstance(features.get("pixel_values"), list): # for pixtral inputs
features = features.data # use default_collate() instead of BatchEncoding.to()
if "image_bound" in features: # for minicpmv inputs
bsz, seq_length = features["input_ids"].shape