fix reserved label len

Former-commit-id: b06d6c05a1911f329252a7572240048e456affdc
This commit is contained in:
hiyouga
2024-02-04 17:54:26 +08:00
parent 5589d0296a
commit e134013e04
3 changed files with 33 additions and 13 deletions

View File

@@ -21,10 +21,10 @@ class DataArguments:
default="train", metadata={"help": "Which dataset split to use for training and evaluation."}
)
cutoff_len: Optional[int] = field(
default=1024, metadata={"help": "The maximum length of the model inputs after tokenization."}
default=1024, metadata={"help": "The cutoff length of the model inputs after tokenization."}
)
reserved_label_len: Optional[int] = field(
default=1, metadata={"help": "The maximum length reserved for label after tokenization."}
default=1, metadata={"help": "The minimum cutoff length reserved for label after tokenization."}
)
train_on_prompt: Optional[bool] = field(
default=False, metadata={"help": "Whether to disable the mask on the prompt or not."}
@@ -57,7 +57,7 @@ class DataArguments:
ignore_pad_token_for_loss: Optional[bool] = field(
default=True,
metadata={
"help": "Whether to ignore the tokens corresponding to padded labels in the loss computation or not."
"help": "Whether or not to ignore the tokens corresponding to padded labels in the loss computation."
},
)
val_size: Optional[float] = field(