add default template

Former-commit-id: c64fb6b83fdbedd62073417213f0215207ff1311
This commit is contained in:
hiyouga
2023-06-16 21:12:17 +08:00
parent 7dc1f06a97
commit de2c418637
3 changed files with 19 additions and 6 deletions

View File

@@ -284,7 +284,7 @@ def prepare_args(
if training_args.do_train and (not training_args.fp16):
logger.warning("We recommend enable fp16 mixed precision training.")
if data_args.prompt_template == "alpaca":
if data_args.prompt_template == "default":
logger.warning("Please specify `prompt_template` if you are using other pre-trained models.")
if training_args.local_rank != -1 and training_args.ddp_find_unused_parameters is None:
@@ -326,7 +326,7 @@ def prepare_infer_args() -> Tuple[ModelArguments, DataTrainingArguments, Finetun
if model_args.quantization_bit is not None and finetuning_args.finetuning_type != "lora":
raise ValueError("Quantization is only compatible with the LoRA method.")
if data_args.prompt_template == "alpaca":
if data_args.prompt_template == "default":
logger.warning("Please specify `prompt_template` if you are using other pre-trained models.")
return model_args, data_args, finetuning_args, generating_args