diff --git a/examples/extras/badam/llama3_lora_sft.yaml b/examples/extras/badam/llama3_lora_sft.yaml index 242e63ab1..a78de2fa6 100644 --- a/examples/extras/badam/llama3_lora_sft.yaml +++ b/examples/extras/badam/llama3_lora_sft.yaml @@ -37,5 +37,5 @@ pure_bf16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/fsdp_qlora/llama3_lora_sft.yaml b/examples/extras/fsdp_qlora/llama3_lora_sft.yaml index 920d8fdb1..348459b86 100644 --- a/examples/extras/fsdp_qlora/llama3_lora_sft.yaml +++ b/examples/extras/fsdp_qlora/llama3_lora_sft.yaml @@ -38,5 +38,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/galore/llama3_full_sft.yaml b/examples/extras/galore/llama3_full_sft.yaml index 3db31fed0..605545de9 100644 --- a/examples/extras/galore/llama3_full_sft.yaml +++ b/examples/extras/galore/llama3_full_sft.yaml @@ -38,5 +38,5 @@ pure_bf16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/llama_pro/llama3_freeze_sft.yaml b/examples/extras/llama_pro/llama3_freeze_sft.yaml index 214f411ae..444a11133 100644 --- a/examples/extras/llama_pro/llama3_freeze_sft.yaml +++ b/examples/extras/llama_pro/llama3_freeze_sft.yaml @@ -36,5 +36,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/loraplus/llama3_lora_sft.yaml b/examples/extras/loraplus/llama3_lora_sft.yaml index 9936bcd30..960f613e3 100644 --- a/examples/extras/loraplus/llama3_lora_sft.yaml +++ b/examples/extras/loraplus/llama3_lora_sft.yaml @@ -35,5 +35,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/extras/mod/llama3_full_sft.yaml b/examples/extras/mod/llama3_full_sft.yaml index edfec44e0..df03c1e02 100644 --- a/examples/extras/mod/llama3_full_sft.yaml +++ b/examples/extras/mod/llama3_full_sft.yaml @@ -35,5 +35,5 @@ pure_bf16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/full_multi_gpu/llama3_full_sft.yaml b/examples/full_multi_gpu/llama3_full_sft.yaml index b8873e3ac..40b62f24f 100644 --- a/examples/full_multi_gpu/llama3_full_sft.yaml +++ b/examples/full_multi_gpu/llama3_full_sft.yaml @@ -37,5 +37,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_multi_gpu/llama3_lora_sft.yaml b/examples/lora_multi_gpu/llama3_lora_sft.yaml index 5e5dd9e6e..9be3c780c 100644 --- a/examples/lora_multi_gpu/llama3_lora_sft.yaml +++ b/examples/lora_multi_gpu/llama3_lora_sft.yaml @@ -37,5 +37,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml b/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml index e8dee2163..411522434 100644 --- a/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml +++ b/examples/lora_multi_gpu/llama3_lora_sft_ds.yaml @@ -38,5 +38,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_multi_npu/llama3_lora_sft_ds.yaml b/examples/lora_multi_npu/llama3_lora_sft_ds.yaml index 825b84503..1ed24d040 100644 --- a/examples/lora_multi_npu/llama3_lora_sft_ds.yaml +++ b/examples/lora_multi_npu/llama3_lora_sft_ds.yaml @@ -38,5 +38,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_dpo.yaml b/examples/lora_single_gpu/llama3_lora_dpo.yaml index 62752e573..158c9e048 100644 --- a/examples/lora_single_gpu/llama3_lora_dpo.yaml +++ b/examples/lora_single_gpu/llama3_lora_dpo.yaml @@ -36,5 +36,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_kto.yaml b/examples/lora_single_gpu/llama3_lora_kto.yaml index 6f6898185..ead221e9f 100644 --- a/examples/lora_single_gpu/llama3_lora_kto.yaml +++ b/examples/lora_single_gpu/llama3_lora_kto.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_pretrain.yaml b/examples/lora_single_gpu/llama3_lora_pretrain.yaml index 54c5d89ae..9167a893a 100644 --- a/examples/lora_single_gpu/llama3_lora_pretrain.yaml +++ b/examples/lora_single_gpu/llama3_lora_pretrain.yaml @@ -33,5 +33,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_reward.yaml b/examples/lora_single_gpu/llama3_lora_reward.yaml index c82f9414c..916630573 100644 --- a/examples/lora_single_gpu/llama3_lora_reward.yaml +++ b/examples/lora_single_gpu/llama3_lora_reward.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llama3_lora_sft.yaml b/examples/lora_single_gpu/llama3_lora_sft.yaml index 429cb6afd..cc93d05a3 100644 --- a/examples/lora_single_gpu/llama3_lora_sft.yaml +++ b/examples/lora_single_gpu/llama3_lora_sft.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/lora_single_gpu/llava1_5_lora_sft.yaml b/examples/lora_single_gpu/llava1_5_lora_sft.yaml index acab48848..95c1d40d1 100644 --- a/examples/lora_single_gpu/llava1_5_lora_sft.yaml +++ b/examples/lora_single_gpu/llava1_5_lora_sft.yaml @@ -35,5 +35,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml b/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml index 53cc12e28..23301de51 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_aqlm.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml b/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml index 1a92f8228..40a290a39 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_awq.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml b/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml index c7f72c66d..6652d8cfd 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_bitsandbytes.yaml @@ -35,5 +35,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml b/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml index 45caf17cc..323ea7c67 100644 --- a/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml +++ b/examples/qlora_single_gpu/llama3_lora_sft_gptq.yaml @@ -34,5 +34,5 @@ fp16: true ### eval val_size: 0.1 per_device_eval_batch_size: 1 -evaluation_strategy: steps +eval_strategy: steps eval_steps: 500 diff --git a/src/llamafactory/extras/env.py b/src/llamafactory/extras/env.py index 27453a6b8..059730f1e 100644 --- a/src/llamafactory/extras/env.py +++ b/src/llamafactory/extras/env.py @@ -51,4 +51,4 @@ def print_env() -> None: info["vLLM version"] = vllm.__version__ - print("\n".join(["- {}: {}".format(key, value) for key, value in info.items()]) + "\n") + print("\n" + "\n".join(["- {}: {}".format(key, value) for key, value in info.items()]) + "\n") diff --git a/src/llamafactory/webui/runner.py b/src/llamafactory/webui/runner.py index 6e1facef0..d35fd9031 100644 --- a/src/llamafactory/webui/runner.py +++ b/src/llamafactory/webui/runner.py @@ -200,7 +200,7 @@ class Runner: # eval config if get("train.val_size") > 1e-6 and args["stage"] != "ppo": args["val_size"] = get("train.val_size") - args["evaluation_strategy"] = "steps" + args["eval_strategy"] = "steps" args["eval_steps"] = args["save_steps"] args["per_device_eval_batch_size"] = args["per_device_train_batch_size"]