update hardware requirements

Former-commit-id: 604b3d10fc1448f702943114b66b97bded21e080
This commit is contained in:
hiyouga
2024-03-09 03:58:18 +08:00
parent 8ed1463236
commit 5d7d8bd55c
4 changed files with 29 additions and 35 deletions

View File

@@ -9,8 +9,8 @@ class FreezeArguments:
Arguments pertaining to the freeze (partial-parameter) training.
"""
name_module_trainable: Optional[str] = field(
default=None,
name_module_trainable: str = field(
default="all",
metadata={
"help": """Name of trainable modules for partial-parameter (freeze) fine-tuning. \
Use commas to separate multiple modules. \
@@ -23,7 +23,7 @@ class FreezeArguments:
},
)
num_layer_trainable: int = field(
default=3,
default=2,
metadata={"help": "The number of trainable layers for partial-parameter (freeze) fine-tuning."},
)
@@ -52,8 +52,8 @@ class LoraArguments:
default=8,
metadata={"help": "The intrinsic dimension for LoRA fine-tuning."},
)
lora_target: Optional[str] = field(
default=None,
lora_target: str = field(
default="all",
metadata={
"help": """Name(s) of target modules to apply LoRA. \
Use commas to separate multiple modules. \

View File

@@ -137,16 +137,6 @@ def get_train_args(args: Optional[Dict[str, Any]] = None) -> _TRAIN_CLS:
if training_args.do_train and training_args.predict_with_generate:
raise ValueError("`predict_with_generate` cannot be set as True while training.")
if (
training_args.do_train
and finetuning_args.finetuning_type == "freeze"
and finetuning_args.name_module_trainable is None
):
raise ValueError("Please specify `name_module_trainable` in Freeze training.")
if training_args.do_train and finetuning_args.finetuning_type == "lora" and finetuning_args.lora_target is None:
raise ValueError("Please specify `lora_target` in LoRA training.")
if training_args.do_train and model_args.use_unsloth and not is_unsloth_available:
raise ValueError("Unsloth was not installed: https://github.com/unslothai/unsloth")