add adam_mini to readme

Former-commit-id: d610c6bcf8a8ba6f4236f5d11f79571b83f4fb11
This commit is contained in:
hiyouga
2024-08-09 20:02:03 +08:00
parent 7e755e9cac
commit 59cbce1a46
12 changed files with 94 additions and 34 deletions

View File

@@ -326,6 +326,10 @@ class FinetuningArguments(FreezeArguments, LoraArguments, RLHFArguments, GaloreA
default=False,
metadata={"help": "Whether or not to make only the parameters in the expanded blocks trainable."},
)
use_adam_mini: bool = field(
default=False,
metadata={"help": "Whether or not to use the Adam-mini optimizer."},
)
freeze_vision_tower: bool = field(
default=True,
metadata={"help": "Whether ot not to freeze vision tower in MLLM training."},
@@ -342,10 +346,6 @@ class FinetuningArguments(FreezeArguments, LoraArguments, RLHFArguments, GaloreA
default=False,
metadata={"help": "Whether or not to save the training loss curves."},
)
use_adammini: bool = field(
default=False,
metadata={"help": "Whether or not to use AdamMini optimizer."},
)
def __post_init__(self):
def split_arg(arg):