add moe aux loss control #3085

Former-commit-id: c9187ebc944e2de454ace3304b7d28eabb1b1a81
This commit is contained in:
hiyouga
2024-04-02 14:26:31 +08:00
parent 03e20bb5c6
commit 117b67ea30
4 changed files with 23 additions and 16 deletions

View File

@@ -73,6 +73,10 @@ class ModelArguments:
default=False,
metadata={"help": "Whether or not to use unsloth's optimization for the LoRA training."},
)
moe_aux_loss_coef: Optional[float] = field(
default=None,
metadata={"help": "Coefficient of the auxiliary router loss in mixture-of-experts model."},
)
disable_gradient_checkpointing: bool = field(
default=False,
metadata={"help": "Whether or not to disable gradient checkpointing."},