lora modules: all by default

Former-commit-id: 52c4ae87c7f4312704c31ef26b079b2c5b95ea5f
This commit is contained in:
hiyouga
2024-06-06 03:53:28 +08:00
parent abc2a73a33
commit 937f49ec3d
23 changed files with 78 additions and 118 deletions

View File

@@ -6,7 +6,7 @@ quantization_bit: 4
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### ddp
ddp_timeout: 180000000

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
loraplus_lr_ratio: 16.0
### dataset

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### ddp
ddp_timeout: 180000000

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### ddp
ddp_timeout: 180000000

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### ddp
ddp_timeout: 180000000

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: dpo
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
pref_beta: 0.1
pref_loss: sigmoid # [sigmoid (dpo), orpo, simpo]

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: kto
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: kto_en_demo

View File

@@ -6,7 +6,7 @@ reward_model: saves/llama3-8b/lora/reward
stage: ppo
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: pt
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: c4_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: rm
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: dpo_en_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo

View File

@@ -6,7 +6,7 @@ visual_inputs: true
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: mllm_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: ISTA-DASLab/Meta-Llama-3-8B-Instruct-AQLM-2Bit-1x16
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: TechxGenus/Meta-Llama-3-8B-Instruct-AWQ
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo

View File

@@ -6,7 +6,7 @@ quantization_bit: 4
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo

View File

@@ -5,7 +5,7 @@ model_name_or_path: TechxGenus/Meta-Llama-3-8B-Instruct-GPTQ
stage: sft
do_train: true
finetuning_type: lora
lora_target: q_proj,v_proj
lora_target: all
### dataset
dataset: identity,alpaca_en_demo