update examples

Former-commit-id: 779603055ae9216ff549f5285caac8c0c0a1e9fb
This commit is contained in:
hiyouga
2024-05-13 20:39:36 +08:00
parent 5f5f948806
commit e4972c8fc4
22 changed files with 36 additions and 37 deletions

View File

@@ -13,7 +13,6 @@ dataset: orca_rlhf
template: llama3
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
@@ -34,6 +33,7 @@ warmup_steps: 0.1
fp16: true
# eval
val_size: 0.1
per_device_eval_batch_size: 1
evaluation_strategy: steps
eval_steps: 500

View File

@@ -12,7 +12,6 @@ dataset: orca_rlhf
template: llama3
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
@@ -33,6 +32,7 @@ warmup_steps: 0.1
fp16: true
# eval
val_size: 0.1
per_device_eval_batch_size: 1
evaluation_strategy: steps
eval_steps: 500

View File

@@ -11,7 +11,6 @@ lora_target: q_proj,v_proj
dataset: c4_demo
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
@@ -32,6 +31,7 @@ warmup_steps: 0.1
fp16: true
# eval
val_size: 0.1
per_device_eval_batch_size: 1
evaluation_strategy: steps
eval_steps: 500

View File

@@ -12,7 +12,6 @@ dataset: orca_rlhf
template: llama3
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
@@ -33,6 +32,7 @@ warmup_steps: 0.1
fp16: true
# eval
val_size: 0.1
per_device_eval_batch_size: 1
evaluation_strategy: steps
eval_steps: 500

View File

@@ -12,7 +12,6 @@ dataset: identity,alpaca_gpt4_en
template: llama3
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
@@ -33,6 +32,7 @@ warmup_steps: 0.1
fp16: true
# eval
val_size: 0.1
per_device_eval_batch_size: 1
evaluation_strategy: steps
eval_steps: 500

View File

@@ -12,7 +12,6 @@ dataset: identity,alpaca_gpt4_en
template: llama3
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
tokenized_path: saves/llama3-8b/dataset/sft

View File

@@ -13,7 +13,6 @@ dataset: mllm_demo
template: vicuna
cutoff_len: 1024
max_samples: 1000
val_size: 0.1
overwrite_cache: true
preprocessing_num_workers: 16
@@ -34,6 +33,7 @@ warmup_steps: 0.1
fp16: true
# eval
val_size: 0.1
per_device_eval_batch_size: 1
evaluation_strategy: steps
eval_steps: 500