[parser] support omegaconf (#7793)

This commit is contained in:
hoshi-hiyouga
2025-04-21 23:30:30 +08:00
committed by GitHub
parent bd7bc31c79
commit 416853dd25
25 changed files with 62 additions and 94 deletions

View File

@@ -6,5 +6,5 @@ trust_remote_code: true
### export
export_dir: output/llama3_full_sft
export_size: 5
export_device: cpu
export_device: cpu # choices: [cpu, auto]
export_legacy_format: false

View File

@@ -6,7 +6,7 @@ trust_remote_code: true
### export
export_dir: output/llama3_gptq
export_quantization_bit: 4
export_quantization_dataset: data/c4_demo.json
export_quantization_dataset: data/c4_demo.jsonl
export_size: 5
export_device: cpu
export_device: cpu # choices: [cpu, auto]
export_legacy_format: false

View File

@@ -9,5 +9,5 @@ trust_remote_code: true
### export
export_dir: output/llama3_lora_sft
export_size: 5
export_device: cpu
export_device: cpu # choices: [cpu, auto]
export_legacy_format: false

View File

@@ -1,7 +1,7 @@
### Note: DO NOT use quantized model or quantization_bit when merging lora adapters
### model
model_name_or_path: Qwen/Qwen2-VL-7B-Instruct
model_name_or_path: Qwen/Qwen2.5-VL-7B-Instruct
adapter_name_or_path: saves/qwen2_vl-7b/lora/sft
template: qwen2_vl
trust_remote_code: true
@@ -9,5 +9,5 @@ trust_remote_code: true
### export
export_dir: output/qwen2_vl_lora_sft
export_size: 5
export_device: cpu
export_device: cpu # choices: [cpu, auto]
export_legacy_format: false