@@ -1,4 +1,4 @@
|
||||
# Note: DO NOT use quantized model or quantization_bit when merging lora weights
|
||||
# Note: DO NOT use quantized model or quantization_bit when merging lora adapters
|
||||
|
||||
# model
|
||||
model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct
|
||||
|
||||
Reference in New Issue
Block a user