[misc] move wechat out (#9223)
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
# FP8 training example with FSDP
|
||||
# FP8 training example with FSDP
|
||||
# This config demonstrates FP8 mixed precision training using HuggingFace Accelerate
|
||||
# with FSDP for distributed training and float8 all-gather optimization
|
||||
|
||||
@@ -48,4 +48,4 @@ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
|
||||
|
||||
### Logging configuration
|
||||
report_to: wandb
|
||||
run_name: llama3_fp8_fsdp_sft
|
||||
run_name: llama3_fp8_fsdp_sft
|
||||
|
||||
Reference in New Issue
Block a user