Former-commit-id: 1eb7118db3ad6054cfd59d5f16a5d882e40e9057
This commit is contained in:
fzc8578
2025-01-10 20:55:52 +08:00
parent 75b3819e43
commit 40382f1387
4 changed files with 10 additions and 9 deletions

View File

@@ -100,7 +100,7 @@ def load_tokenizer(model_args: "ModelArguments") -> "TokenizerModule":
processor = AutoProcessor.from_pretrained(model_args.model_name_or_path, **init_kwargs)
patch_processor(processor, config, tokenizer, model_args)
except Exception as e:
logger.info(f"Processor was not found: {e}.")
logger.debug(f"Processor was not found: {e}.")
processor = None
# Avoid load tokenizer, see:

View File

@@ -138,13 +138,12 @@ def patch_model(
add_valuehead: bool,
) -> None:
gen_config = model.generation_config # check and fix generation config
if gen_config is not None:
if not gen_config.do_sample and (
(gen_config.temperature is not None and gen_config.temperature != 1.0)
or (gen_config.top_p is not None and gen_config.top_p != 1.0)
or (gen_config.typical_p is not None and gen_config.typical_p != 1.0)
):
gen_config.do_sample = True
if not gen_config.do_sample and (
(gen_config.temperature is not None and gen_config.temperature != 1.0)
or (gen_config.top_p is not None and gen_config.top_p != 1.0)
or (gen_config.typical_p is not None and gen_config.typical_p != 1.0)
):
gen_config.do_sample = True
if "GenerationMixin" not in str(model.generate.__func__):
model.generate = MethodType(PreTrainedModel.generate, model)