[Hotfix][VLM] Fixing max position embeddings for Pixtral (#8399)

This commit is contained in:
Roger Wang
2024-09-12 09:28:37 -07:00
committed by GitHub
parent 7de49aa86c
commit 520ca380ae

View File

@ -206,6 +206,8 @@ def load_params_config(model, revision) -> PretrainedConfig:
config_dict["tie_word_embeddings"] = config_dict.get(
"tie_embeddings", False)
config_dict["max_seq_len"] = config_dict.get("max_seq_len", 128_000)
config_dict["max_position_embeddings"] = config_dict.get(
"max_position_embeddings", 128_000)
if config_dict.get("moe") is not None:
config_dict["architectures"] = ["MixtralForCausalLM"]