mirror of
https://github.com/vllm-project/vllm.git
synced 2025-10-20 14:53:52 +08:00
[Bugfix] Use "vision_model" prefix for MllamaVisionModel (#9628)
Signed-off-by: mgoin <michael@neuralmagic.com>
This commit is contained in:
@ -1053,7 +1053,8 @@ class MllamaForConditionalGeneration(nn.Module, SupportsMultiModal):
|
||||
self.image_size = config.vision_config.image_size
|
||||
|
||||
self.vision_model = MllamaVisionModel(config.vision_config,
|
||||
quant_config)
|
||||
quant_config,
|
||||
prefix="vision_model")
|
||||
self.language_model = MllamaForCausalLM(
|
||||
config.text_config,
|
||||
cache_config=cache_config,
|
||||
|
Reference in New Issue
Block a user