diff --git a/backends/exllamav2/model.py b/backends/exllamav2/model.py index bc9142a..a0a7f0e 100644 --- a/backends/exllamav2/model.py +++ b/backends/exllamav2/model.py @@ -476,6 +476,7 @@ class ExllamaV2Container: "prompt_template": self.prompt_template.name if self.prompt_template else None, + "use_vision": self.use_vision, } if self.draft_config: diff --git a/endpoints/core/types/model.py b/endpoints/core/types/model.py index 17fa0a7..ddf1cc2 100644 --- a/endpoints/core/types/model.py +++ b/endpoints/core/types/model.py @@ -21,6 +21,7 @@ class ModelCardParameters(BaseModel): chunk_size: Optional[int] = 2048 prompt_template: Optional[str] = None num_experts_per_token: Optional[int] = None + use_vision: Optional[bool] = False # Draft is another model, so include it in the card params draft: Optional["ModelCard"] = None