OAI: Add fasttensors to model load endpoint

Also fix logging when loading prompt templates.

Signed-off-by: kingbri <bdashore3@proton.me>
This commit is contained in:
kingbri
2024-01-25 01:01:29 -05:00
parent fc4570220c
commit 751627e571
2 changed files with 2 additions and 1 deletions

View File

@@ -90,6 +90,7 @@ class ModelLoadRequest(BaseModel):
prompt_template: Optional[str] = None
num_experts_per_token: Optional[int] = None
use_cfg: Optional[bool] = None
fasttensors: Optional[bool] = False
draft: Optional[DraftModelLoadRequest] = None