diff --git a/OAI/types/common.py b/OAI/types/common.py index 482877d..bc1e341 100644 --- a/OAI/types/common.py +++ b/OAI/types/common.py @@ -19,13 +19,13 @@ class CommonCompletionRequest(BaseModel): model: Optional[str] = None # Extra OAI request stuff - best_of: Optional[int] = None - echo: Optional[bool] = False - logit_bias: Optional[Dict[str, float]] = None - logprobs: Optional[int] = None - n: Optional[int] = 1 - suffix: Optional[str] = None - user: Optional[str] = None + best_of: Optional[int] = Field(description = "Not parsed. Only used for OAI compliance.", default = None) + echo: Optional[bool] = Field(description = "Not parsed. Only used for OAI compliance.", default = False) + logit_bias: Optional[Dict[str, float]] = Field(description = "Not parsed. Only used for OAI compliance.", default = None) + logprobs: Optional[int] = Field(description = "Not parsed. Only used for OAI compliance.", default = None) + n: Optional[int] = Field(description = "Not parsed. Only used for OAI compliance.", default = 1) + suffix: Optional[str] = Field(description = "Not parsed. Only used for OAI compliance.", default = None) + user: Optional[str] = Field(description = "Not parsed. Only used for OAI compliance.", default = None) # Generation info # seed: Optional[int] = -1 @@ -36,7 +36,7 @@ class CommonCompletionRequest(BaseModel): max_tokens: Optional[int] = 150 # Aliased to repetition_penalty - frequency_penalty: Optional[float] = 0.0 + frequency_penalty: Optional[float] = Field(description = "Aliased to Repetition Penalty", default = 0.0) # Sampling params token_healing: Optional[bool] = False diff --git a/OAI/types/model.py b/OAI/types/model.py index f23a4c6..08715ad 100644 --- a/OAI/types/model.py +++ b/OAI/types/model.py @@ -27,7 +27,7 @@ class ModelList(BaseModel): class DraftModelLoadRequest(BaseModel): draft_model_name: str draft_rope_scale: Optional[float] = 1.0 - draft_rope_alpha: Optional[float] = None + draft_rope_alpha: Optional[float] = Field(description = "Automatically calculated if not present", default = None) # TODO: Unify this with ModelCardParams class ModelLoadRequest(BaseModel): @@ -36,7 +36,7 @@ class ModelLoadRequest(BaseModel): gpu_split_auto: Optional[bool] = True gpu_split: Optional[List[float]] = Field(default_factory=list) rope_scale: Optional[float] = 1.0 - rope_alpha: Optional[float] = None + rope_alpha: Optional[float] = Field(description = "Automatically calculated if not present", default = None) no_flash_attention: Optional[bool] = False # low_mem: Optional[bool] = False cache_mode: Optional[str] = "FP16"