From 1f444c8fb77bf20447dbfa17b4fddca4121ca3a1 Mon Sep 17 00:00:00 2001 From: kingbri Date: Wed, 15 Nov 2023 01:00:08 -0500 Subject: [PATCH] Requirements: Add fastchat and override pydantic Use an older version of pydantic to stay compatible Signed-off-by: kingbri --- main.py | 16 ++++++++-------- requirements.txt | Bin 78 -> 122 bytes 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/main.py b/main.py index 73bf0de..cb8cefd 100644 --- a/main.py +++ b/main.py @@ -25,7 +25,7 @@ async def list_models(): model_config = config["model"] models = get_model_list(pathlib.Path(model_config["model_dir"] or "models")) - return models.model_dump_json() + return models.json() @app.get("/v1/model", dependencies=[Depends(check_api_key)]) async def get_current_model(): @@ -33,7 +33,7 @@ async def get_current_model(): return HTTPException(400, "No models are loaded.") model_card = ModelCard(id=model_container.get_model_path().name) - return model_card.model_dump_json() + return model_card.json() @app.post("/v1/model/load", dependencies=[Depends(check_admin_key)]) async def load_model(data: ModelLoadRequest): @@ -61,13 +61,13 @@ async def load_model(data: ModelLoadRequest): module=module, modules=modules, status="processing" - ).model_dump_json() + ).json() yield ModelLoadResponse( module=module, modules=modules, status="finished" - ).model_dump_json() + ).json() return EventSourceResponse(generator()) @@ -86,14 +86,14 @@ async def encode_tokens(data: TokenEncodeRequest): tokens = model_container.get_tokens(data.text, None, **data.get_params())[0].tolist() response = TokenEncodeResponse(tokens=tokens, length=len(tokens)) - return response.model_dump_json() + return response.json() @app.post("/v1/token/decode", dependencies=[Depends(check_api_key)]) async def decode_tokens(data: TokenDecodeRequest): message = model_container.get_tokens(None, data.tokens, **data.get_params()) response = TokenDecodeResponse(text=message) - return response.model_dump_json() + return response.json() @app.post("/v1/completions", dependencies=[Depends(check_api_key)]) async def generate_completion(request: Request, data: CompletionRequest): @@ -106,14 +106,14 @@ async def generate_completion(request: Request, data: CompletionRequest): response = create_completion_response(part, index, model_container.get_model_path().name) - yield response.model_dump_json() + yield response.json() return EventSourceResponse(generator()) else: response_text = model_container.generate(**data.to_gen_params()) response = create_completion_response(response_text, 0, model_container.get_model_path().name) - return response.model_dump_json() + return response.json() if __name__ == "__main__": # Initialize auth keys diff --git a/requirements.txt b/requirements.txt index 25c5fc6a0f6500c32c022bd302677600a6898356..7dd575d1ae64974f918f9006ba70bbb7a525ce8d 100644 GIT binary patch delta 66 zcmeaP(*6H0jUkbtn4ttn7BFNoZ~@6mh7=$#4=9$&kj!AiV8o!qV8>v~U^tQAo-Ylk NIGG^>sGtO>7yuz$4fg;5 delta 21 bcmb>O