From 8662437a9f7d06b4fc6d22713e472428636f7562 Mon Sep 17 00:00:00 2001 From: Aditya Pratap Singh Date: Sat, 20 Jan 2024 04:17:06 +0530 Subject: [PATCH 1/4] Add workaround for gpt-4-vision-preview model --- backend/apps/openai/main.py | 36 ++++++++++++++++++++++++++++++++---- 1 file changed, 32 insertions(+), 4 deletions(-) diff --git a/backend/apps/openai/main.py b/backend/apps/openai/main.py index ef9330c5..cbf3043a 100644 --- a/backend/apps/openai/main.py +++ b/backend/apps/openai/main.py @@ -84,9 +84,37 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): raise HTTPException(status_code=401, detail=ERROR_MESSAGES.API_KEY_NOT_FOUND) - body = await request.body() - # headers = dict(request.headers) - # print(headers) + # Try to decode the body of the request from bytes to a UTF-8 string (Require add max_token to fix gpt-4-vision) + try: + body_str = (await request.body()).decode('utf-8') + except UnicodeDecodeError as e: + print("Error decoding request body:", e) + raise HTTPException(status_code=400, detail="Invalid request body") + # Check if the body is not empty + if body_str: + try: + + body_dict = json.loads(body_str) + except json.JSONDecodeError as e: + print("Error loading request body into a dictionary:", e) + raise HTTPException(status_code=400, detail="Invalid JSON in request body") + + # Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 10000 + # This is a workaround until OpenAI fixes the issue with this model + if body_dict.get("model") == "gpt-4-vision-preview": + body_dict["max_tokens"] = 10000 + print("Modified body_dict:", body_dict) + + # Try to convert the modified body back to JSON + try: + # Convert the modified body back to JSON + body_json = json.dumps(body_dict) + except TypeError as e: + print("Error converting modified body to JSON:", e) + raise HTTPException(status_code=500, detail="Internal server error") + else: + body_json = body_str # If the body is empty, use it as is + headers = {} headers["Authorization"] = f"Bearer {app.state.OPENAI_API_KEY}" @@ -96,7 +124,7 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): r = requests.request( method=request.method, url=target_url, - data=body, + data=body_json, headers=headers, stream=True, ) From 60afd6ecddf4eae0808f193a6c146eb378ba076b Mon Sep 17 00:00:00 2001 From: Aditya Pratap Singh Date: Sat, 20 Jan 2024 04:34:47 +0530 Subject: [PATCH 2/4] Add workaround for gpt-4-vision-preview model that support 4k tokens --- backend/apps/openai/main.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/backend/apps/openai/main.py b/backend/apps/openai/main.py index cbf3043a..1544949e 100644 --- a/backend/apps/openai/main.py +++ b/backend/apps/openai/main.py @@ -99,10 +99,10 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): print("Error loading request body into a dictionary:", e) raise HTTPException(status_code=400, detail="Invalid JSON in request body") - # Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 10000 + # Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000 # This is a workaround until OpenAI fixes the issue with this model if body_dict.get("model") == "gpt-4-vision-preview": - body_dict["max_tokens"] = 10000 + body_dict["max_tokens"] = 4000 print("Modified body_dict:", body_dict) # Try to convert the modified body back to JSON From b26e0fb7e707e202b42a8e3e5fc5067c26968f00 Mon Sep 17 00:00:00 2001 From: "Timothy J. Baek" Date: Mon, 22 Jan 2024 01:37:54 -0800 Subject: [PATCH 3/4] refac --- backend/apps/openai/main.py | 70 +++++++++++++------------------------ 1 file changed, 25 insertions(+), 45 deletions(-) diff --git a/backend/apps/openai/main.py b/backend/apps/openai/main.py index 1544949e..fd6acee7 100644 --- a/backend/apps/openai/main.py +++ b/backend/apps/openai/main.py @@ -37,19 +37,16 @@ async def get_openai_url(user=Depends(get_current_user)): if user and user.role == "admin": return {"OPENAI_API_BASE_URL": app.state.OPENAI_API_BASE_URL} else: - raise HTTPException(status_code=401, - detail=ERROR_MESSAGES.ACCESS_PROHIBITED) + raise HTTPException(status_code=401, detail=ERROR_MESSAGES.ACCESS_PROHIBITED) @app.post("/url/update") -async def update_openai_url(form_data: UrlUpdateForm, - user=Depends(get_current_user)): +async def update_openai_url(form_data: UrlUpdateForm, user=Depends(get_current_user)): if user and user.role == "admin": app.state.OPENAI_API_BASE_URL = form_data.url return {"OPENAI_API_BASE_URL": app.state.OPENAI_API_BASE_URL} else: - raise HTTPException(status_code=401, - detail=ERROR_MESSAGES.ACCESS_PROHIBITED) + raise HTTPException(status_code=401, detail=ERROR_MESSAGES.ACCESS_PROHIBITED) @app.get("/key") @@ -57,19 +54,16 @@ async def get_openai_key(user=Depends(get_current_user)): if user and user.role == "admin": return {"OPENAI_API_KEY": app.state.OPENAI_API_KEY} else: - raise HTTPException(status_code=401, - detail=ERROR_MESSAGES.ACCESS_PROHIBITED) + raise HTTPException(status_code=401, detail=ERROR_MESSAGES.ACCESS_PROHIBITED) @app.post("/key/update") -async def update_openai_key(form_data: KeyUpdateForm, - user=Depends(get_current_user)): +async def update_openai_key(form_data: KeyUpdateForm, user=Depends(get_current_user)): if user and user.role == "admin": app.state.OPENAI_API_KEY = form_data.key return {"OPENAI_API_KEY": app.state.OPENAI_API_KEY} else: - raise HTTPException(status_code=401, - detail=ERROR_MESSAGES.ACCESS_PROHIBITED) + raise HTTPException(status_code=401, detail=ERROR_MESSAGES.ACCESS_PROHIBITED) @app.api_route("/{path:path}", methods=["GET", "POST", "PUT", "DELETE"]) @@ -78,43 +72,29 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): print(target_url, app.state.OPENAI_API_KEY) if user.role not in ["user", "admin"]: - raise HTTPException(status_code=401, - detail=ERROR_MESSAGES.ACCESS_PROHIBITED) + raise HTTPException(status_code=401, detail=ERROR_MESSAGES.ACCESS_PROHIBITED) if app.state.OPENAI_API_KEY == "": - raise HTTPException(status_code=401, - detail=ERROR_MESSAGES.API_KEY_NOT_FOUND) + raise HTTPException(status_code=401, detail=ERROR_MESSAGES.API_KEY_NOT_FOUND) + body = await request.body() + + # TODO: Remove below after gpt-4-vision fix from Open AI # Try to decode the body of the request from bytes to a UTF-8 string (Require add max_token to fix gpt-4-vision) try: - body_str = (await request.body()).decode('utf-8') - except UnicodeDecodeError as e: - print("Error decoding request body:", e) - raise HTTPException(status_code=400, detail="Invalid request body") - # Check if the body is not empty - if body_str: - try: - - body_dict = json.loads(body_str) - except json.JSONDecodeError as e: - print("Error loading request body into a dictionary:", e) - raise HTTPException(status_code=400, detail="Invalid JSON in request body") - + body = body.decode("utf-8") + body = json.loads(body) + # Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000 # This is a workaround until OpenAI fixes the issue with this model - if body_dict.get("model") == "gpt-4-vision-preview": - body_dict["max_tokens"] = 4000 - print("Modified body_dict:", body_dict) - - # Try to convert the modified body back to JSON - try: - # Convert the modified body back to JSON - body_json = json.dumps(body_dict) - except TypeError as e: - print("Error converting modified body to JSON:", e) - raise HTTPException(status_code=500, detail="Internal server error") - else: - body_json = body_str # If the body is empty, use it as is + if body.get("model") == "gpt-4-vision-preview": + body["max_tokens"] = 4000 + print("Modified body_dict:", body) + # Convert the modified body back to JSON + body = json.dumps(body) + except json.JSONDecodeError as e: + print("Error loading request body into a dictionary:", e) + raise HTTPException(status_code=400, detail="Invalid JSON in request body") headers = {} headers["Authorization"] = f"Bearer {app.state.OPENAI_API_KEY}" @@ -124,7 +104,7 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): r = requests.request( method=request.method, url=target_url, - data=body_json, + data=body, headers=headers, stream=True, ) @@ -153,8 +133,8 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): if "openai" in app.state.OPENAI_API_BASE_URL and path == "models": response_data["data"] = list( - filter(lambda model: "gpt" in model["id"], - response_data["data"])) + filter(lambda model: "gpt" in model["id"], response_data["data"]) + ) return response_data except Exception as e: From 83181b7968b462ad5e218802273213b6fa6ca3ab Mon Sep 17 00:00:00 2001 From: "Timothy J. Baek" Date: Mon, 22 Jan 2024 01:41:00 -0800 Subject: [PATCH 4/4] fix: add max_token only when field not present --- backend/apps/openai/main.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/backend/apps/openai/main.py b/backend/apps/openai/main.py index fd6acee7..1529ab93 100644 --- a/backend/apps/openai/main.py +++ b/backend/apps/openai/main.py @@ -87,7 +87,8 @@ async def proxy(path: str, request: Request, user=Depends(get_current_user)): # Check if the model is "gpt-4-vision-preview" and set "max_tokens" to 4000 # This is a workaround until OpenAI fixes the issue with this model if body.get("model") == "gpt-4-vision-preview": - body["max_tokens"] = 4000 + if "max_tokens" not in body: + body["max_tokens"] = 4000 print("Modified body_dict:", body) # Convert the modified body back to JSON