fix(openai-proxy): fixes

This commit is contained in:
Krrish Dholakia 2023-10-23 14:39:06 -07:00
parent 6373f6bddd
commit a930f15233

View file

@ -1,6 +1,5 @@
import litellm
from fastapi import FastAPI, Request, HTTPException
import litellm, os
from fastapi import FastAPI, Request
from fastapi.routing import APIRouter
from fastapi.responses import StreamingResponse, FileResponse
from fastapi.middleware.cors import CORSMiddleware
@ -18,6 +17,8 @@ app.add_middleware(
allow_headers=["*"],
)
if ("LANGUFSE_PUBLIC_KEY" in os.environ and "LANGUFSE_SECRET_KEY" in os.environ) or "LANGFUSE_HOST" in os.environ:
litellm.success_callback = ["langfuse"]
#### API ENDPOINTS ####
@router.post("/v1/models")
@ -58,6 +59,7 @@ async def completion(request: Request):
@router.post("/v1/chat/completions")
@router.post("/chat/completions")
async def chat_completion(request: Request):
<<<<<<< HEAD
try:
data = await request.json()
if "authorization" in request.headers: # if users pass LLM api keys as part of header
@ -72,6 +74,26 @@ async def chat_completion(request: Request):
return response
except Exception as e:
return HTTPException(status_code=500, detail=str(e))
=======
data = await request.json()
api_key = request.headers.get("authorization")
api_key = api_key.split(" ")[1]
## check for special character - '|' <- used for bedrock (aws_access_key + "|" + aws_secret_access_key + "|" + aws_region_name)
if "|" in api_key: ## BEDROCK
aws_keys = api_key.split("|")
data["aws_access_key_id"] = aws_keys[0]
data["aws_secret_access_key"] = aws_keys[1]
data["aws_region_name"] = aws_keys[2]
else: ## ALL OTHER PROVIDERS
data["api_key"] = api_key
response = litellm.completion(
**data
)
if 'stream' in data and data['stream'] == True: # use generate_responses to stream responses
return StreamingResponse(data_generator(response), media_type='text/event-stream')
return response
>>>>>>> 968b835 (fix(openai-proxy): adding langfuse)
@router.get("/")
async def home(request: Request):