diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index fa50b06f1..8778aa360 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -965,15 +965,15 @@ }, "dolphin": { "max_tokens": 4096, - "input_cost_per_token": 0.00002, - "output_cost_per_token": 0.00002, + "input_cost_per_token": 0.0000005, + "output_cost_per_token": 0.0000005, "litellm_provider": "nlp_cloud", "mode": "completion" }, "chatdolphin": { "max_tokens": 4096, - "input_cost_per_token": 0.00002, - "output_cost_per_token": 0.00002, + "input_cost_per_token": 0.0000005, + "output_cost_per_token": 0.0000005, "litellm_provider": "nlp_cloud", "mode": "chat" }, diff --git a/litellm/proxy/proxy_server.py b/litellm/proxy/proxy_server.py index 545540422..c56339638 100644 --- a/litellm/proxy/proxy_server.py +++ b/litellm/proxy/proxy_server.py @@ -2259,8 +2259,13 @@ async def completion( response = await llm_router.atext_completion( **data, specific_deployment=True ) - else: # router is not set + elif user_model is not None: # `litellm --model ` response = await litellm.atext_completion(**data) + else: + raise HTTPException( + status_code=status.HTTP_400_BAD_REQUEST, + detail={"error": "Invalid model name passed in"}, + ) if hasattr(response, "_hidden_params"): model_id = response._hidden_params.get("model_id", None) or ""