remove mocks

This commit is contained in:
yujonglee 2024-06-02 19:49:34 +09:00
parent d464b97810
commit c5e9e89288
2 changed files with 113 additions and 67 deletions

View file

@ -1,6 +1,7 @@
from dataclasses import dataclass, field
from typing import Optional
import os import os
import litellm
from litellm.integrations.custom_logger import CustomLogger from litellm.integrations.custom_logger import CustomLogger
from opentelemetry import trace from opentelemetry import trace
@ -16,18 +17,30 @@ from opentelemetry.sdk.trace.export import (
ConsoleSpanExporter, ConsoleSpanExporter,
) )
LITELLM_TRACER_NAME = "litellm" LITELLM_TRACER_NAME = "litellm"
LITELLM_RESOURCE = {"service.name": "litellm"} LITELLM_RESOURCE = {"service.name": "litellm"}
MOCK_TRACE_PARENT = {"traceparent": "SOMETHING_FROM_PROXY_REQUEST"}
MOCK_SPAN_NAME = "TODO" @dataclass
class OpenTelemetryConfig:
exporter: str = field(default="console")
endpoint: Optional[str] = None
bearer_token: Optional[str] = None
@classmethod
def from_env(cls):
return cls(
exporter=os.getenv("OTEL_EXPORTER", "console"),
endpoint=os.getenv("OTEL_ENDPOINT"),
bearer_token=os.getenv("OTEL_BEARER_TOKEN"),
)
class OpenTelemetry(CustomLogger): class OpenTelemetry(CustomLogger):
def __init__(self): def __init__(self, config=OpenTelemetryConfig.from_env()):
self.config = config
provider = TracerProvider(resource=Resource(attributes=LITELLM_RESOURCE)) provider = TracerProvider(resource=Resource(attributes=LITELLM_RESOURCE))
provider.add_span_processor(self.get_span_processor()) provider.add_span_processor(self._get_span_processor())
trace.set_tracer_provider(provider) trace.set_tracer_provider(provider)
self.tracer = trace.get_tracer(LITELLM_TRACER_NAME) self.tracer = trace.get_tracer(LITELLM_TRACER_NAME)
@ -46,42 +59,56 @@ class OpenTelemetry(CustomLogger):
def _handle_sucess(self, kwargs, response_obj, start_time, end_time): def _handle_sucess(self, kwargs, response_obj, start_time, end_time):
span = self.tracer.start_span( span = self.tracer.start_span(
MOCK_SPAN_NAME, name=self._get_span_name(kwargs),
start_time=self._to_ns(start_time), start_time=self._to_ns(start_time),
context=TraceContextTextMapPropagator().extract(carrier=MOCK_TRACE_PARENT), context=self._get_span_context(kwargs),
) )
span.set_status(Status(StatusCode.OK)) span.set_status(Status(StatusCode.OK))
self._set_attributes(span, kwargs, response_obj) self.set_attributes(span, kwargs, response_obj)
span.end(end_time=self._to_ns(end_time)) span.end(end_time=self._to_ns(end_time))
def _handle_failure(self, kwargs, response_obj, start_time, end_time): def _handle_failure(self, kwargs, response_obj, start_time, end_time):
span = self.tracer.start_span( span = self.tracer.start_span(
MOCK_SPAN_NAME, name=self._get_span_name(kwargs),
start_time=self._to_ns(start_time), start_time=self._to_ns(start_time),
context=TraceContextTextMapPropagator().extract(carrier=MOCK_TRACE_PARENT), context=self._get_span_context(kwargs),
) )
span.set_status(Status(StatusCode.ERROR)) span.set_status(Status(StatusCode.ERROR))
self._set_attributes(span, kwargs, response_obj) self.set_attributes(span, kwargs, response_obj)
span.end(end_time=self._to_ns(end_time)) span.end(end_time=self._to_ns(end_time))
def set_attributes(self, span, kwargs, response_obj):
for key in ["model", "api_base", "api_version"]:
if key in kwargs:
span.set_attribute(key, kwargs[key])
def _to_ns(self, dt): def _to_ns(self, dt):
return int(dt.timestamp() * 1e9) return int(dt.timestamp() * 1e9)
def _set_attributes(self, span, kwargs, response_obj): def _get_span_name(self, kwargs):
keys = ["model", "api_base", "api_version"] f"litellm-{kwargs.get('call_type', 'completion')}"
for key in keys:
if key in kwargs:
span.set_attribute("model", kwargs[key])
def get_span_processor(self): def _get_span_context(self, kwargs):
if litellm.set_verbose: litellm_params = kwargs.get("litellm_params", {}) or {}
BatchSpanProcessor(ConsoleSpanExporter()) proxy_server_request = litellm_params.get("proxy_server_request", {}) or {}
headers = proxy_server_request.get("headers", {}) or {}
traceparent = headers.get("traceparent", None)
if traceparent is None:
return None
else: else:
BatchSpanProcessor( carrier = {"traceparent": traceparent}
return TraceContextTextMapPropagator().extract(carrier=carrier)
def _get_span_processor(self):
if self.config.exporter == "console":
return BatchSpanProcessor(ConsoleSpanExporter())
elif self.config.exporter == "otlp_http":
return BatchSpanProcessor(
OTLPSpanExporter( OTLPSpanExporter(
endpoint=os.getenv("OTEL_ENDPOINT"), endpoint=self.OTEL_ENDPOINT,
headers={ headers={"Authorization": f"Bearer {self.OTEL_BEARER_TOKEN}"},
"Authorization": f"Bearer {os.getenv('OTEL_BEARER_TOKEN')}"
},
) )
) )
else:
return BatchSpanProcessor(ConsoleSpanExporter())

View file

@ -41,7 +41,7 @@ example_completion_result = {
{ {
"message": { "message": {
"content": "Whispers of the wind carry dreams to me.", "content": "Whispers of the wind carry dreams to me.",
"role": "assistant" "role": "assistant",
} }
} }
], ],
@ -69,21 +69,11 @@ example_embedding_result = {
} }
], ],
"model": "text-embedding-3-small", "model": "text-embedding-3-small",
"usage": { "usage": {"prompt_tokens": 5, "total_tokens": 5},
"prompt_tokens": 5,
"total_tokens": 5
}
} }
example_image_generation_result = { example_image_generation_result = {
"created": 1589478378, "created": 1589478378,
"data": [ "data": [{"url": "https://..."}, {"url": "https://..."}],
{
"url": "https://..."
},
{
"url": "https://..."
}
]
} }
@ -185,7 +175,9 @@ def test_engines_model_chat_completions(mock_acompletion, client_no_auth):
} }
print("testing proxy server with chat completions") print("testing proxy server with chat completions")
response = client_no_auth.post("/engines/gpt-3.5-turbo/chat/completions", json=test_data) response = client_no_auth.post(
"/engines/gpt-3.5-turbo/chat/completions", json=test_data
)
mock_acompletion.assert_called_once_with( mock_acompletion.assert_called_once_with(
model="gpt-3.5-turbo", model="gpt-3.5-turbo",
messages=[ messages=[
@ -249,7 +241,9 @@ def test_chat_completion_azure(mock_acompletion, client_no_auth):
@mock_patch_acompletion() @mock_patch_acompletion()
def test_openai_deployments_model_chat_completions_azure(mock_acompletion, client_no_auth): def test_openai_deployments_model_chat_completions_azure(
mock_acompletion, client_no_auth
):
global headers global headers
try: try:
# Your test data # Your test data
@ -388,10 +382,10 @@ def test_img_gen(mock_aimage_generation, client_no_auth):
response = client_no_auth.post("/v1/images/generations", json=test_data) response = client_no_auth.post("/v1/images/generations", json=test_data)
mock_aimage_generation.assert_called_once_with( mock_aimage_generation.assert_called_once_with(
model='dall-e-3', model="dall-e-3",
prompt='A cute baby sea otter', prompt="A cute baby sea otter",
n=1, n=1,
size='1024x1024', size="1024x1024",
metadata=mock.ANY, metadata=mock.ANY,
proxy_server_request=mock.ANY, proxy_server_request=mock.ANY,
) )
@ -608,3 +602,28 @@ def test_load_router_config(mock_cache, fake_env_vars):
# test_load_router_config() # test_load_router_config()
from litellm.integrations.opentelemetry import OpenTelemetry, OpenTelemetryConfig
@mock_patch_acompletion()
def test_otel_with_proxy_server(mock_acompletion, client_no_auth, capsys):
litellm.callbacks = [OpenTelemetry(OpenTelemetryConfig(exporter="console"))]
data = {"model": "gpt-3.5-turbo", "messages": [{"role": "user", "content": "hi"}]}
response = client_no_auth.post("/v1/chat/completions", json=data)
mock_acompletion.assert_called_once_with(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "hi"}],
litellm_call_id=mock.ANY,
litellm_logging_obj=mock.ANY,
request_timeout=mock.ANY,
specific_deployment=True,
metadata=mock.ANY,
proxy_server_request=mock.ANY,
)
assert response.status_code == 200
assert response.json() == example_completion_result
print(capsys.readouterr())