forked from phoenix/litellm-mirror
fix otel use sensible defaults
This commit is contained in:
parent
0295a22561
commit
3376f151c6
3 changed files with 35 additions and 5 deletions
|
@ -650,8 +650,6 @@ class OpenTelemetry(CustomLogger):
|
||||||
OTLPSpanExporterHTTP(
|
OTLPSpanExporterHTTP(
|
||||||
endpoint=self.OTEL_ENDPOINT, headers=_split_otel_headers
|
endpoint=self.OTEL_ENDPOINT, headers=_split_otel_headers
|
||||||
),
|
),
|
||||||
max_queue_size=100,
|
|
||||||
max_export_batch_size=100,
|
|
||||||
)
|
)
|
||||||
elif self.OTEL_EXPORTER == "otlp_grpc":
|
elif self.OTEL_EXPORTER == "otlp_grpc":
|
||||||
verbose_logger.debug(
|
verbose_logger.debug(
|
||||||
|
@ -662,8 +660,6 @@ class OpenTelemetry(CustomLogger):
|
||||||
OTLPSpanExporterGRPC(
|
OTLPSpanExporterGRPC(
|
||||||
endpoint=self.OTEL_ENDPOINT, headers=_split_otel_headers
|
endpoint=self.OTEL_ENDPOINT, headers=_split_otel_headers
|
||||||
),
|
),
|
||||||
max_queue_size=100,
|
|
||||||
max_export_batch_size=100,
|
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
verbose_logger.debug(
|
verbose_logger.debug(
|
||||||
|
|
|
@ -2015,6 +2015,37 @@
|
||||||
"supports_vision": true,
|
"supports_vision": true,
|
||||||
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||||
},
|
},
|
||||||
|
"gemini-1.5-flash-exp-0827": {
|
||||||
|
"max_tokens": 8192,
|
||||||
|
"max_input_tokens": 1000000,
|
||||||
|
"max_output_tokens": 8192,
|
||||||
|
"max_images_per_prompt": 3000,
|
||||||
|
"max_videos_per_prompt": 10,
|
||||||
|
"max_video_length": 1,
|
||||||
|
"max_audio_length_hours": 8.4,
|
||||||
|
"max_audio_per_prompt": 1,
|
||||||
|
"max_pdf_size_mb": 30,
|
||||||
|
"input_cost_per_image": 0.0001315,
|
||||||
|
"input_cost_per_video_per_second": 0.0001315,
|
||||||
|
"input_cost_per_audio_per_second": 0.000125,
|
||||||
|
"input_cost_per_token": 0.0000005,
|
||||||
|
"input_cost_per_character": 0.000000125,
|
||||||
|
"input_cost_per_token_above_128k_tokens": 0.000001,
|
||||||
|
"input_cost_per_character_above_128k_tokens": 0.00000025,
|
||||||
|
"output_cost_per_token": 0.0000015,
|
||||||
|
"output_cost_per_character": 0.000000375,
|
||||||
|
"output_cost_per_token_above_128k_tokens": 0.000003,
|
||||||
|
"output_cost_per_character_above_128k_tokens": 0.00000075,
|
||||||
|
"output_cost_per_image": 0.000263,
|
||||||
|
"output_cost_per_video_per_second": 0.000263,
|
||||||
|
"output_cost_per_audio_per_second": 0.00025,
|
||||||
|
"litellm_provider": "vertex_ai-language-models",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_system_messages": true,
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_vision": true,
|
||||||
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#foundation_models"
|
||||||
|
},
|
||||||
"gemini-1.5-flash-001": {
|
"gemini-1.5-flash-001": {
|
||||||
"max_tokens": 8192,
|
"max_tokens": 8192,
|
||||||
"max_input_tokens": 1000000,
|
"max_input_tokens": 1000000,
|
||||||
|
|
|
@ -14,3 +14,6 @@ model_list:
|
||||||
|
|
||||||
general_settings:
|
general_settings:
|
||||||
master_key: sk-1234
|
master_key: sk-1234
|
||||||
|
|
||||||
|
litellm_settings:
|
||||||
|
callbacks: ["otel"]
|
Loading…
Add table
Add a link
Reference in a new issue