Krish Dholakia
9bdcef238b
Merge pull request #4907 from BerriAI/litellm_proxy_get_secret
...
fix(proxy_server.py): fix get secret for environment_variables
2024-07-26 22:17:11 -07:00
Krish Dholakia
f9c2fec1a6
Merge pull request #4918 from BerriAI/litellm_ollama_tool_calling
...
feat(ollama_chat.py): support ollama tool calling
2024-07-26 22:16:58 -07:00
Krrish Dholakia
77fe8f57cf
docs(ollama.md): add ollama tool calling to docs
2024-07-26 22:12:52 -07:00
Krrish Dholakia
b25d4a8cb3
feat(ollama_chat.py): support ollama tool calling
...
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Ishaan Jaff
2501b4eccd
feat link to model cost map on swagger
2024-07-26 21:34:42 -07:00
Ishaan Jaff
548adea8cf
add litellm_header_name endpoint
2024-07-26 21:04:31 -07:00
Krrish Dholakia
a264d1ca8c
feat(vertex_httpx.py): support logging citation metadata
...
Closes https://github.com/BerriAI/litellm/issues/3230
2024-07-26 20:54:59 -07:00
Krrish Dholakia
fe7f78fbf6
feat(vertex_httpx.py): support logging vertex ai safety results to langfuse
...
Closes https://github.com/BerriAI/litellm/issues/3230
2024-07-26 20:50:43 -07:00
Ishaan Jaff
a7f964b869
Merge pull request #4913 from BerriAI/litellm_fix_error_limit
...
[Proxy-Fix] - raise more descriptive errors when crossing tpm / rpm limits on keys, user, global limits
2024-07-26 20:25:28 -07:00
Ishaan Jaff
3c463ccbe6
Merge pull request #4914 from BerriAI/litellm_fix_batches
...
[Proxy-Fix + Test] - /batches endpoint
2024-07-26 20:12:03 -07:00
Krrish Dholakia
fe0b55f2ca
fix(utils.py): fix cache hits for streaming
...
Fixes https://github.com/BerriAI/litellm/issues/4109
2024-07-26 19:04:08 -07:00
Ishaan Jaff
2541d5f625
add verbose_logger.debug to retrieve batch
2024-07-26 18:26:39 -07:00
Ishaan Jaff
f627fa9b40
fix for GET /v1/batches{batch_id:path}
2024-07-26 18:23:15 -07:00
Ishaan Jaff
56ce7e892d
fix batches inserting metadata
2024-07-26 18:08:54 -07:00
Ishaan Jaff
159a880dcc
fix /v1/batches POST
2024-07-26 18:06:00 -07:00
Ishaan Jaff
c4e4b4675c
fix raise better error when crossing tpm / rpm limits
2024-07-26 17:35:08 -07:00
Idris Mokhtarzada
e8d4234dbd
Better JSON serialization for Datadog logs
...
Dicts are now properly serialized to JSON so that Datadog can parse the child attributes. Also, numbers and nulls are sent as numbers and nulls instead of strings.
2024-07-26 17:02:05 -04:00
Idris Mokhtarzada
a7e877d15f
Use milliseconds for response_time in Datadog logs
...
milliseconds is more commonly used and more standard than seconds
2024-07-26 16:43:21 -04:00
Idris Mokhtarzada
9b89280a90
Use underscores
...
Datadog does not play nice with special characters (as in "(seconds)"). Also just makes sense to standardize on either underscores or camelCase, but not mix-and-match.
2024-07-26 16:38:54 -04:00
Krrish Dholakia
9943c6d607
fix(proxy_server.py): fix get secret for environment_variables
2024-07-26 13:33:02 -07:00
Krrish Dholakia
9a6ed8cabb
fix(bedrock_httpx.py): fix streaming error message
...
Fixes https://github.com/BerriAI/litellm/issues/4900
2024-07-26 10:42:47 -07:00
Krish Dholakia
67115a56c0
Merge pull request #4869 from maamalama/anthropic-tools
...
Fixed tool_call for Helicone integration
2024-07-26 10:42:10 -07:00
Krrish Dholakia
84482703b8
docs(config.md): update wildcard docs
2024-07-26 08:59:53 -07:00
Krrish Dholakia
1d6c39a607
feat(proxy_server.py): handle pydantic mockselvar error
...
Fixes https://github.com/BerriAI/litellm/issues/4898#issuecomment-2252105485
2024-07-26 08:38:51 -07:00
Krrish Dholakia
ce210ddaf6
fix(vertex_ai_llama3.py): Fix llama3 streaming issue
...
Closes https://github.com/BerriAI/litellm/issues/4885
2024-07-25 22:30:55 -07:00
Krrish Dholakia
2f773d9cb6
fix(litellm_cost_calc/google.py): support meta llama vertex ai cost tracking
2024-07-25 22:12:07 -07:00
Ishaan Jaff
079a41fbe1
Merge branch 'main' into litellm_proxy_support_all_providers
2024-07-25 20:15:37 -07:00
Ishaan Jaff
87cebdefd3
Merge pull request #4896 from BerriAI/docs_add_example_usage_with_mistral_python
...
Docs Proxy - add example usage with mistral SDK with Proxy
2024-07-25 20:02:26 -07:00
Krrish Dholakia
826bb125e8
test(test_router.py): handle azure api instability
2024-07-25 19:54:40 -07:00
Krrish Dholakia
a2fd8459fc
fix(utils.py): don't raise error on openai content filter during streaming - return as is
...
Fixes issue where we would raise an error vs. openai who return the chunk with finish reason as 'content_filter'
2024-07-25 19:50:52 -07:00
Ishaan Jaff
68e94f0976
example mistral sdk
2024-07-25 19:48:54 -07:00
Ishaan Jaff
5bec2bf513
Merge pull request #4894 from BerriAI/litellm_logfire_dotenv
...
fix logfire - don't load_dotenv
2024-07-25 19:34:35 -07:00
Ishaan Jaff
693bcfac39
fix using pass_through_all_models
2024-07-25 19:32:49 -07:00
Krish Dholakia
c2086300b7
Merge branch 'main' into litellm_redis_team_object
2024-07-25 19:31:52 -07:00
Krish Dholakia
b6ca4406b6
Merge branch 'main' into bedrock-llama3.1-405b
2024-07-25 19:29:10 -07:00
Ishaan Jaff
a0655b4192
Merge pull request #4884 from Manouchehri/add-mistral-large-2407-bedrock-1
...
Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock
2024-07-25 19:22:46 -07:00
Ishaan Jaff
fcd834b277
fix logfire - don't load_dotenv
2024-07-25 19:22:26 -07:00
Krish Dholakia
c0c934d9cf
Merge pull request #4879 from fracapuano/main
...
Add Single-Token predictions support for Replicate
2024-07-25 19:10:57 -07:00
Krish Dholakia
a306b83b2d
Merge pull request #4887 from BerriAI/litellm_custom_llm
...
feat(custom_llm.py): Support Custom LLM Handlers
2024-07-25 19:05:29 -07:00
Krrish Dholakia
41abd51240
fix(custom_llm.py): pass input params to custom llm
2024-07-25 19:03:52 -07:00
Ishaan Jaff
9863520376
support using */*
2024-07-25 18:48:56 -07:00
Ishaan Jaff
8f4c5437b8
router support setting pass_through_all_models
2024-07-25 18:34:12 -07:00
Ishaan Jaff
e67daf79be
router support setting pass_through_all_models
2024-07-25 18:22:35 -07:00
Krrish Dholakia
bd7af04a72
feat(proxy_server.py): support custom llm handler on proxy
2024-07-25 17:56:34 -07:00
Ishaan Jaff
f2443996d8
feat support audio health checks for azure
2024-07-25 17:30:15 -07:00
Ishaan Jaff
2432c90515
feat - support health check audio_speech
2024-07-25 17:26:14 -07:00
Ishaan Jaff
e3142b4294
fix whisper health check with litellm
2024-07-25 17:22:57 -07:00
Krrish Dholakia
060249c7e0
feat(utils.py): support async streaming for custom llm provider
2024-07-25 17:11:57 -07:00
Krrish Dholakia
b4e3a77ad0
feat(utils.py): support sync streaming for custom llm provider
2024-07-25 16:47:32 -07:00
Krrish Dholakia
9f97436308
fix(custom_llm.py): support async completion calls
2024-07-25 15:51:39 -07:00