Commit graph

9699 commits

Author SHA1 Message Date
Krish Dholakia
9bdcef238b
Merge pull request #4907 from BerriAI/litellm_proxy_get_secret
fix(proxy_server.py): fix get secret for environment_variables
2024-07-26 22:17:11 -07:00
Krish Dholakia
f9c2fec1a6
Merge pull request #4918 from BerriAI/litellm_ollama_tool_calling
feat(ollama_chat.py): support ollama tool calling
2024-07-26 22:16:58 -07:00
Krrish Dholakia
77fe8f57cf docs(ollama.md): add ollama tool calling to docs 2024-07-26 22:12:52 -07:00
Krrish Dholakia
b25d4a8cb3 feat(ollama_chat.py): support ollama tool calling
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Ishaan Jaff
2501b4eccd feat link to model cost map on swagger 2024-07-26 21:34:42 -07:00
Ishaan Jaff
548adea8cf add litellm_header_name endpoint 2024-07-26 21:04:31 -07:00
Krrish Dholakia
a264d1ca8c feat(vertex_httpx.py): support logging citation metadata
Closes https://github.com/BerriAI/litellm/issues/3230
2024-07-26 20:54:59 -07:00
Krrish Dholakia
fe7f78fbf6 feat(vertex_httpx.py): support logging vertex ai safety results to langfuse
Closes https://github.com/BerriAI/litellm/issues/3230
2024-07-26 20:50:43 -07:00
Ishaan Jaff
a7f964b869
Merge pull request #4913 from BerriAI/litellm_fix_error_limit
[Proxy-Fix] - raise more descriptive errors when crossing tpm / rpm limits on keys, user, global limits
2024-07-26 20:25:28 -07:00
Ishaan Jaff
3c463ccbe6
Merge pull request #4914 from BerriAI/litellm_fix_batches
[Proxy-Fix + Test] - /batches endpoint
2024-07-26 20:12:03 -07:00
Krrish Dholakia
fe0b55f2ca fix(utils.py): fix cache hits for streaming
Fixes https://github.com/BerriAI/litellm/issues/4109
2024-07-26 19:04:08 -07:00
Ishaan Jaff
2541d5f625 add verbose_logger.debug to retrieve batch 2024-07-26 18:26:39 -07:00
Ishaan Jaff
f627fa9b40 fix for GET /v1/batches{batch_id:path} 2024-07-26 18:23:15 -07:00
Ishaan Jaff
56ce7e892d fix batches inserting metadata 2024-07-26 18:08:54 -07:00
Ishaan Jaff
159a880dcc fix /v1/batches POST 2024-07-26 18:06:00 -07:00
Ishaan Jaff
c4e4b4675c fix raise better error when crossing tpm / rpm limits 2024-07-26 17:35:08 -07:00
Idris Mokhtarzada
e8d4234dbd
Better JSON serialization for Datadog logs
Dicts are now properly serialized to JSON so that Datadog can parse the child attributes.  Also, numbers and nulls are sent as numbers and nulls instead of strings.
2024-07-26 17:02:05 -04:00
Idris Mokhtarzada
a7e877d15f
Use milliseconds for response_time in Datadog logs
milliseconds is more commonly used and more standard than seconds
2024-07-26 16:43:21 -04:00
Idris Mokhtarzada
9b89280a90
Use underscores
Datadog does not play nice with special characters (as in "(seconds)").  Also just makes sense to standardize on either underscores or camelCase, but not mix-and-match.
2024-07-26 16:38:54 -04:00
Krrish Dholakia
9943c6d607 fix(proxy_server.py): fix get secret for environment_variables 2024-07-26 13:33:02 -07:00
Krrish Dholakia
9a6ed8cabb fix(bedrock_httpx.py): fix streaming error message
Fixes https://github.com/BerriAI/litellm/issues/4900
2024-07-26 10:42:47 -07:00
Krish Dholakia
67115a56c0
Merge pull request #4869 from maamalama/anthropic-tools
Fixed tool_call for Helicone integration
2024-07-26 10:42:10 -07:00
Krrish Dholakia
84482703b8 docs(config.md): update wildcard docs 2024-07-26 08:59:53 -07:00
Krrish Dholakia
1d6c39a607 feat(proxy_server.py): handle pydantic mockselvar error
Fixes https://github.com/BerriAI/litellm/issues/4898#issuecomment-2252105485
2024-07-26 08:38:51 -07:00
Krrish Dholakia
ce210ddaf6 fix(vertex_ai_llama3.py): Fix llama3 streaming issue
Closes https://github.com/BerriAI/litellm/issues/4885
2024-07-25 22:30:55 -07:00
Krrish Dholakia
2f773d9cb6 fix(litellm_cost_calc/google.py): support meta llama vertex ai cost tracking 2024-07-25 22:12:07 -07:00
Ishaan Jaff
079a41fbe1
Merge branch 'main' into litellm_proxy_support_all_providers 2024-07-25 20:15:37 -07:00
Ishaan Jaff
87cebdefd3
Merge pull request #4896 from BerriAI/docs_add_example_usage_with_mistral_python
Docs Proxy - add example usage with mistral SDK with Proxy
2024-07-25 20:02:26 -07:00
Krrish Dholakia
826bb125e8 test(test_router.py): handle azure api instability 2024-07-25 19:54:40 -07:00
Krrish Dholakia
a2fd8459fc fix(utils.py): don't raise error on openai content filter during streaming - return as is
Fixes issue where we would raise an error vs. openai who return the chunk with finish reason as 'content_filter'
2024-07-25 19:50:52 -07:00
Ishaan Jaff
68e94f0976 example mistral sdk 2024-07-25 19:48:54 -07:00
Ishaan Jaff
5bec2bf513
Merge pull request #4894 from BerriAI/litellm_logfire_dotenv
fix logfire - don't load_dotenv
2024-07-25 19:34:35 -07:00
Ishaan Jaff
693bcfac39 fix using pass_through_all_models 2024-07-25 19:32:49 -07:00
Krish Dholakia
c2086300b7
Merge branch 'main' into litellm_redis_team_object 2024-07-25 19:31:52 -07:00
Krish Dholakia
b6ca4406b6
Merge branch 'main' into bedrock-llama3.1-405b 2024-07-25 19:29:10 -07:00
Ishaan Jaff
a0655b4192
Merge pull request #4884 from Manouchehri/add-mistral-large-2407-bedrock-1
Add mistral.mistral-large-2407-v1:0 on Amazon Bedrock
2024-07-25 19:22:46 -07:00
Ishaan Jaff
fcd834b277 fix logfire - don't load_dotenv 2024-07-25 19:22:26 -07:00
Krish Dholakia
c0c934d9cf
Merge pull request #4879 from fracapuano/main
Add Single-Token predictions support for Replicate
2024-07-25 19:10:57 -07:00
Krish Dholakia
a306b83b2d
Merge pull request #4887 from BerriAI/litellm_custom_llm
feat(custom_llm.py): Support Custom LLM Handlers
2024-07-25 19:05:29 -07:00
Krrish Dholakia
41abd51240 fix(custom_llm.py): pass input params to custom llm 2024-07-25 19:03:52 -07:00
Ishaan Jaff
9863520376 support using */* 2024-07-25 18:48:56 -07:00
Ishaan Jaff
8f4c5437b8 router support setting pass_through_all_models 2024-07-25 18:34:12 -07:00
Ishaan Jaff
e67daf79be router support setting pass_through_all_models 2024-07-25 18:22:35 -07:00
Krrish Dholakia
bd7af04a72 feat(proxy_server.py): support custom llm handler on proxy 2024-07-25 17:56:34 -07:00
Ishaan Jaff
f2443996d8 feat support audio health checks for azure 2024-07-25 17:30:15 -07:00
Ishaan Jaff
2432c90515 feat - support health check audio_speech 2024-07-25 17:26:14 -07:00
Ishaan Jaff
e3142b4294 fix whisper health check with litellm 2024-07-25 17:22:57 -07:00
Krrish Dholakia
060249c7e0 feat(utils.py): support async streaming for custom llm provider 2024-07-25 17:11:57 -07:00
Krrish Dholakia
b4e3a77ad0 feat(utils.py): support sync streaming for custom llm provider 2024-07-25 16:47:32 -07:00
Krrish Dholakia
9f97436308 fix(custom_llm.py): support async completion calls 2024-07-25 15:51:39 -07:00