Ishaan Jaff
b93152e978
assume index is not always in stream chunk
2024-08-15 19:42:03 -07:00
Ishaan Jaff
fa569aaf6f
feat add support for aws_region_name
2024-08-15 19:32:59 -07:00
Ishaan Jaff
b4ba12e22c
show bedrock, sagemaker creds in verbose mode
2024-08-15 19:10:11 -07:00
Ishaan Jaff
b96a27b230
add verbose logging on test
2024-08-15 19:05:23 -07:00
Ishaan Jaff
40dc27e72c
fix sagemaker tests
2024-08-15 18:34:20 -07:00
Ishaan Jaff
0d374fb7c0
fix sagemaker test
2024-08-15 18:23:41 -07:00
Ishaan Jaff
df4ea8fba6
refactor sagemaker to be async
2024-08-15 18:18:02 -07:00
Krrish Dholakia
28d7de30de
fix(litellm_logging.py): wrap function to safely fail
2024-08-15 18:05:06 -07:00
Krrish Dholakia
f6dba82882
feat(litellm_logging.py): cleanup payload + add response cost to logged payload
2024-08-15 17:53:25 -07:00
Krrish Dholakia
3ddeb3297d
fix(litellm_logging.py): fix standard payload
2024-08-15 17:33:40 -07:00
Krrish Dholakia
cda50e5d47
fix(s3.py): fix s3 logging payload to have valid json values
...
Previously pydantic objects were being stringified, making them unparsable
2024-08-15 17:09:02 -07:00
Ishaan Jaff
e4638ee4e4
Merge pull request #5233 from BerriAI/litellm_make_sagemaker_async
...
[Fix] use BaseAWSLLM for bedrock, sagemaker
2024-08-15 15:13:29 -07:00
Ishaan Jaff
b1aed699ea
test sync sagemaker calls
2024-08-15 15:12:31 -07:00
Ishaan Jaff
2c9e3e9bd7
run mock tests for test_completion_sagemaker
2024-08-15 14:49:21 -07:00
Ishaan Jaff
b58c2bef1c
add non-stream mock tests for sagemaker
2024-08-15 14:48:24 -07:00
Ishaan Jaff
c32cb365d0
Merge pull request #5226 from BerriAI/litellm_always_log_cache_key
...
[Fix-Proxy + Langfuse] Always log cache_key on hits/misses
2024-08-15 14:01:15 -07:00
Ishaan Jaff
e217eda303
use BaseAWSLLM for bedrock getcredentials
2024-08-15 13:58:47 -07:00
Ishaan Jaff
5c1c9f7616
fix ImportError
2024-08-15 13:02:44 -07:00
Paul Lang
97fce022e4
Fixes the tool_use
chunk mapping
2024-08-15 15:44:09 -04:00
Ishaan Jaff
5f693971f7
fix - don't require boto3 on the cli
2024-08-15 12:36:38 -07:00
Ishaan Jaff
08922041da
Merge pull request #5224 from BerriAI/litellm_use_common_helper_llm_call
...
[Feat-Proxy] - user common helper to `route_request` for making llm call
2024-08-15 11:49:29 -07:00
Ishaan Jaff
a59ed00fd3
litellm always log cache_key on hits/misses
2024-08-15 09:59:58 -07:00
Ishaan Jaff
7a17b2132f
fix /moderations endpoint
2024-08-15 08:58:28 -07:00
Ishaan Jaff
c50a60004f
fix test proxy exception mapping
2024-08-15 08:52:28 -07:00
Ishaan Jaff
58828403ea
refactor use 1 util for llm routing
2024-08-15 08:42:20 -07:00
Ishaan Jaff
d50f26d73d
simplify logic for routing llm request
2024-08-15 08:29:28 -07:00
Ishaan Jaff
fdd6664420
use route_request for making llm call
2024-08-15 08:16:44 -07:00
Kyrylo Yefimenko
e3cd8e169b
Use sepecific llama2 and llama3 model names in Ollama
2024-08-15 12:27:51 +01:00
lidongming
e1f53fcc80
Fix incorrect message length check in cost calculator
2024-08-15 16:59:38 +08:00
Krrish Dholakia
eb6a0a32f1
docs(bedrock.md): add guardrails on config.yaml to docs
2024-08-14 22:11:19 -07:00
Krrish Dholakia
c7fd626805
docs(team_logging.md): add key-based logging to docs
2024-08-14 21:49:55 -07:00
Krrish Dholakia
3487d84fcc
docs(pass_through.md): add doc on using langfuse client sdk w/ litellm proxy
2024-08-14 21:43:31 -07:00
Krrish Dholakia
28faafadb1
feat(pass_through_endpoints.py): initial commit of crud endpoints for pass through endpoints
2024-08-14 21:36:07 -07:00
Krrish Dholakia
a020563149
feat(proxy_server.py): support returning available fields for pass_through_endpoints via `/config/field/list
2024-08-14 19:07:10 -07:00
Ishaan Jaff
d8ef882905
fix langfuse log_provider_specific_information_as_span
2024-08-14 17:54:18 -07:00
Krish Dholakia
970892fea7
Merge pull request #5209 from BerriAI/litellm_azure_ai_openai_support
...
fix(utils.py): support calling openai models via `azure_ai/`
2024-08-14 17:53:33 -07:00
Krish Dholakia
435d39bb24
Merge branch 'main' into litellm_azure_ai_openai_support
2024-08-14 17:53:27 -07:00
Ishaan Jaff
9c039a9064
bump: version 1.43.12 → 1.43.13
2024-08-14 17:47:20 -07:00
Ishaan Jaff
1f631606a5
Merge pull request #5210 from BerriAI/litellm_add_prompt_caching_support
...
[Feat] Add Anthropic API Prompt Caching Support
2024-08-14 17:43:01 -07:00
Ishaan Jaff
912acb1cae
docs using proxy with context caaching anthropic
2024-08-14 17:42:48 -07:00
Krish Dholakia
c5647ac596
Merge pull request #5211 from msabramo/use-AZURE_API_VERSION-as-default-azure-openai-version
...
Use `AZURE_API_VERSION` env var as default azure openai version
2024-08-14 17:37:55 -07:00
Ishaan Jaff
2267b8a59f
docs add examples with litellm proxy
2024-08-14 17:13:26 -07:00
Ishaan Jaff
fd122aa7a3
docs add examples doing context caching anthropic sdk
2024-08-14 17:07:51 -07:00
Ishaan Jaff
78a2013e51
add test for large context in system message for anthropic
2024-08-14 17:03:10 -07:00
Ishaan Jaff
76a5f5d433
move claude prompt caching to diff file
2024-08-14 16:28:12 -07:00
Ishaan Jaff
e0ff4823d0
add test for caching tool calls
2024-08-14 16:19:14 -07:00
Marc Abramowitz
ac9aa1ab67
Use AZURE_API_VERSION as default azure openai version
...
Without this change, the default version of the Azure OpenAI API is hardcoded in
the code as an old version, `"2024-02-01"`. This change allows the user to set
the default version of the Azure OpenAI API by setting the environment variable
`AZURE_API_VERSION` or by using the command-line parameter `--api_version`.
2024-08-14 15:47:57 -07:00
Krrish Dholakia
6333b04be3
fix(factory.py): handle assistant null content
2024-08-14 15:44:38 -07:00
Ishaan Jaff
fccc6dc928
fix bedrock test
2024-08-14 15:27:20 -07:00
Ishaan Jaff
45e367d4d4
docs Caching - Continuing Multi-Turn Convo
2024-08-14 15:26:25 -07:00