Krish Dholakia
5f13d2ee64
Merge pull request #5029 from BerriAI/litellm_azure_ui_fix
...
fix(utils.py): Fix adding azure models on ui
2024-08-02 22:12:19 -07:00
Ishaan Jaff
057bb37a39
test vertex ft jobs
2024-08-02 22:03:52 -07:00
Joe Cheng
1fbfc09b44
Add unit test
2024-08-02 20:51:08 -07:00
Ishaan Jaff
0463e73ee3
fix typing
2024-08-02 18:49:16 -07:00
Ishaan Jaff
f194aa3a93
fix typing
2024-08-02 18:46:43 -07:00
Ishaan Jaff
4fc27e87c5
add vertex ai ft on proxy
2024-08-02 18:26:36 -07:00
Ishaan Jaff
40430dde10
add support for sync vertex ft
2024-08-02 18:16:14 -07:00
Ishaan Jaff
db3b2a3ee0
fix translating response
2024-08-02 18:13:59 -07:00
Ishaan Jaff
0e70aca776
convert response obj from vertex ai
2024-08-02 18:13:25 -07:00
Ishaan Jaff
960de8a3d0
test ft response vertex ai
2024-08-02 18:02:34 -07:00
Ishaan Jaff
ac6224c2b1
translate response from vertex to openai
2024-08-02 18:02:24 -07:00
Krrish Dholakia
5d96ff6694
fix(utils.py): handle scenario where model="azure/*" and custom_llm_provider="azure"
...
Fixes https://github.com/BerriAI/litellm/issues/4912
2024-08-02 17:48:53 -07:00
Ishaan Jaff
7ec1f241fc
Merge pull request #5026 from BerriAI/litellm_fix_whisper_caching
...
[Fix] Whisper Caching - Use correct cache keys for checking request in cache
2024-08-02 17:26:28 -07:00
Ishaan Jaff
d7ab440280
example vertex ai.jsonl
2024-08-02 17:25:32 -07:00
Ishaan Jaff
cb8a592ccf
test vertex ft
2024-08-02 17:25:18 -07:00
Ishaan Jaff
693ab3759c
add vertex ft support
2024-08-02 17:25:03 -07:00
Ishaan Jaff
d364d76bd0
add vertex FT spec
2024-08-02 17:24:25 -07:00
Ishaan Jaff
653cf55850
add fine tuning for vertex
2024-08-02 17:23:58 -07:00
Krish Dholakia
3ee8ae231c
Merge pull request #5027 from BerriAI/litellm_admin_viewer_team_tab
...
build(ui): allow admin_viewer to view teams tab
2024-08-02 17:18:48 -07:00
Ishaan Jaff
cff7050147
add vertex_credentials in router param
2024-08-02 16:58:17 -07:00
Krrish Dholakia
186e5f55e8
fix(user_api_key_auth.py): fix linting errors
2024-08-02 16:34:32 -07:00
Krish Dholakia
3d4ee61358
Merge pull request #5019 from BerriAI/litellm_deepseek_caching
...
fix(types/utils.py): Support deepseek prompt caching
2024-08-02 16:26:49 -07:00
Krrish Dholakia
1080c7014e
build(ui): allow admin_viewer to view teams tab
...
Allows admin viewe role to see available teams on proxy ui
2024-08-02 16:19:40 -07:00
Ishaan Jaff
7fa25c443a
docs caching
2024-08-02 15:19:27 -07:00
Ishaan Jaff
aadc29aab8
docs call types
2024-08-02 15:17:09 -07:00
Ishaan Jaff
2fd523a0f9
docs supported call types
2024-08-02 15:14:38 -07:00
Ishaan Jaff
87a15b408f
fix test
2024-08-02 15:08:44 -07:00
Ishaan Jaff
d6c3560ff1
test whisper
2024-08-02 15:07:38 -07:00
Ishaan Jaff
ec3b0d0d0b
return cache hit True on cache hits
2024-08-02 15:07:05 -07:00
Ishaan Jaff
b6b19dc128
use file name when getting cache key
2024-08-02 14:52:08 -07:00
Ishaan Jaff
1b3bc32090
log correct file name on langfuse
2024-08-02 14:49:25 -07:00
Joe Cheng
33f4411f17
Fix tool call coalescing
...
The previous code seemed to assume that the tool call index property
started at 0, but Anthropic sometimes returns them starting at 1.
This was causing an extra null-ish tool call to be materialized.
2024-08-02 13:05:23 -07:00
Joe Cheng
90dd60fa71
fix(main.py): Handle bedrock tool calling in stream_chunk_builder
...
Fixes #5022 .
The streaming chunks from Anthropic seem to violate an assumption
that is implicit in the stream_chunk_builder implementation: that
only tool_calls OR function_calls OR content will appear in a
streamed response. The repro in #5022 shows that you can get
content followed by tool calls.
These changes properly handle these combinations by building
separate lists of each type of chunk (note that in theory a chunk
could appear in multiple lists, e.g. both delta.tool_calls and
delta.content being present on one chunk).
2024-08-02 12:41:13 -07:00
Ishaan Jaff
7f1d32a75b
use regular ci/cd pipeline
2024-08-02 12:13:15 -07:00
Ishaan Jaff
fc8a87efec
ci/cd run again
2024-08-02 11:46:45 -07:00
Ishaan Jaff
ffa7818339
fix config.yaml
2024-08-02 11:41:53 -07:00
Ishaan Jaff
459964decf
fix config
2024-08-02 11:39:54 -07:00
Ishaan Jaff
53d57509ab
bump: version 1.42.10 → 1.42.11
2024-08-02 11:29:56 -07:00
Ishaan Jaff
d42549d6e4
temp testing ci/cd
2024-08-02 11:29:41 -07:00
Ishaan Jaff
b3196df0b8
queue stable release testing after new GH release
2024-08-02 10:57:56 -07:00
Krrish Dholakia
0ce6c038ee
fix(anthropic.py): fix linting error
2024-08-02 10:37:53 -07:00
Krish Dholakia
70c45767c2
Merge pull request #5020 from BerriAI/litellm_openai_raw_response_headers
...
refactor(openai/azure.py): move to returning openai/azure response headers by default
2024-08-02 10:36:14 -07:00
Krrish Dholakia
fe2aa706e8
refactor(openai/azure.py): move to returning openai/azure response headers by default
...
Allows token tracking to work more reliably across multiple azure/openai deployments
2024-08-02 09:42:08 -07:00
Krrish Dholakia
0a30ba9674
fix(types/utils.py): support passing prompt cache usage stats in usage object
...
Passes deepseek prompt caching values through to end user
2024-08-02 09:30:50 -07:00
Krrish Dholakia
cd073d5ad3
test: handle anthropic rate limit error
2024-08-02 08:57:09 -07:00
Ishaan Jaff
48419ca8bc
add step to ghcr deploy
2024-08-02 08:06:09 -07:00
Krrish Dholakia
fe7e68adc8
fix(utils.py): fix codestral streaming
2024-08-02 07:38:06 -07:00
Ishaan Jaff
75fba18c9f
fix langfuse hardcoded public key
2024-08-02 07:21:02 -07:00
Krrish Dholakia
4c2ef8ea64
fix(bedrock_httpx.py): fix ai21 streaming
2024-08-01 22:03:24 -07:00
Krrish Dholakia
57e3044974
bump: version 1.42.9 → 1.42.10
2024-08-01 21:29:36 -07:00