Krrish Dholakia
|
c0204310ee
|
fix(main.py): fix translation to text_completions format for async text completion calls
|
2024-03-30 09:02:51 -07:00 |
|
Krrish Dholakia
|
63271846c2
|
refactor(main.py): trigger new build
|
2024-03-29 09:24:47 -07:00 |
|
Krrish Dholakia
|
d547944556
|
fix(sagemaker.py): support 'model_id' param for sagemaker
allow passing inference component param to sagemaker in the same format as we handle this for bedrock
|
2024-03-29 08:43:17 -07:00 |
|
Krish Dholakia
|
28905c85b6
|
Merge pull request #2720 from onukura/ollama-batch-embedding
Batch embedding for Ollama
|
2024-03-28 14:58:55 -07:00 |
|
Krrish Dholakia
|
664663f301
|
refactor(main.py): trigger new build
|
2024-03-28 14:52:47 -07:00 |
|
onukura
|
f86472518d
|
Add a feature to ollama aembedding to accept batch input
|
2024-03-27 21:39:19 +00:00 |
|
Krrish Dholakia
|
9375b131ee
|
refactor(main.py): trigger new build
|
2024-03-26 21:18:51 -07:00 |
|
Krish Dholakia
|
7eb2c7942c
|
Merge pull request #2675 from onukura/ollama-embedding
Fix Ollama embedding
|
2024-03-26 16:08:28 -07:00 |
|
onukura
|
ef69eefcdb
|
Fix ollama api_base to enable remote url
|
2024-03-25 16:26:40 +00:00 |
|
Krrish Dholakia
|
f98aead602
|
feat(main.py): support router.chat.completions.create
allows using router with instructor
https://github.com/BerriAI/litellm/issues/2673
|
2024-03-25 08:26:28 -07:00 |
|
Krrish Dholakia
|
b7321ae4ee
|
fix(router.py): fix pre call check logic
|
2024-03-23 18:56:08 -07:00 |
|
Krrish Dholakia
|
94f55aa6d9
|
fix(bedrock.py): support claude 3 function calling when stream=true
https://github.com/BerriAI/litellm/issues/2615
|
2024-03-21 18:39:03 -07:00 |
|
Krrish Dholakia
|
af27a61d76
|
refactor(main.py): trigger new build
|
2024-03-21 10:56:44 -07:00 |
|
Krrish Dholakia
|
d6624bf6c3
|
refactor(main.py): trigger new build
|
2024-03-19 21:05:53 -07:00 |
|
Krish Dholakia
|
c840fecdeb
|
Merge pull request #2142 from vilmar-hillow/azure_embedding_ad_token
Fixed azure ad token not being processed properly in embedding models
|
2024-03-19 11:51:28 -07:00 |
|
Krish Dholakia
|
c4dbd0407e
|
Merge pull request #2561 from BerriAI/litellm_batch_writing_db
fix(proxy/utils.py): move to batch writing db updates
|
2024-03-18 21:50:47 -07:00 |
|
Krrish Dholakia
|
2827acc487
|
refactor(main.py): trigger new build
|
2024-03-18 21:27:32 -07:00 |
|
Krrish Dholakia
|
693b5eb376
|
docs(main.py): add timeout to docstring
|
2024-03-18 21:23:46 -07:00 |
|
Bincheng Li
|
e605b04927
|
fix bug: custom prompt templates registered are never applied to vllm provider
|
2024-03-17 15:21:14 +08:00 |
|
Krrish Dholakia
|
c69ae8efce
|
refactor(main.py): trigger new build
|
2024-03-16 18:49:54 -07:00 |
|
Krish Dholakia
|
32ca306123
|
Merge pull request #2535 from BerriAI/litellm_fireworks_ai_support
feat(utils.py): add native fireworks ai support
|
2024-03-15 10:02:53 -07:00 |
|
Krrish Dholakia
|
860b06d273
|
refactor(main.py): trigger new build
|
2024-03-15 09:42:23 -07:00 |
|
Krrish Dholakia
|
9909f44015
|
feat(utils.py): add native fireworks ai support
addresses - https://github.com/BerriAI/litellm/issues/777, https://github.com/BerriAI/litellm/issues/2486
|
2024-03-15 09:09:59 -07:00 |
|
Krrish Dholakia
|
0b6cf3d5cf
|
refactor(main.py): trigger new build
|
2024-03-14 13:01:18 -07:00 |
|
Krrish Dholakia
|
bdd2004691
|
refactor(main.py): trigger new build
|
2024-03-14 12:10:39 -07:00 |
|
Krrish Dholakia
|
7876aa2d75
|
fix(parallel_request_limiter.py): handle metadata being none
|
2024-03-14 10:02:41 -07:00 |
|
Krrish Dholakia
|
16e3aaced5
|
docs(enterprise.md): add prompt injection detection to docs
|
2024-03-13 12:37:32 -07:00 |
|
Krish Dholakia
|
9f2d540ebf
|
Merge pull request #2472 from BerriAI/litellm_anthropic_streaming_tool_calling
fix(anthropic.py): support claude-3 streaming with function calling
|
2024-03-12 21:36:01 -07:00 |
|
Dmitry Supranovich
|
57ebb9582e
|
Fixed azure ad token not being processed properly in embedding models
|
2024-03-12 21:29:24 -04:00 |
|
Krish Dholakia
|
0d18f3c0ca
|
Merge pull request #2473 from BerriAI/litellm_fix_compatible_provider_model_name
fix(openai.py): return model name with custom llm provider for openai-compatible endpoints (e.g. mistral, together ai, etc.)
|
2024-03-12 12:58:29 -07:00 |
|
Ishaan Jaff
|
5172fb1de9
|
Merge pull request #2474 from BerriAI/litellm_support_command_r
[New-Model] Cohere/command-r
|
2024-03-12 11:11:56 -07:00 |
|
Krrish Dholakia
|
d2286fb93c
|
fix(main.py): trigger new build
|
2024-03-12 11:07:14 -07:00 |
|
Krrish Dholakia
|
0033613b9e
|
fix(openai.py): return model name with custom llm provider for openai compatible endpoints
|
2024-03-12 10:30:10 -07:00 |
|
ishaan-jaff
|
7635c764cf
|
(feat) cohere_chat provider
|
2024-03-12 10:29:26 -07:00 |
|
Krrish Dholakia
|
86ed0aaba8
|
fix(anthropic.py): support streaming with function calling
|
2024-03-12 09:52:11 -07:00 |
|
ishaan-jaff
|
b193b01f40
|
(feat) support azure/gpt-instruct models
|
2024-03-12 09:30:15 -07:00 |
|
Krrish Dholakia
|
e07174736f
|
refactor(main.py): trigger new build
|
2024-03-11 13:57:40 -07:00 |
|
Krrish Dholakia
|
942b5e4145
|
fix(main.py): trigger new build
|
2024-03-10 09:48:06 -07:00 |
|
Krish Dholakia
|
caa99f43bf
|
Merge branch 'main' into litellm_load_balancing_transcription_endpoints
|
2024-03-08 23:08:47 -08:00 |
|
Krish Dholakia
|
e245b1c98a
|
Merge pull request #2401 from BerriAI/litellm_transcription_endpoints
feat(main.py): support openai transcription endpoints
|
2024-03-08 23:07:48 -08:00 |
|
Krrish Dholakia
|
0fb7afe820
|
feat(proxy_server.py): working /audio/transcription endpoint
|
2024-03-08 18:20:27 -08:00 |
|
ishaan-jaff
|
ddd231a8c2
|
(feat) use no-log as a litellm param
|
2024-03-08 16:46:38 -08:00 |
|
ishaan-jaff
|
986a526790
|
(feat) disable logging per request
|
2024-03-08 16:25:54 -08:00 |
|
Krrish Dholakia
|
ae54b398d2
|
feat(router.py): add load balancing for async transcription calls
|
2024-03-08 13:58:15 -08:00 |
|
Krrish Dholakia
|
6b1049217e
|
feat(azure.py): add support for calling whisper endpoints on azure
|
2024-03-08 13:48:38 -08:00 |
|
Krrish Dholakia
|
696eb54455
|
feat(main.py): support openai transcription endpoints
enable user to load balance between openai + azure transcription endpoints
|
2024-03-08 10:25:19 -08:00 |
|
Krrish Dholakia
|
2f9a39f30c
|
refactor(main.py): trigger new build
|
2024-03-08 08:12:22 -08:00 |
|
Krrish Dholakia
|
b9854a99d2
|
test: increase time before checking budget reset - avoid deadlocking
|
2024-03-06 22:16:59 -08:00 |
|
Krrish Dholakia
|
cdb960eb34
|
fix(vertex_ai.py): correctly parse optional params and pass vertex ai project
|
2024-03-06 14:00:50 -08:00 |
|
Krrish Dholakia
|
387864662e
|
fix(main.py): trigger new build
|
2024-03-05 15:50:40 -08:00 |
|