Commit graph

150 commits

Author SHA1 Message Date
Krrish Dholakia
5d3a0ace4b fix(openai.py): fix client caching logic 2024-06-01 16:45:56 -07:00
Ishaan Jaff
1a69682282 fix used hashed api key 2024-06-01 09:24:16 -07:00
Ishaan Jaff
a89a6f2ee9 fix - in memory client cache 2024-06-01 08:58:22 -07:00
Ishaan Jaff
d2efc8c63b fix cache openai client for embeddings, text, speech 2024-05-31 21:35:03 -07:00
Ishaan Jaff
9ff540870c fix - linting error 2024-05-31 21:24:14 -07:00
Ishaan Jaff
506251ecdf feat - cache openai clients 2024-05-31 21:22:06 -07:00
Krish Dholakia
c049b6b4af Merge pull request #3936 from BerriAI/litellm_assistants_api_proxy
feat(proxy_server.py): add assistants api endpoints to proxy server
2024-05-31 18:43:22 -07:00
Krrish Dholakia
2fdf4a7bb4 feat(proxy_server.py): add assistants api endpoints to proxy server 2024-05-30 22:44:43 -07:00
Krrish Dholakia
eb159b64e1 fix(openai.py): fix openai response for /audio/speech endpoint 2024-05-30 16:41:06 -07:00
Krrish Dholakia
1e89a1f56e feat(main.py): support openai tts endpoint
Closes https://github.com/BerriAI/litellm/issues/3094
2024-05-30 14:28:28 -07:00
Krrish Dholakia
5f01dce284 fix(main.py): pass api key and api base to openai.py for audio transcription call 2024-05-29 21:29:01 -07:00
Krrish Dholakia
915e599772 fix(openai.py): only allow 'user' as optional param if openai model 2024-05-29 15:15:02 -07:00
Ishaan Jaff
cd4a3627e8 feat - add afile_content, file_content 2024-05-28 20:58:22 -07:00
Ishaan Jaff
fe704e5857 feat - add aretrieve_batch 2024-05-28 17:12:41 -07:00
Ishaan Jaff
c580fe03a0 feat - add acreate_batch 2024-05-28 17:03:29 -07:00
Ishaan Jaff
80cae343cc feat - add litellm.acreate_file 2024-05-28 16:47:27 -07:00
Ishaan Jaff
fc4ca265b8 working create_batch 2024-05-28 15:45:23 -07:00
Ishaan Jaff
4dc7bfebd4 feat - import batches in __init__ 2024-05-28 15:35:11 -07:00
Krrish Dholakia
2b729624e6 fix(openai.py): fix deepinfra config optional param 2024-05-27 18:36:34 -07:00
Krrish Dholakia
23542fc1d2 fix(utils.py): support deepinfra optional params
Fixes https://github.com/BerriAI/litellm/issues/3855
2024-05-27 09:16:56 -07:00
Krrish Dholakia
e3c5e004c5 feat(databricks.py): add embedding model support 2024-05-23 18:22:03 -07:00
Krrish Dholakia
143a44823a feat(databricks.py): adds databricks support - completion, async, streaming
Closes https://github.com/BerriAI/litellm/issues/2160
2024-05-23 16:29:46 -07:00
Krish Dholakia
dc8f238449 Merge pull request #3657 from phact/patch-4
Another dictionary changed size during iteration error
2024-05-20 17:45:50 -07:00
Krrish Dholakia
0016477d3b feat(proxy_server.py): new 'supported_openai_params' endpoint
get supported openai params for a given model
2024-05-20 08:39:50 -07:00
Krrish Dholakia
7da925cada fix(utils.py): drop response_format if 'drop_params=True' for gpt-4 2024-05-18 13:02:48 -07:00
Sebastián Estévez
d4edf0f0ae Another dictionary changed size during iteration error
```
ImportError while loading conftest '/astra-assistants-api/tests/openai-sdk/conftest.py'.
conftest.py:13: in <module>
    from impl.astra_vector import CassandraClient
../../impl/astra_vector.py:45: in <module>
    from impl.services.inference_utils import get_embeddings
../../impl/services/inference_utils.py:5: in <module>
    import litellm
.cache/pypoetry/virtualenvs/astra-assistants-api-eiSmbCzm-py3.10/lib/python3.10/site-packages/litellm/__init__.py:678: in <module>
    from .main import *  # type: ignore
.cache/pypoetry/virtualenvs/astra-assistants-api-eiSmbCzm-py3.10/lib/python3.10/site-packages/litellm/main.py:73: in <module>
    from .llms.azure_text import AzureTextCompletion
.cache/pypoetry/virtualenvs/astra-assistants-api-eiSmbCzm-py3.10/lib/python3.10/site-packages/litellm/llms/azure_text.py:23: in <module>
    openai_text_completion_config = OpenAITextCompletionConfig()
.cache/pypoetry/virtualenvs/astra-assistants-api-eiSmbCzm-py3.10/lib/python3.10/site-packages/litellm/llms/openai.py:192: in __init__
    for key, value in locals_.items():
E   RuntimeError: dictionary changed size during iteration
```
2024-05-15 17:06:54 -04:00
Krrish Dholakia
96336cdd49 fix(openai.py): creat MistralConfig with response_format mapping for mistral api 2024-05-13 13:29:58 -07:00
Ishaan Jaff
953805aac5 stream_options for text-completionopenai 2024-05-09 08:37:40 -07:00
Ishaan Jaff
1b9da4a533 support stream_options for chat completion models 2024-05-08 21:52:25 -07:00
Krrish Dholakia
5f93cae3ff feat(proxy_server.py): return litellm version in response headers 2024-05-08 16:00:08 -07:00
Krrish Dholakia
91e96ce4b1 fix(openai.py): fix typing import for python 3.8 2024-05-04 21:49:30 -07:00
Krrish Dholakia
e7669652b5 fix(openai.py): fix linting error 2024-05-04 21:48:42 -07:00
Krrish Dholakia
a04b7578f5 feat(assistants/main.py): support litellm.get_assistants() and litellm.get_messages() 2024-05-04 21:30:28 -07:00
Krrish Dholakia
6b061e3cc5 feat(assistants/main.py): support 'litellm.get_threads' 2024-05-04 21:14:03 -07:00
Krrish Dholakia
300ad1dec1 feat(assistants/main.py): add 'add_message' endpoint 2024-05-04 19:56:11 -07:00
Krrish Dholakia
b0845d82cd fix(assistants/main.py): support litellm.create_thread() call 2024-05-04 19:35:37 -07:00
Krrish Dholakia
92d548b50b feat(openai.py): add support for openai assistants
v0 commit. Closes https://github.com/BerriAI/litellm/issues/2842
2024-05-04 17:27:48 -07:00
Krrish Dholakia
7715a9d333 fix(bedrock.py): convert httpx.timeout to boto3 valid timeout
Closes https://github.com/BerriAI/litellm/issues/3398
2024-05-03 16:24:21 -07:00
Krrish Dholakia
f1b2405fe0 fix(router.py): fix default retry logic 2024-04-25 11:57:27 -07:00
Krrish Dholakia
b10f03706d fix(utils.py): fix streaming to not return usage dict
Fixes https://github.com/BerriAI/litellm/issues/3237
2024-04-24 08:06:07 -07:00
Krrish Dholakia
7880dcbdf7 fix(openai.py): support passing prompt as list instead of concat string 2024-04-03 15:23:20 -07:00
Krrish Dholakia
f3fcfe3007 fix(proxy_server.py): return original model response via response headers - /v1/completions
to help devs with debugging
2024-04-03 13:05:43 -07:00
Krrish Dholakia
de3b314620 fix(openai.py): switch to using openai sdk for text completion calls 2024-04-02 15:08:12 -07:00
Krrish Dholakia
ed46af19ec fix(openai.py): return logprobs for text completion calls 2024-04-02 14:05:56 -07:00
Krrish Dholakia
5546f9f10a fix(main.py): support max retries for transcription calls 2024-04-01 18:37:53 -07:00
Krrish Dholakia
e94c4f818c fix(openai.py): return model name with custom llm provider for openai compatible endpoints 2024-03-12 10:30:10 -07:00
Krrish Dholakia
548e9a3590 fix(utils.py): fix model name checking 2024-03-09 18:22:26 -08:00
Krrish Dholakia
b2ce963498 feat: add cost tracking + caching for transcription calls 2024-03-09 15:43:38 -08:00
Krrish Dholakia
5b5f28dee8 fix(openai.py): fix async audio transcription 2024-03-08 23:33:54 -08:00
Krish Dholakia
f461352908 Merge branch 'main' into litellm_load_balancing_transcription_endpoints 2024-03-08 23:08:47 -08:00