forked from phoenix/litellm-mirror
Update README.md
This commit is contained in:
parent
96708d7c76
commit
ecb8d4c428
1 changed files with 10 additions and 5 deletions
15
README.md
15
README.md
|
@ -37,9 +37,6 @@
|
||||||
<a href="https://docs.litellm.ai/docs/providers" target="_blank">100+ Supported Models</a> |
|
<a href="https://docs.litellm.ai/docs/providers" target="_blank">100+ Supported Models</a> |
|
||||||
<a href="https://docs.litellm.ai/docs/" target="_blank">Docs</a> |
|
<a href="https://docs.litellm.ai/docs/" target="_blank">Docs</a> |
|
||||||
</h4>
|
</h4>
|
||||||
<a target="_blank" href="https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat">
|
|
||||||
<p>📣<b>1-click deploy</b> your own LLM proxy server. Grab time, if you're interested!</p>
|
|
||||||
</a>
|
|
||||||
|
|
||||||
LiteLLM manages
|
LiteLLM manages
|
||||||
- Translating inputs to the provider's completion and embedding endpoints
|
- Translating inputs to the provider's completion and embedding endpoints
|
||||||
|
@ -78,8 +75,6 @@ response = completion(model="command-nightly", messages=messages)
|
||||||
print(response)
|
print(response)
|
||||||
```
|
```
|
||||||
|
|
||||||
**Don't have a key? We'll give you access 👉 https://docs.litellm.ai/docs/proxy_api**
|
|
||||||
|
|
||||||
## Streaming
|
## Streaming
|
||||||
liteLLM supports streaming the model response back, pass `stream=True` to get a streaming iterator in response.
|
liteLLM supports streaming the model response back, pass `stream=True` to get a streaming iterator in response.
|
||||||
Streaming is supported for OpenAI, Azure, Anthropic, Huggingface models
|
Streaming is supported for OpenAI, Azure, Anthropic, Huggingface models
|
||||||
|
@ -93,6 +88,16 @@ result = completion('claude-2', messages, stream=True)
|
||||||
for chunk in result:
|
for chunk in result:
|
||||||
print(chunk['choices'][0]['delta'])
|
print(chunk['choices'][0]['delta'])
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## OpenAI Proxy Server
|
||||||
|
Spin up a local server to translate openai api calls to any non-openai model (e.g. Huggingface, TogetherAI, Ollama, etc.)
|
||||||
|
|
||||||
|
This works for async + streaming as well.
|
||||||
|
```python
|
||||||
|
litellm --model <model_name>
|
||||||
|
```
|
||||||
|
Running your model locally or on a custom endpoint ? Set the `--api-base` parameter [see how](https://docs.litellm.ai/docs/proxy_server#advanced-setting-api-base)
|
||||||
|
|
||||||
# Contributing
|
# Contributing
|
||||||
To contribute: Clone the repo locally -> Make a change -> Submit a PR with the change.
|
To contribute: Clone the repo locally -> Make a change -> Submit a PR with the change.
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue