forked from phoenix/litellm-mirror
Update README.md
This commit is contained in:
parent
866acdae4e
commit
8e4a167629
1 changed files with 2 additions and 34 deletions
36
README.md
36
README.md
|
@ -14,8 +14,8 @@
|
||||||
<img src="https://deploy.cloud.run/button.svg" width=200/>
|
<img src="https://deploy.cloud.run/button.svg" width=200/>
|
||||||
</a>
|
</a>
|
||||||
</h4>
|
</h4>
|
||||||
<a href="https://docs.litellm.ai/docs/simple_proxy" target="_blank">
|
<a href="https://github.com/BerriAI/litellm/tree/main/openai-proxy" target="_blank">
|
||||||
<h4 align="center">1-click OpenAI Proxy</h4>
|
<h4 align="center">OpenAI Proxy Server</h4>
|
||||||
</a>
|
</a>
|
||||||
<h4 align="center">
|
<h4 align="center">
|
||||||
<a href="https://pypi.org/project/litellm/" target="_blank">
|
<a href="https://pypi.org/project/litellm/" target="_blank">
|
||||||
|
@ -85,38 +85,6 @@ for chunk in result:
|
||||||
print(chunk['choices'][0]['delta'])
|
print(chunk['choices'][0]['delta'])
|
||||||
```
|
```
|
||||||
|
|
||||||
## OpenAI Proxy Server ([Docs](https://docs.litellm.ai/docs/proxy_server))
|
|
||||||
Create an OpenAI API compatible server to call any non-openai model (e.g. Huggingface, TogetherAI, Ollama, etc.)
|
|
||||||
|
|
||||||
This works for async + streaming as well.
|
|
||||||
```python
|
|
||||||
litellm --model <model_name>
|
|
||||||
|
|
||||||
#INFO: litellm proxy running on http://0.0.0.0:8000
|
|
||||||
```
|
|
||||||
Running your model locally or on a custom endpoint ? Set the `--api-base` parameter [see how](https://docs.litellm.ai/docs/proxy_server)
|
|
||||||
|
|
||||||
### Self-host server ([Docs](https://docs.litellm.ai/docs/proxy_server#deploy-proxy))
|
|
||||||
|
|
||||||
1. Clone the repo
|
|
||||||
```shell
|
|
||||||
git clone https://github.com/BerriAI/litellm.git
|
|
||||||
```
|
|
||||||
|
|
||||||
2. Modify `template_secrets.toml`
|
|
||||||
```shell
|
|
||||||
[keys]
|
|
||||||
OPENAI_API_KEY="sk-..."
|
|
||||||
|
|
||||||
[general]
|
|
||||||
default_model = "gpt-3.5-turbo"
|
|
||||||
```
|
|
||||||
|
|
||||||
3. Deploy
|
|
||||||
```shell
|
|
||||||
docker build -t litellm . && docker run -p 8000:8000 litellm
|
|
||||||
```
|
|
||||||
|
|
||||||
## Supported Provider ([Docs](https://docs.litellm.ai/docs/providers))
|
## Supported Provider ([Docs](https://docs.litellm.ai/docs/providers))
|
||||||
| Provider | [Completion](https://docs.litellm.ai/docs/#basic-usage) | [Streaming](https://docs.litellm.ai/docs/completion/stream#streaming-responses) | [Async Completion](https://docs.litellm.ai/docs/completion/stream#async-completion) | [Async Streaming](https://docs.litellm.ai/docs/completion/stream#async-streaming) |
|
| Provider | [Completion](https://docs.litellm.ai/docs/#basic-usage) | [Streaming](https://docs.litellm.ai/docs/completion/stream#streaming-responses) | [Async Completion](https://docs.litellm.ai/docs/completion/stream#async-completion) | [Async Streaming](https://docs.litellm.ai/docs/completion/stream#async-streaming) |
|
||||||
| ------------- | ------------- | ------------- | ------------- | ------------- |
|
| ------------- | ------------- | ------------- | ------------- | ------------- |
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue