forked from phoenix/litellm-mirror
docs(vertex.md): fix codestral fim placement (#5946)
This commit is contained in:
parent
bbf4db79c1
commit
71f68ac185
1 changed files with 79 additions and 80 deletions
|
@ -983,86 +983,6 @@ curl --location 'http://0.0.0.0:4000/chat/completions' \
|
||||||
</Tabs>
|
</Tabs>
|
||||||
|
|
||||||
|
|
||||||
## AI21 Models
|
|
||||||
|
|
||||||
| Model Name | Function Call |
|
|
||||||
|------------------|--------------------------------------|
|
|
||||||
| jamba-1.5-mini@001 | `completion(model='vertex_ai/jamba-1.5-mini@001', messages)` |
|
|
||||||
| jamba-1.5-large@001 | `completion(model='vertex_ai/jamba-1.5-large@001', messages)` |
|
|
||||||
|
|
||||||
### Usage
|
|
||||||
|
|
||||||
<Tabs>
|
|
||||||
<TabItem value="sdk" label="SDK">
|
|
||||||
|
|
||||||
```python
|
|
||||||
from litellm import completion
|
|
||||||
import os
|
|
||||||
|
|
||||||
os.environ["GOOGLE_APPLICATION_CREDENTIALS"] = ""
|
|
||||||
|
|
||||||
model = "meta/jamba-1.5-mini@001"
|
|
||||||
|
|
||||||
vertex_ai_project = "your-vertex-project" # can also set this as os.environ["VERTEXAI_PROJECT"]
|
|
||||||
vertex_ai_location = "your-vertex-location" # can also set this as os.environ["VERTEXAI_LOCATION"]
|
|
||||||
|
|
||||||
response = completion(
|
|
||||||
model="vertex_ai/" + model,
|
|
||||||
messages=[{"role": "user", "content": "hi"}],
|
|
||||||
vertex_ai_project=vertex_ai_project,
|
|
||||||
vertex_ai_location=vertex_ai_location,
|
|
||||||
)
|
|
||||||
print("\nModel Response", response)
|
|
||||||
```
|
|
||||||
</TabItem>
|
|
||||||
<TabItem value="proxy" label="Proxy">
|
|
||||||
|
|
||||||
**1. Add to config**
|
|
||||||
|
|
||||||
```yaml
|
|
||||||
model_list:
|
|
||||||
- model_name: jamba-1.5-mini
|
|
||||||
litellm_params:
|
|
||||||
model: vertex_ai/jamba-1.5-mini@001
|
|
||||||
vertex_ai_project: "my-test-project"
|
|
||||||
vertex_ai_location: "us-east-1"
|
|
||||||
- model_name: jamba-1.5-large
|
|
||||||
litellm_params:
|
|
||||||
model: vertex_ai/jamba-1.5-large@001
|
|
||||||
vertex_ai_project: "my-test-project"
|
|
||||||
vertex_ai_location: "us-west-1"
|
|
||||||
```
|
|
||||||
|
|
||||||
**2. Start proxy**
|
|
||||||
|
|
||||||
```bash
|
|
||||||
litellm --config /path/to/config.yaml
|
|
||||||
|
|
||||||
# RUNNING at http://0.0.0.0:4000
|
|
||||||
```
|
|
||||||
|
|
||||||
**3. Test it!**
|
|
||||||
|
|
||||||
```bash
|
|
||||||
curl --location 'http://0.0.0.0:4000/chat/completions' \
|
|
||||||
--header 'Authorization: Bearer sk-1234' \
|
|
||||||
--header 'Content-Type: application/json' \
|
|
||||||
--data '{
|
|
||||||
"model": "jamba-1.5-large",
|
|
||||||
"messages": [
|
|
||||||
{
|
|
||||||
"role": "user",
|
|
||||||
"content": "what llm are you"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
}'
|
|
||||||
```
|
|
||||||
|
|
||||||
</TabItem>
|
|
||||||
</Tabs>
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
### Usage - Codestral FIM
|
### Usage - Codestral FIM
|
||||||
|
|
||||||
Call Codestral on VertexAI via the OpenAI [`/v1/completion`](https://platform.openai.com/docs/api-reference/completions/create) endpoint for FIM tasks.
|
Call Codestral on VertexAI via the OpenAI [`/v1/completion`](https://platform.openai.com/docs/api-reference/completions/create) endpoint for FIM tasks.
|
||||||
|
@ -1150,6 +1070,85 @@ curl -X POST 'http://0.0.0.0:4000/completions' \
|
||||||
</Tabs>
|
</Tabs>
|
||||||
|
|
||||||
|
|
||||||
|
## AI21 Models
|
||||||
|
|
||||||
|
| Model Name | Function Call |
|
||||||
|
|------------------|--------------------------------------|
|
||||||
|
| jamba-1.5-mini@001 | `completion(model='vertex_ai/jamba-1.5-mini@001', messages)` |
|
||||||
|
| jamba-1.5-large@001 | `completion(model='vertex_ai/jamba-1.5-large@001', messages)` |
|
||||||
|
|
||||||
|
### Usage
|
||||||
|
|
||||||
|
<Tabs>
|
||||||
|
<TabItem value="sdk" label="SDK">
|
||||||
|
|
||||||
|
```python
|
||||||
|
from litellm import completion
|
||||||
|
import os
|
||||||
|
|
||||||
|
os.environ["GOOGLE_APPLICATION_CREDENTIALS"] = ""
|
||||||
|
|
||||||
|
model = "meta/jamba-1.5-mini@001"
|
||||||
|
|
||||||
|
vertex_ai_project = "your-vertex-project" # can also set this as os.environ["VERTEXAI_PROJECT"]
|
||||||
|
vertex_ai_location = "your-vertex-location" # can also set this as os.environ["VERTEXAI_LOCATION"]
|
||||||
|
|
||||||
|
response = completion(
|
||||||
|
model="vertex_ai/" + model,
|
||||||
|
messages=[{"role": "user", "content": "hi"}],
|
||||||
|
vertex_ai_project=vertex_ai_project,
|
||||||
|
vertex_ai_location=vertex_ai_location,
|
||||||
|
)
|
||||||
|
print("\nModel Response", response)
|
||||||
|
```
|
||||||
|
</TabItem>
|
||||||
|
<TabItem value="proxy" label="Proxy">
|
||||||
|
|
||||||
|
**1. Add to config**
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
model_list:
|
||||||
|
- model_name: jamba-1.5-mini
|
||||||
|
litellm_params:
|
||||||
|
model: vertex_ai/jamba-1.5-mini@001
|
||||||
|
vertex_ai_project: "my-test-project"
|
||||||
|
vertex_ai_location: "us-east-1"
|
||||||
|
- model_name: jamba-1.5-large
|
||||||
|
litellm_params:
|
||||||
|
model: vertex_ai/jamba-1.5-large@001
|
||||||
|
vertex_ai_project: "my-test-project"
|
||||||
|
vertex_ai_location: "us-west-1"
|
||||||
|
```
|
||||||
|
|
||||||
|
**2. Start proxy**
|
||||||
|
|
||||||
|
```bash
|
||||||
|
litellm --config /path/to/config.yaml
|
||||||
|
|
||||||
|
# RUNNING at http://0.0.0.0:4000
|
||||||
|
```
|
||||||
|
|
||||||
|
**3. Test it!**
|
||||||
|
|
||||||
|
```bash
|
||||||
|
curl --location 'http://0.0.0.0:4000/chat/completions' \
|
||||||
|
--header 'Authorization: Bearer sk-1234' \
|
||||||
|
--header 'Content-Type: application/json' \
|
||||||
|
--data '{
|
||||||
|
"model": "jamba-1.5-large",
|
||||||
|
"messages": [
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": "what llm are you"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
}'
|
||||||
|
```
|
||||||
|
|
||||||
|
</TabItem>
|
||||||
|
</Tabs>
|
||||||
|
|
||||||
|
|
||||||
## Model Garden
|
## Model Garden
|
||||||
| Model Name | Function Call |
|
| Model Name | Function Call |
|
||||||
|------------------|--------------------------------------|
|
|------------------|--------------------------------------|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue