mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
added documentation for snowflake
This commit is contained in:
parent
4e4ad41de8
commit
65aff4bf54
2 changed files with 102 additions and 0 deletions
101
docs/my-website/docs/providers/snowflake.md
Normal file
101
docs/my-website/docs/providers/snowflake.md
Normal file
|
@ -0,0 +1,101 @@
|
|||
import Tabs from '@theme/Tabs';
|
||||
import TabItem from '@theme/TabItem';
|
||||
|
||||
|
||||
# Snowflake
|
||||
LiteLLM supports all Snowflake models.
|
||||
- `snowflake/deepseek-r1`
|
||||
- `snowflake/claude-3-5-sonnet`
|
||||
- `snowflake/llama3.2-1b`
|
||||
- `snowflake/llama3.2-3b`
|
||||
- `snowflake/llama3.1-8b`
|
||||
- `snowflake/llama3.1-70b`
|
||||
- `snowflake/llama3.3-70b`
|
||||
- `snowflake/snowflake-llama-3.3-70b`
|
||||
- `snowflake/llama3.1-405b`
|
||||
- `snowflake/snowflake-llama-3.1-405b`
|
||||
- `snowflake/snowflake-arctic`
|
||||
- `snowflake/reka-core`
|
||||
- `snowflake/reka-flash`
|
||||
- `snowflake/mistral-large2`
|
||||
- `snowflake/mixtral-8x7b`
|
||||
- `snowflake/mistral-7b`
|
||||
- `snowflake/jamba-instruct`
|
||||
- `snowflake/jamba-1.5-mini`
|
||||
- `snowflake/jamba-1.5-large`
|
||||
- `snowflake/gemma-7b`
|
||||
|
||||
Currently, Snowflake's REST API does not have an endpoint for `snowflake-arctic-embed` embedding models. If you want to use these embedding models with Litellm, you can call them through our Hugging Face provider.
|
||||
|
||||
Find the Arctic Embed models [here](https://huggingface.co/collections/Snowflake/arctic-embed-661fd57d50fab5fc314e4c18) on Hugging Face.
|
||||
## Supported OpenAI Parameters
|
||||
```
|
||||
"temperature",
|
||||
"max_tokens",
|
||||
"top_p",
|
||||
"response_format"
|
||||
```
|
||||
|
||||
## API KEYS
|
||||
|
||||
Snowflake does have API keys. Instead, you access the Snowflake API with your JWT token and account identifier.
|
||||
|
||||
```python
|
||||
import os
|
||||
os.environ["SNOWFLAKE_JWT"] = "YOUR JWT"
|
||||
os.environ["SNOWFLAKE_ACCOUNT_ID"] = "YOUR ACCOUNT IDENTIFIER"
|
||||
```
|
||||
## Usage
|
||||
|
||||
```python
|
||||
from litellm import completion
|
||||
|
||||
## set ENV variables
|
||||
os.environ["SNOWFLAKE_JWT"] = "YOUR JWT"
|
||||
os.environ["SNOWFLAKE_ACCOUNT_ID"] = "YOUR ACCOUNT IDENTIFIER"
|
||||
|
||||
# Snowflake call
|
||||
response = completion(
|
||||
model="snowflake/mistral-7b",
|
||||
messages = [{ "content": "Hello, how are you?","role": "user"}]
|
||||
)
|
||||
```
|
||||
|
||||
## Usage with LiteLLM Proxy
|
||||
|
||||
#### 1. Required env variables
|
||||
```bash
|
||||
export SNOWFLAKE_JWT=""
|
||||
export SNOWFLAKE_ACCOUNT_ID = ""
|
||||
```
|
||||
|
||||
#### 2. Start the proxy~
|
||||
```yaml
|
||||
model_list:
|
||||
- model_name: mistral-7b
|
||||
litellm_params:
|
||||
model: snowflake/mistral-7b
|
||||
api_key: YOUR_API_KEY
|
||||
api_base: https://YOUR-ACCOUNT-ID.snowflakecomputing.com/api/v2/cortex/inference:complete
|
||||
|
||||
```
|
||||
|
||||
```bash
|
||||
litellm --config /path/to/config.yaml
|
||||
```
|
||||
|
||||
#### 3. Test it
|
||||
```shell
|
||||
curl --location 'http://0.0.0.0:4000/chat/completions' \
|
||||
--header 'Content-Type: application/json' \
|
||||
--data ' {
|
||||
"model": "snowflake/mistral-7b",
|
||||
"messages": [
|
||||
{
|
||||
"role": "user",
|
||||
"content": "Hello, how are you?"
|
||||
}
|
||||
]
|
||||
}
|
||||
'
|
||||
```
|
|
@ -230,6 +230,7 @@ const sidebars = {
|
|||
"providers/sambanova",
|
||||
"providers/custom_llm_server",
|
||||
"providers/petals",
|
||||
"providers/snowflake"
|
||||
],
|
||||
},
|
||||
{
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue