mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
Added a guide for users who want to use LiteLLM with AI/ML API. (#7058)
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 13s
All checks were successful
Read Version from pyproject.toml / read-version (push) Successful in 13s
* Added a guide for users who want to use LiteLLM with AI/ML. * Minor changes * Minor changes * Fix sidebars.js --------- Co-authored-by: Krish Dholakia <krrishdholakia@gmail.com>
This commit is contained in:
parent
8d3a942fbd
commit
fbe3c58372
3 changed files with 165 additions and 4 deletions
|
@ -303,6 +303,7 @@ curl 'http://0.0.0.0:4000/key/generate' \
|
||||||
|-------------------------------------------------------------------------------------|---------------------------------------------------------|---------------------------------------------------------------------------------|-------------------------------------------------------------------------------------|-----------------------------------------------------------------------------------|-------------------------------------------------------------------------------|-------------------------------------------------------------------------|
|
|-------------------------------------------------------------------------------------|---------------------------------------------------------|---------------------------------------------------------------------------------|-------------------------------------------------------------------------------------|-----------------------------------------------------------------------------------|-------------------------------------------------------------------------------|-------------------------------------------------------------------------|
|
||||||
| [openai](https://docs.litellm.ai/docs/providers/openai) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
| [openai](https://docs.litellm.ai/docs/providers/openai) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
||||||
| [azure](https://docs.litellm.ai/docs/providers/azure) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
| [azure](https://docs.litellm.ai/docs/providers/azure) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
||||||
|
| [AI/ML API](https://docs.litellm.ai/docs/providers/aiml) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
||||||
| [aws - sagemaker](https://docs.litellm.ai/docs/providers/aws_sagemaker) | ✅ | ✅ | ✅ | ✅ | ✅ | |
|
| [aws - sagemaker](https://docs.litellm.ai/docs/providers/aws_sagemaker) | ✅ | ✅ | ✅ | ✅ | ✅ | |
|
||||||
| [aws - bedrock](https://docs.litellm.ai/docs/providers/bedrock) | ✅ | ✅ | ✅ | ✅ | ✅ | |
|
| [aws - bedrock](https://docs.litellm.ai/docs/providers/bedrock) | ✅ | ✅ | ✅ | ✅ | ✅ | |
|
||||||
| [google - vertex_ai](https://docs.litellm.ai/docs/providers/vertex) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
| [google - vertex_ai](https://docs.litellm.ai/docs/providers/vertex) | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
|
||||||
|
|
160
docs/my-website/docs/providers/aiml.md
Normal file
160
docs/my-website/docs/providers/aiml.md
Normal file
|
@ -0,0 +1,160 @@
|
||||||
|
# AI/ML API
|
||||||
|
|
||||||
|
Getting started with the AI/ML API is simple. Follow these steps to set up your integration:
|
||||||
|
|
||||||
|
### 1. Get Your API Key
|
||||||
|
To begin, you need an API key. You can obtain yours here:
|
||||||
|
🔑 [Get Your API Key](https://aimlapi.com/app/keys/?utm_source=aimlapi&utm_medium=github&utm_campaign=integration)
|
||||||
|
|
||||||
|
### 2. Explore Available Models
|
||||||
|
Looking for a different model? Browse the full list of supported models:
|
||||||
|
📚 [Full List of Models](https://docs.aimlapi.com/api-overview/model-database/text-models?utm_source=aimlapi&utm_medium=github&utm_campaign=integration)
|
||||||
|
|
||||||
|
### 3. Read the Documentation
|
||||||
|
For detailed setup instructions and usage guidelines, check out the official documentation:
|
||||||
|
📖 [AI/ML API Docs](https://docs.aimlapi.com/quickstart/setting-up?utm_source=aimlapi&utm_medium=github&utm_campaign=integration)
|
||||||
|
|
||||||
|
### 4. Need Help?
|
||||||
|
If you have any questions, feel free to reach out. We’re happy to assist! 🚀 [Discord](https://discord.gg/hvaUsJpVJf)
|
||||||
|
|
||||||
|
## Usage
|
||||||
|
You can choose from LLama, Qwen, Flux, and 200+ other open and closed-source models on aimlapi.com/models. For example:
|
||||||
|
|
||||||
|
```python
|
||||||
|
import litellm
|
||||||
|
|
||||||
|
response = litellm.completion(
|
||||||
|
model="openai/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", # The model name must include prefix "openai" + the model name from ai/ml api
|
||||||
|
api_key="", # your aiml api-key
|
||||||
|
api_base="https://api.aimlapi.com/v2",
|
||||||
|
messages=[
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": "Hey, how's it going?",
|
||||||
|
}
|
||||||
|
],
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
## Streaming
|
||||||
|
|
||||||
|
```python
|
||||||
|
import litellm
|
||||||
|
|
||||||
|
response = litellm.completion(
|
||||||
|
model="openai/Qwen/Qwen2-72B-Instruct", # The model name must include prefix "openai" + the model name from ai/ml api
|
||||||
|
api_key="", # your aiml api-key
|
||||||
|
api_base="https://api.aimlapi.com/v2",
|
||||||
|
messages=[
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": "Hey, how's it going?",
|
||||||
|
}
|
||||||
|
],
|
||||||
|
stream=True,
|
||||||
|
)
|
||||||
|
for chunk in response:
|
||||||
|
print(chunk)
|
||||||
|
```
|
||||||
|
|
||||||
|
## Async Completion
|
||||||
|
|
||||||
|
```python
|
||||||
|
import asyncio
|
||||||
|
|
||||||
|
import litellm
|
||||||
|
|
||||||
|
|
||||||
|
async def main():
|
||||||
|
response = await litellm.acompletion(
|
||||||
|
model="openai/anthropic/claude-3-5-haiku", # The model name must include prefix "openai" + the model name from ai/ml api
|
||||||
|
api_key="", # your aiml api-key
|
||||||
|
api_base="https://api.aimlapi.com/v2",
|
||||||
|
messages=[
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": "Hey, how's it going?",
|
||||||
|
}
|
||||||
|
],
|
||||||
|
)
|
||||||
|
print(response)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
asyncio.run(main())
|
||||||
|
```
|
||||||
|
|
||||||
|
## Async Streaming
|
||||||
|
|
||||||
|
```python
|
||||||
|
import asyncio
|
||||||
|
import traceback
|
||||||
|
|
||||||
|
import litellm
|
||||||
|
|
||||||
|
|
||||||
|
async def main():
|
||||||
|
try:
|
||||||
|
print("test acompletion + streaming")
|
||||||
|
response = await litellm.acompletion(
|
||||||
|
model="openai/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF", # The model name must include prefix "openai" + the model name from ai/ml api
|
||||||
|
api_key="", # your aiml api-key
|
||||||
|
api_base="https://api.aimlapi.com/v2",
|
||||||
|
messages=[{"content": "Hey, how's it going?", "role": "user"}],
|
||||||
|
stream=True,
|
||||||
|
)
|
||||||
|
print(f"response: {response}")
|
||||||
|
async for chunk in response:
|
||||||
|
print(chunk)
|
||||||
|
except:
|
||||||
|
print(f"error occurred: {traceback.format_exc()}")
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
asyncio.run(main())
|
||||||
|
```
|
||||||
|
|
||||||
|
## Async Embedding
|
||||||
|
|
||||||
|
```python
|
||||||
|
import asyncio
|
||||||
|
|
||||||
|
import litellm
|
||||||
|
|
||||||
|
|
||||||
|
async def main():
|
||||||
|
response = await litellm.aembedding(
|
||||||
|
model="openai/text-embedding-3-small", # The model name must include prefix "openai" + the model name from ai/ml api
|
||||||
|
api_key="", # your aiml api-key
|
||||||
|
api_base="https://api.aimlapi.com/v1", # 👈 the URL has changed from v2 to v1
|
||||||
|
input="Your text string",
|
||||||
|
)
|
||||||
|
print(response)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
asyncio.run(main())
|
||||||
|
```
|
||||||
|
|
||||||
|
## Async Image Generation
|
||||||
|
|
||||||
|
```python
|
||||||
|
import asyncio
|
||||||
|
|
||||||
|
import litellm
|
||||||
|
|
||||||
|
|
||||||
|
async def main():
|
||||||
|
response = await litellm.aimage_generation(
|
||||||
|
model="openai/dall-e-3", # The model name must include prefix "openai" + the model name from ai/ml api
|
||||||
|
api_key="", # your aiml api-key
|
||||||
|
api_base="https://api.aimlapi.com/v1", # 👈 the URL has changed from v2 to v1
|
||||||
|
prompt="A cute baby sea otter",
|
||||||
|
)
|
||||||
|
print(response)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
asyncio.run(main())
|
||||||
|
```
|
|
@ -45,7 +45,7 @@ const sidebars = {
|
||||||
"proxy/health",
|
"proxy/health",
|
||||||
"proxy/debugging",
|
"proxy/debugging",
|
||||||
"proxy/spending_monitoring",
|
"proxy/spending_monitoring",
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
"proxy/demo",
|
"proxy/demo",
|
||||||
{
|
{
|
||||||
|
@ -142,7 +142,7 @@ const sidebars = {
|
||||||
"proxy/guardrails/secret_detection",
|
"proxy/guardrails/secret_detection",
|
||||||
"proxy/guardrails/custom_guardrail",
|
"proxy/guardrails/custom_guardrail",
|
||||||
"prompt_injection"
|
"prompt_injection"
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
type: "category",
|
type: "category",
|
||||||
|
@ -162,7 +162,6 @@ const sidebars = {
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
"proxy/caching",
|
"proxy/caching",
|
||||||
|
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
@ -181,6 +180,7 @@ const sidebars = {
|
||||||
"providers/openai_compatible",
|
"providers/openai_compatible",
|
||||||
"providers/azure",
|
"providers/azure",
|
||||||
"providers/azure_ai",
|
"providers/azure_ai",
|
||||||
|
"providers/aiml",
|
||||||
"providers/vertex",
|
"providers/vertex",
|
||||||
"providers/gemini",
|
"providers/gemini",
|
||||||
"providers/anthropic",
|
"providers/anthropic",
|
||||||
|
@ -328,7 +328,7 @@ const sidebars = {
|
||||||
description: "Learn how to load balance, route, and set fallbacks for your LLM requests",
|
description: "Learn how to load balance, route, and set fallbacks for your LLM requests",
|
||||||
slug: "/routing-load-balancing",
|
slug: "/routing-load-balancing",
|
||||||
},
|
},
|
||||||
items: ["routing", "scheduler", "proxy/load_balancing", "proxy/reliability", "proxy/timeout", "proxy/tag_routing", "proxy/provider_budget_routing", "wildcard_routing"],
|
items: ["routing", "scheduler", "proxy/load_balancing", "proxy/reliability", "proxy/timeout", "proxy/tag_routing", "proxy/provider_budget_routing", "wildcard_routing"],
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
type: "category",
|
type: "category",
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue