forked from phoenix/litellm-mirror
resolve comments
This commit is contained in:
parent
fe8d59f5eb
commit
1442adbc33
2 changed files with 5 additions and 5 deletions
|
@ -1,12 +1,12 @@
|
||||||
# FriendliAI
|
# FriendliAI
|
||||||
https://friendli.ai/
|
https://suite.friendli.ai/
|
||||||
|
|
||||||
**We support ALL FriendliAI models, just set `friendli_ai/` as a prefix when sending completion requests**
|
**We support ALL FriendliAI models, just set `friendli_ai/` as a prefix when sending completion requests**
|
||||||
|
|
||||||
## API Key
|
## API Key
|
||||||
```python
|
```python
|
||||||
# env variable
|
# env variable
|
||||||
os.environ['FRIENDLI_AI_API_KEY']
|
os.environ['FRIENDLI_TOKEN']
|
||||||
```
|
```
|
||||||
|
|
||||||
## Sample Usage
|
## Sample Usage
|
||||||
|
@ -14,7 +14,7 @@ os.environ['FRIENDLI_AI_API_KEY']
|
||||||
from litellm import completion
|
from litellm import completion
|
||||||
import os
|
import os
|
||||||
|
|
||||||
os.environ['FRIENDLI_AI_API_KEY'] = ""
|
os.environ['FRIENDLI_TOKEN'] = ""
|
||||||
response = completion(
|
response = completion(
|
||||||
model="friendli_ai/mixtral-8x7b-instruct-v0-1",
|
model="friendli_ai/mixtral-8x7b-instruct-v0-1",
|
||||||
messages=[
|
messages=[
|
||||||
|
@ -29,7 +29,7 @@ print(response)
|
||||||
from litellm import completion
|
from litellm import completion
|
||||||
import os
|
import os
|
||||||
|
|
||||||
os.environ['FRIENDLI_AI_API_KEY'] = ""
|
os.environ['FRIENDLI_TOKEN'] = ""
|
||||||
response = completion(
|
response = completion(
|
||||||
model="friendli_ai/mixtral-8x7b-instruct-v0-1",
|
model="friendli_ai/mixtral-8x7b-instruct-v0-1",
|
||||||
messages=[
|
messages=[
|
||||||
|
|
|
@ -4288,7 +4288,7 @@ def speech(
|
||||||
response: Optional[HttpxBinaryResponseContent] = None
|
response: Optional[HttpxBinaryResponseContent] = None
|
||||||
if custom_llm_provider == "openai":
|
if custom_llm_provider == "openai":
|
||||||
api_base = (
|
api_base = (
|
||||||
api_base # for deepinfra/perplexity/anyscale/groq/friendli we check in get_llm_provider and pass in the api base from there
|
api_base # for deepinfra/perplexity/anyscale/groq/friendli_ai we check in get_llm_provider and pass in the api base from there
|
||||||
or litellm.api_base
|
or litellm.api_base
|
||||||
or get_secret("OPENAI_API_BASE")
|
or get_secret("OPENAI_API_BASE")
|
||||||
or "https://api.openai.com/v1"
|
or "https://api.openai.com/v1"
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue