forked from phoenix/litellm-mirror
(docs) pass metatadata to proxy + openai client
This commit is contained in:
parent
5b9973136e
commit
8b29f9a48b
1 changed files with 18 additions and 6 deletions
|
@ -524,6 +524,8 @@ curl --location 'http://0.0.0.0:8000/chat/completions' \
|
||||||
</TabItem>
|
</TabItem>
|
||||||
<TabItem value="openai" label="OpenAI v1.0.0+">
|
<TabItem value="openai" label="OpenAI v1.0.0+">
|
||||||
|
|
||||||
|
Set `extra_body={"metadata": { }}` to `metadata` you want to pass
|
||||||
|
|
||||||
```python
|
```python
|
||||||
import openai
|
import openai
|
||||||
client = openai.OpenAI(
|
client = openai.OpenAI(
|
||||||
|
@ -532,15 +534,25 @@ client = openai.OpenAI(
|
||||||
)
|
)
|
||||||
|
|
||||||
# request sent to model set on litellm proxy, `litellm --model`
|
# request sent to model set on litellm proxy, `litellm --model`
|
||||||
response = client.chat.completions.create(model="gpt-3.5-turbo", messages = [
|
response = client.chat.completions.create(
|
||||||
{
|
model="gpt-3.5-turbo",
|
||||||
"role": "user",
|
messages = [
|
||||||
"content": "this is a test request, write a short poem"
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": "this is a test request, write a short poem"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
extra_body={
|
||||||
|
"metadata": {
|
||||||
|
"generation_name": "ishaan-generation-openai-client",
|
||||||
|
"generation_id": "openai-client-gen-id22",
|
||||||
|
"trace_id": "openai-client-trace-id22",
|
||||||
|
"trace_user_id": "openai-client-user-id2"
|
||||||
|
}
|
||||||
}
|
}
|
||||||
])
|
)
|
||||||
|
|
||||||
print(response)
|
print(response)
|
||||||
|
|
||||||
```
|
```
|
||||||
</TabItem>
|
</TabItem>
|
||||||
<TabItem value="langchain" label="Langchain">
|
<TabItem value="langchain" label="Langchain">
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue