mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 12:07:34 +00:00
Fix passthrough streaming, send headers properly not part of body :facepalm
This commit is contained in:
parent
039861f1c7
commit
a0e61a3c7a
2 changed files with 36 additions and 45 deletions
|
@ -50,6 +50,7 @@ class InferenceClient(Inference):
|
|||
headers={"Content-Type": "application/json"},
|
||||
timeout=20,
|
||||
) as response:
|
||||
print("Headers", response.headers)
|
||||
if response.status_code != 200:
|
||||
content = await response.aread()
|
||||
cprint(
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue