forked from phoenix-oss/llama-stack-mirror
File "/usr/local/lib/python3.10/site-packages/llama_stack/providers/adapters/inference/vllm/vllm.py", line 136, in _stream_chat_completion async for chunk in process_chat_completion_stream_response( TypeError: process_chat_completion_stream_response() takes 2 positional arguments but 3 were given This corrects the error by deleting the request variable |
||
---|---|---|
.. | ||
bedrock | ||
databricks | ||
fireworks | ||
ollama | ||
sample | ||
tgi | ||
together | ||
vllm | ||
__init__.py |