llama-stack-mirror/llama_stack/providers/adapters/inference
Steve Grubb 122793ab92
Correct a traceback in vllm (#366)
File "/usr/local/lib/python3.10/site-packages/llama_stack/providers/adapters/inference/vllm/vllm.py", line 136, in _stream_chat_completion
async for chunk in process_chat_completion_stream_response(
TypeError: process_chat_completion_stream_response() takes 2 positional arguments but 3 were given

This corrects the error by deleting the request variable
2024-11-04 20:49:35 -08:00
..
bedrock fix bedrock impl (#359) 2024-11-03 07:32:30 -08:00
databricks completion() for tgi (#295) 2024-10-24 16:02:41 -07:00
fireworks Significantly simpler and malleable test setup (#360) 2024-11-04 17:36:43 -08:00
ollama completion() for tgi (#295) 2024-10-24 16:02:41 -07:00
sample Remove "routing_table" and "routing_key" concepts for the user (#201) 2024-10-10 10:24:13 -07:00
tgi Avoid warnings from pydantic for overriding schema 2024-10-28 21:39:48 -07:00
together Significantly simpler and malleable test setup (#360) 2024-11-04 17:36:43 -08:00
vllm Correct a traceback in vllm (#366) 2024-11-04 20:49:35 -08:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00