llama-stack-mirror/llama_stack/providers/remote/inference
Dinesh Yeduguru 787e2034b7
model registration in ollama and vllm check against the available models in the provider (#446)
tests:
pytest -v -s -m "ollama"
llama_stack/providers/tests/inference/test_text_inference.py

pytest -v -s -m vllm_remote
llama_stack/providers/tests/inference/test_text_inference.py --env
VLLM_URL="http://localhost:9798/v1"

---------
2024-11-13 13:04:06 -08:00
..
bedrock Inference to use provider resource id to register and validate (#428) 2024-11-12 20:02:00 -08:00
databricks Inference to use provider resource id to register and validate (#428) 2024-11-12 20:02:00 -08:00
fireworks Inference to use provider resource id to register and validate (#428) 2024-11-12 20:02:00 -08:00
ollama model registration in ollama and vllm check against the available models in the provider (#446) 2024-11-13 13:04:06 -08:00
sample migrate model to Resource and new registration signature (#410) 2024-11-08 16:12:57 -08:00
tgi migrate model to Resource and new registration signature (#410) 2024-11-08 16:12:57 -08:00
together Inference to use provider resource id to register and validate (#428) 2024-11-12 20:02:00 -08:00
vllm model registration in ollama and vllm check against the available models in the provider (#446) 2024-11-13 13:04:06 -08:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00