llama-stack/llama_stack/providers/utils/inference
Dinesh Yeduguru 787e2034b7
model registration in ollama and vllm check against the available models in the provider (#446)
tests:
pytest -v -s -m "ollama"
llama_stack/providers/tests/inference/test_text_inference.py

pytest -v -s -m vllm_remote
llama_stack/providers/tests/inference/test_text_inference.py --env
VLLM_URL="http://localhost:9798/v1"

---------
2024-11-13 13:04:06 -08:00
..
__init__.py Use inference APIs for executing Llama Guard (#121) 2024-09-28 15:40:06 -07:00
model_registry.py model registration in ollama and vllm check against the available models in the provider (#446) 2024-11-13 13:04:06 -08:00
openai_compat.py Enable vision models for (Together, Fireworks, Meta-Reference, Ollama) (#376) 2024-11-05 16:22:33 -08:00
prompt_adapter.py Inference to use provider resource id to register and validate (#428) 2024-11-12 20:02:00 -08:00