llama-stack-mirror/llama_stack/providers/adapters/inference/vllm
2024-10-23 19:11:04 -07:00
..
__init__.py Add vLLM inference provider for OpenAI compatible vLLM server (#178) 2024-10-20 18:43:25 -07:00
config.py Add vLLM inference provider for OpenAI compatible vLLM server (#178) 2024-10-20 18:43:25 -07:00
vllm.py refactor get_max_tokens and build_options 2024-10-23 19:11:04 -07:00