llama-stack-mirror/llama_stack/providers/utils/inference
Yuan Tang a1da09e166
feat: Support "stop" parameter in remote:vLLM
Signed-off-by: Yuan Tang <terrytangyuan@gmail.com>
2025-03-19 22:41:34 -04:00
..
__init__.py chore: move all Llama Stack types from llama-models to llama-stack (#1098) 2025-02-14 09:10:59 -08:00
embedding_mixin.py fix: dont assume SentenceTransformer is imported 2025-02-25 16:53:01 -08:00
litellm_openai_mixin.py fix: agents with non-llama model (#1550) 2025-03-17 22:11:06 -07:00
model_registry.py feat(providers): support non-llama models for inference providers (#1200) 2025-02-21 13:21:28 -08:00
openai_compat.py feat: Support "stop" parameter in remote:vLLM 2025-03-19 22:41:34 -04:00
prompt_adapter.py feat(logging): implement category-based logging (#1362) 2025-03-07 11:34:30 -08:00