llama-stack/llama_stack/providers/inline/inference
2025-01-22 18:31:59 -08:00
..
meta_reference Fix meta-reference GPU implementation for inference 2025-01-22 18:31:59 -08:00
sentence_transformers remove conflicting default for tool prompt format in chat completion (#742) 2025-01-10 10:41:53 -08:00
vllm Convert SamplingParams.strategy to a union (#767) 2025-01-15 05:38:51 -08:00
__init__.py precommit 2024-11-08 17:58:58 -08:00