llama-stack/llama_stack/providers/remote/inference
Yuan Tang f9ca441974
chore: Link to Groq docs in the warning message for preview model (#1060)
This should be `llama-3.2-3b` instead of `llama-3.2-3b-instruct`.
2025-02-13 12:14:57 -05:00
..
bedrock perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
cerebras perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
databricks perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
fireworks perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
groq chore: Link to Groq docs in the warning message for preview model (#1060) 2025-02-13 12:14:57 -05:00
nvidia feat: Add a new template for dell (#978) 2025-02-06 14:14:39 -08:00
ollama perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
runpod perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
sambanova perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
sample [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
tgi perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
together perf: ensure ToolCall in ChatCompletionResponse is subset of ChatCompletionRequest.tools (#1041) 2025-02-11 18:31:35 -08:00
vllm fix: logprobs support in remote-vllm provider (#1074) 2025-02-13 11:00:00 -05:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00