mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 18:00:36 +00:00
This patch ensures if max tokens is not defined, then is set to None instead of 0 when calling openai_chat_completion. This way some providers (like gemini) that cannot handle the `max_tokens = 0` will not fail Issue: #3666 |
||
|---|---|---|
| .. | ||
| img | ||
| providers/vector_io | ||
| deprecated-llama-stack-spec.html | ||
| deprecated-llama-stack-spec.yaml | ||
| experimental-llama-stack-spec.html | ||
| experimental-llama-stack-spec.yaml | ||
| llama-stack-spec.html | ||
| llama-stack-spec.yaml | ||
| remote_or_local.gif | ||
| safety_system.webp | ||
| site.webmanifest | ||
| stainless-llama-stack-spec.html | ||
| stainless-llama-stack-spec.yaml | ||