llama-stack-mirror/client-sdks/stainless
Luis Tomas Bolivar f18b5eb537
fix: Avoid BadRequestError due to invalid max_tokens (#3667)
This patch ensures if max tokens is not defined, then is set to None
instead of 0 when calling openai_chat_completion. This way some
providers (like gemini) that cannot handle the `max_tokens = 0` will not
fail

Issue: #3666
2025-10-27 09:27:21 -07:00
..
openapi.stainless.yml chore: add beta group to stainless (#3866) 2025-10-20 16:26:06 -07:00
openapi.yml fix: Avoid BadRequestError due to invalid max_tokens (#3667) 2025-10-27 09:27:21 -07:00
README.md feat(stainless): add stainless source of truth config (#3860) 2025-10-20 14:32:20 -07:00

These are the source-of-truth configuration files used to generate the Stainless client SDKs via Stainless.

  • openapi.yml: this is the OpenAPI specification for the Llama Stack API.
  • openapi.stainless.yml: this is the Stainless configuration which instructs Stainless how to generate the client SDKs.

A small side note: notice the .yml suffixes since Stainless uses that suffix typically for its configuration files.

These files go hand-in-hand. As of now, only the openapi.yml file is automatically generated using the run_openapi_generator.sh script.