llama-stack-mirror/llama_stack/providers/utils/inference
Ben Browning a4b573d750 Fix OpenAI API response format handling
This fixes the schema of OpenAI API chat completion response formats,
including how those response formats (and other nested parameters in
the chat completion request) get translated into paramters for calls
to the backend OpenAI-compatible providers.

Signed-off-by: Ben Browning <bbrownin@redhat.com>
2025-04-13 13:39:56 -04:00
..
__init__.py refactor: move all llama code to models/llama out of meta reference (#1887) 2025-04-07 15:03:58 -07:00
embedding_mixin.py fix: dont assume SentenceTransformer is imported 2025-02-25 16:53:01 -08:00
litellm_openai_mixin.py Fix OpenAI API response format handling 2025-04-13 13:39:56 -04:00
model_registry.py test: verification on provider's OAI endpoints (#1893) 2025-04-07 23:06:28 -07:00
openai_compat.py Fix OpenAI API response format handling 2025-04-13 13:39:56 -04:00
prompt_adapter.py refactor: move all llama code to models/llama out of meta reference (#1887) 2025-04-07 15:03:58 -07:00