llama-stack-mirror/tests/integration/inference
Sébastien Han 73e99b6eab
fix: add token to the openai request
OpenAIMixin expects to use an API key and creates its own AsyncOpenAI
client. So our code now authenticate with the Google service, retrieves
a token and pass it to the OpenAI client.
Falls back to an empty string if credentials can't be obtained (letting
LiteLLM handle ADC directly).

Signed-off-by: Sébastien Han <seb@redhat.com>
2025-09-10 15:17:37 +02:00
..
__init__.py fix: remove ruff N999 (#1388) 2025-03-07 11:14:04 -08:00
dog.png refactor: tests/unittests -> tests/unit; tests/api -> tests/integration 2025-03-04 09:57:00 -08:00
test_batch_inference.py feat: add batch inference API to llama stack inference (#1945) 2025-04-12 11:41:12 -07:00
test_embedding.py fix: fix the error type in embedding test case (#3197) 2025-08-21 16:19:51 -07:00
test_openai_completion.py fix: add token to the openai request 2025-09-10 15:17:37 +02:00
test_openai_embeddings.py chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
test_text_inference.py feat: Add Google Vertex AI inference provider support (#2841) 2025-08-11 08:22:04 -04:00
test_vision_inference.py feat(ci): add support for running vision inference tests (#2972) 2025-07-31 11:50:42 -07:00
vision_test_1.jpg feat: introduce llama4 support (#1877) 2025-04-05 11:53:35 -07:00
vision_test_2.jpg feat: introduce llama4 support (#1877) 2025-04-05 11:53:35 -07:00
vision_test_3.jpg feat: introduce llama4 support (#1877) 2025-04-05 11:53:35 -07:00