mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-22 20:02:25 +00:00
Summary: ### THIS DIFF ### PLAN ### CONTEXT ### DESIGN Test Plan: # Test Setup **Type checker and check that the build compiles** **Unit Tests** **E2E Tests** // Screenshots and videos | Before | After | |--| | … | … | # Monitoring Plan.
22 lines
524 B
YAML
22 lines
524 B
YAML
inference_providers:
|
|
- ollama
|
|
- fireworks
|
|
- together
|
|
- tgi
|
|
- vllm
|
|
|
|
test_models:
|
|
text: meta-llama/Llama-3.1-8B-Instruct
|
|
vision: meta-llama/Llama-3.2-11B-Vision-Instruct
|
|
|
|
|
|
inference/test_vision_inference.py
|
|
test_vision_chat_completion_streaming
|
|
test_vision_chat_completion_non_streaming
|
|
|
|
inference/test_text_inference.py
|
|
test_structured_output
|
|
test_chat_completion_streaming
|
|
test_chat_completion_non_streaming
|
|
test_chat_completion_with_tool_calling
|
|
test_chat_completion_with_tool_calling_streaming
|