mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-01 12:08:39 +00:00
# What does this PR do? current passthrough impl returns chatcompletion_message.content as a TextItem() , not a straight string. so it's not compatible with other providers, and causes parsing error downstream. change away from the generic pydantic conversion, and explicitly parse out content.text ## Test Plan setup llama server with passthrough ``` llama-stack-client eval run-benchmark "MMMU_Pro_standard" --model-id meta-llama/Llama-3-8B --output-dir /tmp/ --num-examples 20 ``` works without parsing error |
||
---|---|---|
.. | ||
anthropic | ||
bedrock | ||
cerebras | ||
databricks | ||
fireworks | ||
gemini | ||
groq | ||
nvidia | ||
ollama | ||
openai | ||
passthrough | ||
runpod | ||
sambanova | ||
tgi | ||
together | ||
vllm | ||
__init__.py |