forked from phoenix-oss/llama-stack-mirror
fix: OAI compat endpoint for meta reference inference provider (#1962)
Test plan: python tests/verifications/generate_report.py --providers fireworks,together,llama_meta_ref,openai Co-authored-by: Eric Huang <erichuang@fb.com>
This commit is contained in:
parent
8bd6665775
commit
2976b5d992
8 changed files with 1184 additions and 44 deletions
|
@ -515,7 +515,8 @@ class MetaReferenceInferenceImpl(
|
|||
stop_reason = None
|
||||
ipython = False
|
||||
|
||||
for token_result in self.generator.chat_completion(request):
|
||||
for token_results in self.generator.chat_completion([request]):
|
||||
token_result = token_results[0]
|
||||
if os.environ.get("LLAMA_MODELS_DEBUG", "0") == "1":
|
||||
cprint(token_result.text, "cyan", end="")
|
||||
if os.environ.get("LLAMA_MODELS_DEBUG", "0") == "2":
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue