mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-05 04:17:32 +00:00
bug fixes to make this work, trace creation worked - spans dont yet
This commit is contained in:
parent
84ebed9c9f
commit
6e5ca1350e
3 changed files with 25 additions and 10 deletions
|
@ -6,19 +6,20 @@
|
|||
|
||||
import asyncio
|
||||
import json
|
||||
from typing import Any, AsyncGenerator
|
||||
from typing import Any, AsyncGenerator, List, Optional
|
||||
|
||||
import fire
|
||||
import httpx
|
||||
from pydantic import BaseModel
|
||||
|
||||
from llama_models.llama3.api import * # noqa: F403
|
||||
from llama_stack.apis.inference import * # noqa: F403
|
||||
from termcolor import cprint
|
||||
|
||||
from llama_stack.distribution.datatypes import RemoteProviderConfig
|
||||
|
||||
from .event_logger import EventLogger
|
||||
|
||||
from llama_stack.apis.inference import * # noqa: F403
|
||||
|
||||
|
||||
async def get_client_impl(config: RemoteProviderConfig, _deps: Any) -> Inference:
|
||||
return InferenceClient(config.url)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue