mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-23 00:52:26 +00:00
Fix Dell distro and TGI inference
This commit is contained in:
parent
9b70bb9d4b
commit
ad66968909
2 changed files with 4 additions and 1 deletions
|
|
@ -309,7 +309,7 @@ class TGIAdapter(_HfAdapter):
|
|||
raise ValueError("You must provide a URL in run.yaml (or via the TGI_URL environment variable) to use TGI.")
|
||||
log.info(f"Initializing TGI client with url={config.url}")
|
||||
self.client = AsyncInferenceClient(
|
||||
model=config.url,
|
||||
model=config.url,provider="hf-inference"
|
||||
)
|
||||
endpoint_info = await self.client.get_endpoint_info()
|
||||
self.max_tokens = endpoint_info["max_total_tokens"]
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue