mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-24 02:48:04 +00:00
fix: llama stack run starter in conda
# What does this PR do? ## Test Plan
This commit is contained in:
parent
84fa83b788
commit
c35a4a884e
2 changed files with 40 additions and 34 deletions
|
|
@ -197,7 +197,7 @@ def llama_stack_client(request, provider_data):
|
|||
server_process = start_llama_stack_server(config_name)
|
||||
|
||||
# Wait for server to be ready
|
||||
if not wait_for_server_ready(base_url, timeout=30, process=server_process):
|
||||
if not wait_for_server_ready(base_url, timeout=120, process=server_process):
|
||||
print("Server failed to start within timeout")
|
||||
server_process.terminate()
|
||||
raise RuntimeError(
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue