mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 12:07:34 +00:00
clean up
This commit is contained in:
parent
a6baa7b3d4
commit
0b9480a0e3
5 changed files with 0 additions and 18 deletions
|
@ -6,7 +6,6 @@
|
|||
|
||||
import base64
|
||||
import struct
|
||||
import time
|
||||
|
||||
import pytest
|
||||
from openai import OpenAI
|
||||
|
@ -14,13 +13,6 @@ from openai import OpenAI
|
|||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||
|
||||
|
||||
@pytest.fixture(autouse=True)
|
||||
def rate_limit_delay():
|
||||
"""Add delay between tests to avoid rate limiting from providers like Fireworks"""
|
||||
yield
|
||||
time.sleep(30) # 30 second delay after each test
|
||||
|
||||
|
||||
def decode_base64_to_floats(base64_string: str) -> list[float]:
|
||||
"""Helper function to decode base64 string to list of float32 values."""
|
||||
embedding_bytes = base64.b64decode(base64_string)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue