forked from phoenix/litellm-mirror
Merge pull request #1809 from BerriAI/litellm_embedding_caching_updates
Support caching individual items in embedding list (Async embedding only)
This commit is contained in:
commit
28df60b609
13 changed files with 638 additions and 196 deletions
|
@ -1994,6 +1994,7 @@ def test_completion_palm_stream():
|
|||
|
||||
|
||||
def test_completion_together_ai_stream():
|
||||
litellm.set_verbose = True
|
||||
user_message = "Write 1pg about YC & litellm"
|
||||
messages = [{"content": user_message, "role": "user"}]
|
||||
try:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue