Ishaan Jaff
|
12639b7ccf
|
fix sagemaker streaming error
|
2025-03-24 21:29:29 -07:00 |
|
Ishaan Jaff
|
e2d612efd9
|
Bug fix - String data: stripped from entire content in streamed Gemini responses (#9070)
* _strip_sse_data_from_chunk
* use _strip_sse_data_from_chunk
* use _strip_sse_data_from_chunk
* use _strip_sse_data_from_chunk
* _strip_sse_data_from_chunk
* test_strip_sse_data_from_chunk
* _strip_sse_data_from_chunk
* testing
* _strip_sse_data_from_chunk
|
2025-03-07 21:06:39 -08:00 |
|
Ishaan Jaff
|
7a5dd29fe0
|
(fix) unable to pass input_type parameter to Voyage AI embedding mode (#7276)
Read Version from pyproject.toml / read-version (push) Successful in 46s
* VoyageEmbeddingConfig
* fix voyage logic to get params
* add voyage embedding transformation
* add get_provider_embedding_config
* use BaseEmbeddingConfig
* voyage clean up
* use llm http handler for embedding transformations
* test_voyage_ai_embedding_extra_params
* add voyage async
* test_voyage_ai_embedding_extra_params
* add async for llm http handler
* update BaseLLMEmbeddingTest
* test_voyage_ai_embedding_extra_params
* fix linting
* fix get_provider_embedding_config
* fix anthropic text test
* update location of base/chat/transformation
* fix import path
* fix IBMWatsonXAIConfig
|
2024-12-17 19:23:49 -08:00 |
|
Krish Dholakia
|
e903fe6038
|
refactor(sagemaker/): separate chat + completion routes + make them b… (#7151)
* refactor(sagemaker/): separate chat + completion routes + make them both use base llm config
Addresses https://github.com/andrewyng/aisuite/issues/113#issuecomment-2512369132
* fix(main.py): pass hf model name + custom prompt dict to litellm params
|
2024-12-10 19:40:05 -08:00 |
|