Krish Dholakia
|
b82add11ba
|
LITELLM: Remove requests library usage (#7235)
* fix(generic_api_callback.py): remove requests lib usage
* fix(budget_manager.py): remove requests lib usgae
* fix(main.py): cleanup requests lib usage
* fix(utils.py): remove requests lib usage
* fix(argilla.py): fix argilla test
* fix(athina.py): replace 'requests' lib usage with litellm module
* fix(greenscale.py): replace 'requests' lib usage with httpx
* fix: remove unused 'requests' lib import + replace usage in some places
* fix(prompt_layer.py): remove 'requests' lib usage from prompt layer
* fix(ollama_chat.py): remove 'requests' lib usage
* fix(baseten.py): replace 'requests' lib usage
* fix(codestral/): replace 'requests' lib usage
* fix(predibase/): replace 'requests' lib usage
* refactor: cleanup unused 'requests' lib imports
* fix(oobabooga.py): cleanup 'requests' lib usage
* fix(invoke_handler.py): remove unused 'requests' lib usage
* refactor: cleanup unused 'requests' lib import
* fix: fix linting errors
* refactor(ollama/): move ollama to using base llm http handler
removes 'requests' lib dep for ollama integration
* fix(ollama_chat.py): fix linting errors
* fix(ollama/completion/transformation.py): convert non-jpeg/png image to jpeg/png before passing to ollama
|
2024-12-17 12:50:04 -08:00 |
|
Ishaan Jaff
|
b5d55688e5
|
(Refactor) Code Quality improvement - remove /prompt_templates/ , base_aws_llm.py from /llms folder (#7164)
* fix move base_aws_llm
* fix import
* update enforce llms folder style
* move prompt_templates
* update prompt_templates location
* fix imports
* fix imports
* fix imports
* fix imports
* fix checks
|
2024-12-11 00:02:46 -08:00 |
|
Krish Dholakia
|
350cfc36f7
|
Litellm merge pr (#7161)
* build: merge branch
* test: fix openai naming
* fix(main.py): fix openai renaming
* style: ignore function length for config factory
* fix(sagemaker/): fix routing logic
* fix: fix imports
* fix: fix override
|
2024-12-10 22:49:26 -08:00 |
|
Krish Dholakia
|
e903fe6038
|
refactor(sagemaker/): separate chat + completion routes + make them b… (#7151)
* refactor(sagemaker/): separate chat + completion routes + make them both use base llm config
Addresses https://github.com/andrewyng/aisuite/issues/113#issuecomment-2512369132
* fix(main.py): pass hf model name + custom prompt dict to litellm params
|
2024-12-10 19:40:05 -08:00 |
|