litellm-mirror/litellm/llms/huggingface
Krish Dholakia c8aa876785
fix(proxy_server.py): fix get model info when litellm_model_id is set + move model analytics to free (#7886)
* fix(proxy_server.py): fix get model info when litellm_model_id is set

Fixes https://github.com/BerriAI/litellm/issues/7873

* test(test_models.py): add test to ensure get model info on specific deployment has same value as all model info

Fixes https://github.com/BerriAI/litellm/issues/7873

* fix(usage.tsx): make model analytics free

Fixes @iqballx's feedback

* fix(fix(invoke_handler.py):-fix-bedrock-error-chunk-parsing): return correct bedrock status code and error message if chunk in stream

Improves bedrock stream error handling

* fix(proxy_server.py): fix linting errors

* test(test_auth_checks.py): remove redundant test

* fix(proxy_server.py): fix linting errors

* test: fix flaky test

* test: fix test
2025-01-21 08:19:07 -08:00
..
chat fix(proxy_server.py): fix get model info when litellm_model_id is set + move model analytics to free (#7886) 2025-01-21 08:19:07 -08:00
huggingface_llms_metadata sec fix minor (#7810) 2025-01-16 22:03:28 -08:00
common_utils.py (fix) unable to pass input_type parameter to Voyage AI embedding mode (#7276) 2024-12-17 19:23:49 -08:00