litellm-mirror/cookbook
Krish Dholakia 34bdf36eab
Add inference providers support for Hugging Face (#8258) (#9738) (#9773)
* Add inference providers support for Hugging Face (#8258)

* add first version of inference providers for huggingface

* temporarily skipping tests

* Add documentation

* Fix titles

* remove max_retries from params and clean up

* add suggestions

* use llm http handler

* update doc

* add suggestions

* run formatters

* add tests

* revert

* revert

* rename file

* set maxsize for lru cache

* fix embeddings

* fix inference url

* fix tests following breaking change in main

* use ChatCompletionRequest

* fix tests and lint

* [Hugging Face] Remove outdated chat completion tests and fix embedding tests (#9749)

* remove or fix tests

* fix link in doc

* fix(config_settings.md): document hf api key

---------

Co-authored-by: célina <hanouticelina@gmail.com>
2025-04-05 10:50:15 -07:00
..
benchmark (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
codellama-server (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
community-resources (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
litellm-ollama-docker-image build(deps): bump litellm in /cookbook/litellm-ollama-docker-image 2025-03-20 21:03:29 +00:00
litellm_proxy_server Update readme.md 2024-10-07 18:15:15 +05:30
litellm_router (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
litellm_router_load_test (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
logging_observability docs litellm x langfuse cookbook 2025-02-14 09:31:51 -08:00
misc fix dev release.txt 2025-04-01 12:02:51 -07:00
Benchmarking_LLMs_by_use_case.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
Claude_(Anthropic)_with_Streaming_liteLLM_Examples.ipynb (docs) update how to use litellm streaming 2023-11-21 11:19:32 -08:00
Evaluating_LLMs.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
liteLLM_A121_Jurrasic_example.ipynb add AI21 tutorial 2023-08-22 20:34:18 -07:00
LiteLLM_Azure_and_OpenAI_example.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
liteLLM_Baseten.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
LiteLLM_batch_completion.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
LiteLLM_Bedrock.ipynb (docs) bedrock cookbook 2023-11-06 18:00:52 -08:00
liteLLM_clarifai_Demo.ipynb Clarifai - Added streaming and async completion support 2024-05-03 14:03:38 +00:00
LiteLLM_Comparing_LLMs.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
LiteLLM_Completion_Cost.ipynb add completion costs notebook 2023-09-06 18:43:09 -07:00
liteLLM_function_calling.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
liteLLM_Getting_Started.ipynb getting started cookbook litellm 2023-09-23 10:38:59 -07:00
LiteLLM_HuggingFace.ipynb Add inference providers support for Hugging Face (#8258) (#9738) (#9773) 2025-04-05 10:50:15 -07:00
liteLLM_IBM_Watsonx.ipynb (docs) updated watsonx cookbook 2024-04-24 17:19:02 +02:00
liteLLM_Langchain_Demo.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
litellm_model_fallback.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
liteLLM_Ollama.ipynb docs 2023-09-21 12:33:39 -07:00
LiteLLM_OpenRouter.ipynb or docs 2023-09-20 21:43:03 -07:00
LiteLLM_Petals.ipynb cookbook petals 2023-09-20 09:34:33 -07:00
LiteLLM_PromptLayer.ipynb edit promptlayer cookbook 2023-08-26 13:37:08 -07:00
liteLLM_Replicate_Demo.ipynb fix replicate docs 2023-09-06 12:35:50 -07:00
liteLLM_Streaming_Demo.ipynb add streaming example cookbook 2023-08-09 14:29:23 -07:00
litellm_test_multiple_llm_demo.ipynb Rename Untitled155.ipynb to litellm_test_multiple_llm_demo.ipynb 2023-08-10 18:56:26 -07:00
litellm_Test_Multiple_Providers.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
LiteLLM_User_Based_Rate_Limits.ipynb add notebook on using litellm budget manager 2023-09-12 20:53:55 -07:00
liteLLM_VertextAI_Example.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
Migrating_to_LiteLLM_Proxy_from_OpenAI_Azure_OpenAI.ipynb build: cookbook on migrating to litellm proxy from openai/azure sdk 2024-07-27 16:09:50 -07:00
mlflow_langchain_tracing_litellm_proxy.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
Parallel_function_calling.ipynb (docs) parallel function calling with Azure OpenAI 2023-11-18 14:36:56 -08:00
Proxy_Batch_Users.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
result.html (docs) add embedding() profile 2023-11-30 19:04:51 -08:00
TogetherAI_liteLLM.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
Using_Nemo_Guardrails_with_LiteLLM_Server.ipynb (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
VLLM_Model_Testing.ipynb Add files via upload 2023-09-06 19:27:06 -07:00