litellm-mirror/docs/my-website
Krish Dholakia 1a4910f6c0 fix(health.md): add rerank model health check information (#7295)
* fix(health.md): add rerank model health check information

* build(model_prices_and_context_window.json): add gemini 2.0 for google ai studio - pricing + commercial rate limits

* build(model_prices_and_context_window.json): add gemini-2.0 supports audio output = true

* docs(team_model_add.md): clarify allowing teams to add models is an enterprise feature

* fix(o1_transformation.py): add support for 'n', 'response_format' and 'stop' params for o1 and 'stream_options' param for o1-mini

* build(model_prices_and_context_window.json): add 'supports_system_message' to supporting openai models

needed as o1-preview, and o1-mini models don't support 'system message

* fix(o1_transformation.py): translate system message based on if o1 model supports it

* fix(o1_transformation.py): return 'stream' param support if o1-mini/o1-preview

o1 currently doesn't support streaming, but the other model versions do

Fixes https://github.com/BerriAI/litellm/issues/7292

* fix(o1_transformation.py): return tool calling/response_format in supported params if model map says so

Fixes https://github.com/BerriAI/litellm/issues/7292

* fix: fix linting errors

* fix: update '_transform_messages'

* fix(o1_transformation.py): fix provider passed for supported param checks

* test(base_llm_unit_tests.py): skip test if api takes >5s to respond

* fix(utils.py): return false in 'supports_factory' if can't find value

* fix(o1_transformation.py): always return stream + stream_options as supported params + handle stream options being passed in for azure o1

* feat(openai.py): support stream faking natively in openai handler

Allows o1 calls to be faked for just the "o1" model, allows native streaming for o1-mini, o1-preview

 Fixes https://github.com/BerriAI/litellm/issues/7292

* fix(openai.py): use inference param instead of original optional param
2024-12-18 19:18:10 -08:00
..
blog/2021-08-26-welcome Update index.md 2023-10-21 12:22:41 +05:30
docs fix(health.md): add rerank model health check information (#7295) 2024-12-18 19:18:10 -08:00
img (feat) Add Azure Blob Storage Logging Integration (#7265) 2024-12-16 22:18:22 -08:00
src bye (#6982) 2024-12-05 13:38:10 -08:00
static v1 2023-08-17 15:31:20 -07:00
.gitignore Add docs to export logs to Laminar (#6674) 2024-11-11 12:15:47 -08:00
babel.config.js updating docs 2023-08-12 11:30:32 -07:00
Dockerfile (docs) new dockerfile for litellm proxy 2023-11-17 17:39:07 -08:00
docusaurus.config.js (docs) simplify left nav names + use a section for making llm requests (#6799) 2024-11-18 12:53:43 -08:00
index.md fix keys 2023-08-17 16:13:52 -07:00
package-lock.json fix(main.py): fix retries being multiplied when using openai sdk (#7221) 2024-12-14 11:56:55 -08:00
package.json bye (#6982) 2024-12-05 13:38:10 -08:00
README.md updating docs 2023-08-12 11:30:32 -07:00
sidebars.js fix(main.py): fix retries being multiplied when using openai sdk (#7221) 2024-12-14 11:56:55 -08:00

Website

This website is built using Docusaurus 2, a modern static website generator.

Installation

$ yarn

Local Development

$ yarn start

This command starts a local development server and opens up a browser window. Most changes are reflected live without having to restart the server.

Build

$ yarn build

This command generates static content into the build directory and can be served using any static contents hosting service.

Deployment

Using SSH:

$ USE_SSH=true yarn deploy

Not using SSH:

$ GIT_USER=<Your GitHub username> yarn deploy

If you are using GitHub pages for hosting, this command is a convenient way to build the website and push to the gh-pages branch.