litellm/docs/my-website
Krish Dholakia 7e5085dc7b
Litellm dev 11 21 2024 (#6837)
* Fix Vertex AI function calling invoke: use JSON format instead of protobuf text format. (#6702)

* test: test tool_call conversion when arguments is empty dict

Fixes https://github.com/BerriAI/litellm/issues/6833

* fix(openai_like/handler.py): return more descriptive error message

Fixes https://github.com/BerriAI/litellm/issues/6812

* test: skip overloaded model

* docs(anthropic.md): update anthropic docs to show how to route to any new model

* feat(groq/): fake stream when 'response_format' param is passed

Groq doesn't support streaming when response_format is set

* feat(groq/): add response_format support for groq

Closes https://github.com/BerriAI/litellm/issues/6845

* fix(o1_handler.py): remove fake streaming for o1

Closes https://github.com/BerriAI/litellm/issues/6801

* build(model_prices_and_context_window.json): add groq llama3.2b model pricing

Closes https://github.com/BerriAI/litellm/issues/6807

* fix(utils.py): fix handling ollama response format param

Fixes https://github.com/BerriAI/litellm/issues/6848#issuecomment-2491215485

* docs(sidebars.js): refactor chat endpoint placement

* fix: fix linting errors

* test: fix test

* test: fix test

* fix(openai_like/handler): handle max retries

* fix(streaming_handler.py): fix streaming check for openai-compatible providers

* test: update test

* test: correctly handle model is overloaded error

* test: update test

* test: fix test

* test: mark flaky test

---------

Co-authored-by: Guowang Li <Guowang@users.noreply.github.com>
2024-11-22 01:53:52 +05:30
..
blog/2021-08-26-welcome Update index.md 2023-10-21 12:22:41 +05:30
docs Litellm dev 11 21 2024 (#6837) 2024-11-22 01:53:52 +05:30
img Litellm dev 11 20 2024 (#6838) 2024-11-21 05:20:37 +05:30
src remove ask mode (#6271) 2024-10-16 22:01:50 -07:00
static v1 2023-08-17 15:31:20 -07:00
.gitignore Add docs to export logs to Laminar (#6674) 2024-11-11 12:15:47 -08:00
babel.config.js updating docs 2023-08-12 11:30:32 -07:00
Dockerfile (docs) new dockerfile for litellm proxy 2023-11-17 17:39:07 -08:00
docusaurus.config.js (docs) simplify left nav names + use a section for making llm requests (#6799) 2024-11-18 12:53:43 -08:00
index.md fix keys 2023-08-17 16:13:52 -07:00
package-lock.json build(deps): bump cookie and express in /docs/my-website (#6566) 2024-11-04 07:22:54 -08:00
package.json bump (#6187) 2024-10-14 18:22:54 +05:30
README.md updating docs 2023-08-12 11:30:32 -07:00
sidebars.js Litellm dev 11 21 2024 (#6837) 2024-11-22 01:53:52 +05:30

Website

This website is built using Docusaurus 2, a modern static website generator.

Installation

$ yarn

Local Development

$ yarn start

This command starts a local development server and opens up a browser window. Most changes are reflected live without having to restart the server.

Build

$ yarn build

This command generates static content into the build directory and can be served using any static contents hosting service.

Deployment

Using SSH:

$ USE_SSH=true yarn deploy

Not using SSH:

$ GIT_USER=<Your GitHub username> yarn deploy

If you are using GitHub pages for hosting, this command is a convenient way to build the website and push to the gh-pages branch.