--- title: v1.57.7 slug: v1.57.7 date: 2025-01-10T10:00:00 authors: - name: Krrish Dholakia title: CEO, LiteLLM url: https://www.linkedin.com/in/krish-d/ image_url: https://media.licdn.com/dms/image/v2/D4D03AQGrlsJ3aqpHmQ/profile-displayphoto-shrink_400_400/B4DZSAzgP7HYAg-/0/1737327772964?e=1749686400&v=beta&t=Hkl3U8Ps0VtvNxX0BNNq24b4dtX5wQaPFp6oiKCIHD8 - name: Ishaan Jaffer title: CTO, LiteLLM url: https://www.linkedin.com/in/reffajnaahsi/ image_url: https://media.licdn.com/dms/image/v2/D4D03AQGiM7ZrUwqu_Q/profile-displayphoto-shrink_800_800/profile-displayphoto-shrink_800_800/0/1675971026692?e=1741824000&v=beta&t=eQnRdXPJo4eiINWTZARoYTfqh064pgZ-E21pQTSy8jc tags: [langfuse, management endpoints, ui, prometheus, secret management] hide_table_of_contents: false --- `langfuse`, `management endpoints`, `ui`, `prometheus`, `secret management` ## Langfuse Prompt Management Langfuse Prompt Management is being labelled as BETA. This allows us to iterate quickly on the feedback we're receiving, and making the status clearer to users. We expect to make this feature to be stable by next month (February 2025). Changes: - Include the client message in the LLM API Request. (Previously only the prompt template was sent, and the client message was ignored). - Log the prompt template in the logged request (e.g. to s3/langfuse). - Log the 'prompt_id' and 'prompt_variables' in the logged request (e.g. to s3/langfuse). [Start Here](https://docs.litellm.ai/docs/proxy/prompt_management) ## Team/Organization Management + UI Improvements Managing teams and organizations on the UI is now easier. Changes: - Support for editing user role within team on UI. - Support updating team member role to admin via api - `/team/member_update` - Show team admins all keys for their team. - Add organizations with budgets - Assign teams to orgs on the UI - Auto-assign SSO users to teams [Start Here](https://docs.litellm.ai/docs/proxy/self_serve) ## Hashicorp Vault Support We now support writing LiteLLM Virtual API keys to Hashicorp Vault. [Start Here](https://docs.litellm.ai/docs/proxy/vault) ## Custom Prometheus Metrics Define custom prometheus metrics, and track usage/latency/no. of requests against them This allows for more fine-grained tracking - e.g. on prompt template passed in request metadata [Start Here](https://docs.litellm.ai/docs/proxy/prometheus#beta-custom-metrics)