forked from phoenix/litellm-mirror
ui - clean up chat ui
This commit is contained in:
parent
fa0ef8271e
commit
0e323b507d
2 changed files with 4 additions and 124 deletions
|
@ -13,12 +13,12 @@ import {
|
||||||
TabGroup,
|
TabGroup,
|
||||||
TabList,
|
TabList,
|
||||||
TabPanel,
|
TabPanel,
|
||||||
|
TabPanels,
|
||||||
Metric,
|
Metric,
|
||||||
Col,
|
Col,
|
||||||
Text,
|
Text,
|
||||||
SelectItem,
|
SelectItem,
|
||||||
TextInput,
|
TextInput,
|
||||||
TabPanels,
|
|
||||||
Button,
|
Button,
|
||||||
} from "@tremor/react";
|
} from "@tremor/react";
|
||||||
|
|
||||||
|
@ -201,7 +201,6 @@ const ChatUI: React.FC<ChatUIProps> = ({
|
||||||
<TabGroup>
|
<TabGroup>
|
||||||
<TabList>
|
<TabList>
|
||||||
<Tab>Chat</Tab>
|
<Tab>Chat</Tab>
|
||||||
<Tab>API Reference</Tab>
|
|
||||||
</TabList>
|
</TabList>
|
||||||
|
|
||||||
<TabPanels>
|
<TabPanels>
|
||||||
|
@ -272,124 +271,7 @@ const ChatUI: React.FC<ChatUIProps> = ({
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</TabPanel>
|
</TabPanel>
|
||||||
<TabPanel>
|
|
||||||
<TabGroup>
|
|
||||||
<TabList>
|
|
||||||
<Tab>OpenAI Python SDK</Tab>
|
|
||||||
<Tab>LlamaIndex</Tab>
|
|
||||||
<Tab>Langchain Py</Tab>
|
|
||||||
</TabList>
|
|
||||||
<TabPanels>
|
|
||||||
<TabPanel>
|
|
||||||
<SyntaxHighlighter language="python">
|
|
||||||
{`
|
|
||||||
import openai
|
|
||||||
client = openai.OpenAI(
|
|
||||||
api_key="your_api_key",
|
|
||||||
base_url="http://0.0.0.0:4000" # proxy base url
|
|
||||||
)
|
|
||||||
|
|
||||||
response = client.chat.completions.create(
|
|
||||||
model="gpt-3.5-turbo", # model to use from Models Tab
|
|
||||||
messages = [
|
|
||||||
{
|
|
||||||
"role": "user",
|
|
||||||
"content": "this is a test request, write a short poem"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
extra_body={
|
|
||||||
"metadata": {
|
|
||||||
"generation_name": "ishaan-generation-openai-client",
|
|
||||||
"generation_id": "openai-client-gen-id22",
|
|
||||||
"trace_id": "openai-client-trace-id22",
|
|
||||||
"trace_user_id": "openai-client-user-id2"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
print(response)
|
|
||||||
`}
|
|
||||||
</SyntaxHighlighter>
|
|
||||||
</TabPanel>
|
|
||||||
<TabPanel>
|
|
||||||
<SyntaxHighlighter language="python">
|
|
||||||
{`
|
|
||||||
import os, dotenv
|
|
||||||
|
|
||||||
from llama_index.llms import AzureOpenAI
|
|
||||||
from llama_index.embeddings import AzureOpenAIEmbedding
|
|
||||||
from llama_index import VectorStoreIndex, SimpleDirectoryReader, ServiceContext
|
|
||||||
|
|
||||||
llm = AzureOpenAI(
|
|
||||||
engine="azure-gpt-3.5", # model_name on litellm proxy
|
|
||||||
temperature=0.0,
|
|
||||||
azure_endpoint="http://0.0.0.0:4000", # litellm proxy endpoint
|
|
||||||
api_key="sk-1234", # litellm proxy API Key
|
|
||||||
api_version="2023-07-01-preview",
|
|
||||||
)
|
|
||||||
|
|
||||||
embed_model = AzureOpenAIEmbedding(
|
|
||||||
deployment_name="azure-embedding-model",
|
|
||||||
azure_endpoint="http://0.0.0.0:4000",
|
|
||||||
api_key="sk-1234",
|
|
||||||
api_version="2023-07-01-preview",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
documents = SimpleDirectoryReader("llama_index_data").load_data()
|
|
||||||
service_context = ServiceContext.from_defaults(llm=llm, embed_model=embed_model)
|
|
||||||
index = VectorStoreIndex.from_documents(documents, service_context=service_context)
|
|
||||||
|
|
||||||
query_engine = index.as_query_engine()
|
|
||||||
response = query_engine.query("What did the author do growing up?")
|
|
||||||
print(response)
|
|
||||||
|
|
||||||
`}
|
|
||||||
</SyntaxHighlighter>
|
|
||||||
</TabPanel>
|
|
||||||
<TabPanel>
|
|
||||||
<SyntaxHighlighter language="python">
|
|
||||||
{`
|
|
||||||
from langchain.chat_models import ChatOpenAI
|
|
||||||
from langchain.prompts.chat import (
|
|
||||||
ChatPromptTemplate,
|
|
||||||
HumanMessagePromptTemplate,
|
|
||||||
SystemMessagePromptTemplate,
|
|
||||||
)
|
|
||||||
from langchain.schema import HumanMessage, SystemMessage
|
|
||||||
|
|
||||||
chat = ChatOpenAI(
|
|
||||||
openai_api_base="http://0.0.0.0:8000",
|
|
||||||
model = "gpt-3.5-turbo",
|
|
||||||
temperature=0.1,
|
|
||||||
extra_body={
|
|
||||||
"metadata": {
|
|
||||||
"generation_name": "ishaan-generation-langchain-client",
|
|
||||||
"generation_id": "langchain-client-gen-id22",
|
|
||||||
"trace_id": "langchain-client-trace-id22",
|
|
||||||
"trace_user_id": "langchain-client-user-id2"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
messages = [
|
|
||||||
SystemMessage(
|
|
||||||
content="You are a helpful assistant that im using to make a test request to."
|
|
||||||
),
|
|
||||||
HumanMessage(
|
|
||||||
content="test from litellm. tell me why it's amazing in 1 sentence"
|
|
||||||
),
|
|
||||||
]
|
|
||||||
response = chat(messages)
|
|
||||||
|
|
||||||
print(response)
|
|
||||||
|
|
||||||
`}
|
|
||||||
</SyntaxHighlighter>
|
|
||||||
</TabPanel>
|
|
||||||
</TabPanels>
|
|
||||||
</TabGroup>
|
|
||||||
</TabPanel>
|
|
||||||
</TabPanels>
|
</TabPanels>
|
||||||
</TabGroup>
|
</TabGroup>
|
||||||
</Card>
|
</Card>
|
||||||
|
|
|
@ -63,15 +63,13 @@ const Sidebar: React.FC<SidebarProps> = ({
|
||||||
Test Key
|
Test Key
|
||||||
</Text>
|
</Text>
|
||||||
</Menu.Item>
|
</Menu.Item>
|
||||||
{
|
|
||||||
userRole == "App User" ? (
|
|
||||||
<Menu.Item key="11" onClick={() => setPage("api_ref")}>
|
<Menu.Item key="11" onClick={() => setPage("api_ref")}>
|
||||||
<Text>
|
<Text>
|
||||||
API Reference
|
API Reference
|
||||||
</Text>
|
</Text>
|
||||||
</Menu.Item>
|
</Menu.Item>
|
||||||
) : null
|
|
||||||
}
|
|
||||||
{
|
{
|
||||||
userRole == "Admin" ? (
|
userRole == "Admin" ? (
|
||||||
<Menu.Item key="2" onClick={() => setPage("models")}>
|
<Menu.Item key="2" onClick={() => setPage("models")}>
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue