mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
122 lines
3.6 KiB
TypeScript
122 lines
3.6 KiB
TypeScript
import React, { useState, useEffect } from "react";
|
|
import { Card, Title, Subtitle, Table, TableHead, TableRow, TableCell, TableBody, Metric, Grid } from "@tremor/react";
|
|
import { modelInfoCall } from "./networking";
|
|
|
|
interface ModelDashboardProps {
|
|
accessToken: string;
|
|
token: string | null;
|
|
userRole: string | null;
|
|
userID: string | null;
|
|
}
|
|
|
|
const ModelDashboard: React.FC<ModelDashboardProps> = ({
|
|
accessToken,
|
|
token,
|
|
userRole,
|
|
userID,
|
|
}) => {
|
|
const [modelData, setModelData] = useState<any>({ data: [] });
|
|
|
|
useEffect(() => {
|
|
const fetchData = async () => {
|
|
try {
|
|
// Replace with your actual API call for model data
|
|
const modelDataResponse = await modelInfoCall(accessToken, userID, userRole);
|
|
console.log("Model data response:", modelDataResponse.data);
|
|
setModelData(modelDataResponse);
|
|
} catch (error) {
|
|
console.error("There was an error fetching the model data", error);
|
|
}
|
|
};
|
|
|
|
if (accessToken && token && userRole && userID) {
|
|
fetchData();
|
|
}
|
|
}, [accessToken, token, userRole, userID]);
|
|
|
|
if (!modelData) {
|
|
return <div>Loading...</div>;
|
|
}
|
|
|
|
// loop through model data and edit each row
|
|
for (let i = 0; i < modelData.data.length; i++) {
|
|
let curr_model = modelData.data[i];
|
|
let litellm_model_name = curr_model?.litellm_params?.model;
|
|
|
|
let model_info = curr_model?.model_info;
|
|
|
|
let defaultProvider = "openai";
|
|
let provider = "";
|
|
let input_cost = "Undefined"
|
|
let output_cost = "Undefined"
|
|
let max_tokens = "Undefined"
|
|
|
|
// Check if litellm_model_name is null or undefined
|
|
if (litellm_model_name) {
|
|
// Split litellm_model_name based on "/"
|
|
let splitModel = litellm_model_name.split("/");
|
|
|
|
// Get the first element in the split
|
|
let firstElement = splitModel[0];
|
|
|
|
// If there is only one element, default provider to openai
|
|
provider = splitModel.length === 1 ? defaultProvider : firstElement;
|
|
|
|
console.log("Provider:", provider);
|
|
} else {
|
|
// litellm_model_name is null or undefined, default provider to openai
|
|
provider = defaultProvider;
|
|
console.log("Provider:", provider);
|
|
}
|
|
|
|
if (model_info) {
|
|
input_cost = model_info?.input_cost_per_token;
|
|
output_cost = model_info?.output_cost_per_token;
|
|
max_tokens = model_info?.max_tokens;
|
|
|
|
}
|
|
modelData.data[i].provider = provider
|
|
modelData.data[i].input_cost = input_cost
|
|
modelData.data[i].output_cost = output_cost
|
|
modelData.data[i].max_tokens = max_tokens
|
|
|
|
}
|
|
|
|
|
|
return (
|
|
<div style={{ width: "100%" }}>
|
|
<Grid className="gap-2 p-10 h-[75vh] w-full">
|
|
<Card>
|
|
<Title>Available Models</Title>
|
|
<Table className="mt-5">
|
|
<TableHead>
|
|
<TableRow>
|
|
<TableCell>Model Name</TableCell>
|
|
<TableCell>Provider</TableCell>
|
|
<TableCell>Input Price per token ($)</TableCell>
|
|
<TableCell>Output Price per token ($)</TableCell>
|
|
<TableCell>Max Tokens</TableCell>
|
|
</TableRow>
|
|
</TableHead>
|
|
<TableBody>
|
|
{modelData.data.map((model: any) => (
|
|
<TableRow key={model.model_name}>
|
|
|
|
<TableCell><p>{model.model_name}</p></TableCell>
|
|
<TableCell>{model.provider}</TableCell>
|
|
<TableCell>{model.input_cost}</TableCell>
|
|
<TableCell>{model.output_cost}</TableCell>
|
|
<TableCell>{model.max_tokens}</TableCell>
|
|
|
|
|
|
</TableRow>
|
|
))}
|
|
</TableBody>
|
|
</Table>
|
|
</Card>
|
|
</Grid>
|
|
</div>
|
|
);
|
|
};
|
|
|
|
export default ModelDashboard;
|