forked from phoenix-oss/llama-stack-mirror
# What does this PR do? This PR updates how the `AgentType` gets set using the radio button on the tools page of the playground. This change is needed due to the fact with its current implementation, the chat interface will resets after every input, preventing users from having a multi-turn conversation with the agent. ## Test Plan Run the Playground without these changes: ```bash streamlit run llama_stack/distribution/ui/app.py ``` Navigate to the tools page and attempt to have a multi-turn conversation. You should see the conversation reset after asking a second question. Repeat the steps above with these changes and you will see that it works as expected when asking the agent multiple questions. Signed-off-by: Michael Clifford <mcliffor@redhat.com> |
||
---|---|---|
.. | ||
modules | ||
page | ||
__init__.py | ||
app.py | ||
Containerfile | ||
README.md | ||
requirements.txt |
(Experimental) LLama Stack UI
Docker Setup
⚠️ This is a work in progress.
Developer Setup
- Start up Llama Stack API server. More details here.
llama stack build --template together --image-type conda
llama stack run together
- (Optional) Register datasets and eval tasks as resources. If you want to run pre-configured evaluation flows (e.g. Evaluations (Generation + Scoring) Page).
llama-stack-client datasets register \
--dataset-id "mmlu" \
--provider-id "huggingface" \
--url "https://huggingface.co/datasets/llamastack/evals" \
--metadata '{"path": "llamastack/evals", "name": "evals__mmlu__details", "split": "train"}' \
--schema '{"input_query": {"type": "string"}, "expected_answer": {"type": "string", "chat_completion_input": {"type": "string"}}}'
llama-stack-client benchmarks register \
--eval-task-id meta-reference-mmlu \
--provider-id meta-reference \
--dataset-id mmlu \
--scoring-functions basic::regex_parser_multiple_choice_answer
- Start Streamlit UI
uv run --with ".[ui]" streamlit run llama_stack/distribution/ui/app.py
Environment Variables
Environment Variable | Description | Default Value |
---|---|---|
LLAMA_STACK_ENDPOINT | The endpoint for the Llama Stack | http://localhost:8321 |
FIREWORKS_API_KEY | API key for Fireworks provider | (empty string) |
TOGETHER_API_KEY | API key for Together provider | (empty string) |
SAMBANOVA_API_KEY | API key for SambaNova provider | (empty string) |
OPENAI_API_KEY | API key for OpenAI provider | (empty string) |