mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-24 13:19:54 +00:00
# What does this PR do? - fix cerebras template ## Test Plan ``` llama stack build --template cerebras --image-type conda llama stack run cerebras LLAMA_STACK_BASE_URL="http://localhost:5000" pytest -v tests/client-sdk/ --html=report.html --self-contained-html ``` ## Sources Please link relevant resources if necessary. ## Before submitting - [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case). - [ ] Ran pre-commit to handle lint / formatting issues. - [ ] Read the [contributor guideline](https://github.com/meta-llama/llama-stack/blob/main/CONTRIBUTING.md), Pull Request section? - [ ] Updated relevant documentation. - [ ] Wrote necessary unit or integration tests.
30 lines
654 B
YAML
30 lines
654 B
YAML
version: '2'
|
|
name: cerebras
|
|
distribution_spec:
|
|
description: Use Cerebras for running LLM inference
|
|
providers:
|
|
inference:
|
|
- remote::cerebras
|
|
safety:
|
|
- inline::llama-guard
|
|
memory:
|
|
- inline::meta-reference
|
|
agents:
|
|
- inline::meta-reference
|
|
eval:
|
|
- inline::meta-reference
|
|
datasetio:
|
|
- remote::huggingface
|
|
- inline::localfs
|
|
scoring:
|
|
- inline::basic
|
|
- inline::llm-as-judge
|
|
- inline::braintrust
|
|
telemetry:
|
|
- inline::meta-reference
|
|
tool_runtime:
|
|
- remote::brave-search
|
|
- remote::tavily-search
|
|
- inline::code-interpreter
|
|
- inline::memory-runtime
|
|
image_type: conda
|