autogen build/run

This commit is contained in:
Xi Yan 2024-12-03 17:04:35 -08:00
parent 95187891ca
commit 3fc6b10d22
33 changed files with 369 additions and 6 deletions

View file

@ -2,9 +2,11 @@
"hf-serverless": [ "hf-serverless": [
"aiohttp", "aiohttp",
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -13,6 +15,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -29,9 +32,11 @@
], ],
"together": [ "together": [
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -39,6 +44,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -56,9 +62,11 @@
], ],
"vllm-gpu": [ "vllm-gpu": [
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -66,6 +74,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -110,9 +119,11 @@
], ],
"fireworks": [ "fireworks": [
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -121,6 +132,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -138,9 +150,11 @@
"tgi": [ "tgi": [
"aiohttp", "aiohttp",
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -149,6 +163,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -165,10 +180,12 @@
], ],
"bedrock": [ "bedrock": [
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"boto3", "boto3",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -176,6 +193,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -193,9 +211,11 @@
"meta-reference-gpu": [ "meta-reference-gpu": [
"accelerate", "accelerate",
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"fairscale", "fairscale",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
@ -205,6 +225,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -225,9 +246,11 @@
"meta-reference-quantized-gpu": [ "meta-reference-quantized-gpu": [
"accelerate", "accelerate",
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"fairscale", "fairscale",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
@ -238,6 +261,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -259,9 +283,11 @@
"ollama": [ "ollama": [
"aiohttp", "aiohttp",
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -270,6 +296,7 @@
"nltk", "nltk",
"numpy", "numpy",
"ollama", "ollama",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",
@ -287,9 +314,11 @@
"hf-endpoint": [ "hf-endpoint": [
"aiohttp", "aiohttp",
"aiosqlite", "aiosqlite",
"autoevals",
"blobfile", "blobfile",
"chardet", "chardet",
"chromadb-client", "chromadb-client",
"datasets",
"faiss-cpu", "faiss-cpu",
"fastapi", "fastapi",
"fire", "fire",
@ -298,6 +327,7 @@
"matplotlib", "matplotlib",
"nltk", "nltk",
"numpy", "numpy",
"openai",
"pandas", "pandas",
"pillow", "pillow",
"psycopg2-binary", "psycopg2-binary",

View file

@ -1,6 +1,3 @@
---
orphan: true
---
# Bedrock Distribution # Bedrock Distribution
```{toctree} ```{toctree}
@ -15,9 +12,12 @@ The `llamastack/distribution-bedrock` distribution consists of the following pro
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `remote::bedrock` | | inference | `remote::bedrock` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `remote::bedrock` | | safety | `remote::bedrock` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |

View file

@ -15,9 +15,12 @@ The `llamastack/distribution-fireworks` distribution consists of the following p
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `remote::fireworks` | | inference | `remote::fireworks` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `inline::llama-guard` | | safety | `inline::llama-guard` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |

View file

@ -15,9 +15,12 @@ The `llamastack/distribution-meta-reference-gpu` distribution consists of the fo
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `inline::meta-reference` | | inference | `inline::meta-reference` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `inline::llama-guard` | | safety | `inline::llama-guard` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |

View file

@ -15,9 +15,12 @@ The `llamastack/distribution-meta-reference-quantized-gpu` distribution consists
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `inline::meta-reference-quantized` | | inference | `inline::meta-reference-quantized` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `inline::llama-guard` | | safety | `inline::llama-guard` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |

View file

@ -15,9 +15,12 @@ The `llamastack/distribution-ollama` distribution consists of the following prov
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `remote::ollama` | | inference | `remote::ollama` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `inline::llama-guard` | | safety | `inline::llama-guard` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |
@ -118,9 +121,9 @@ llama stack run ./run-with-safety.yaml \
### (Optional) Update Model Serving Configuration ### (Optional) Update Model Serving Configuration
```{note} > [!NOTE]
Please check the [model_aliases](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/remote/inference/ollama/ollama.py#L45) variable for supported Ollama models. > Please check the [OLLAMA_SUPPORTED_MODELS](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers.remote/inference/ollama/ollama.py) for the supported Ollama models.
```
To serve a new model with `ollama` To serve a new model with `ollama`
```bash ```bash

View file

@ -16,9 +16,12 @@ The `llamastack/distribution-tgi` distribution consists of the following provide
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `remote::tgi` | | inference | `remote::tgi` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `inline::llama-guard` | | safety | `inline::llama-guard` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |

View file

@ -15,9 +15,12 @@ The `llamastack/distribution-together` distribution consists of the following pr
| API | Provider(s) | | API | Provider(s) |
|-----|-------------| |-----|-------------|
| agents | `inline::meta-reference` | | agents | `inline::meta-reference` |
| datasetio | `remote::huggingface`, `inline::localfs` |
| eval | `inline::meta-reference` |
| inference | `remote::together` | | inference | `remote::together` |
| memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | | memory | `inline::faiss`, `remote::chromadb`, `remote::pgvector` |
| safety | `inline::llama-guard` | | safety | `inline::llama-guard` |
| scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` |
| telemetry | `inline::meta-reference` | | telemetry | `inline::meta-reference` |

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: bedrock conda_env: bedrock
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -37,6 +40,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: fireworks conda_env: fireworks
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -39,6 +42,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: hf-endpoint conda_env: hf-endpoint
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -44,6 +47,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: hf-endpoint conda_env: hf-endpoint
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -39,6 +42,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: hf-serverless conda_env: hf-serverless
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -44,6 +47,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: hf-serverless conda_env: hf-serverless
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -39,6 +42,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: meta-reference-gpu conda_env: meta-reference-gpu
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -46,6 +49,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: meta-reference-gpu conda_env: meta-reference-gpu
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -40,6 +43,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: meta-reference-quantized-gpu conda_env: meta-reference-quantized-gpu
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -42,6 +45,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: ollama conda_env: ollama
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -38,6 +41,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: ollama conda_env: ollama
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -38,6 +41,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: tgi conda_env: tgi
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -42,6 +45,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: tgi conda_env: tgi
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -38,6 +41,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: together conda_env: together
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -39,6 +42,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite

View file

@ -16,4 +16,13 @@ distribution_spec:
- inline::meta-reference - inline::meta-reference
telemetry: telemetry:
- inline::meta-reference - inline::meta-reference
eval:
- inline::meta-reference
datasetio:
- remote::huggingface
- inline::localfs
scoring:
- inline::basic
- inline::llm-as-judge
- inline::braintrust
image_type: conda image_type: conda

View file

@ -4,9 +4,12 @@ docker_image: null
conda_env: vllm-gpu conda_env: vllm-gpu
apis: apis:
- agents - agents
- datasetio
- eval
- inference - inference
- memory - memory
- safety - safety
- scoring
- telemetry - telemetry
providers: providers:
inference: inference:
@ -42,6 +45,18 @@ providers:
- provider_id: meta-reference - provider_id: meta-reference
provider_type: inline::meta-reference provider_type: inline::meta-reference
config: {} config: {}
eval:
- provider_id: meta-reference
provider_type: inline::meta-reference
config: {}
datasetio:
- provider_id: huggingface
provider_type: remote::huggingface
config: {}
scoring:
- provider_id: basic
provider_type: inline::basic
config: {}
metadata_store: metadata_store:
namespace: null namespace: null
type: sqlite type: sqlite